Given a collection of rankings of a set of items, rank aggregation seeks to compute a ranking that can serve as a single best representative of the collection. Rank aggregation is a well-studied problem and a number of effective algorithmic solutions have been proposed in the literature. However, when individuals are asked to contribute a ranking, they may be concerned that their personal preferences will be disclosed inappropriately to others. This acts as a disincentive to individuals to respond honestly in expressing their preferences and impedes data collection and data sharing. We address this problem by investigating rank aggregation under differential privacy, which requires that a released output (here, the aggregate ranking computed from individuals' rankings) remain almost the same if any one individual's ranking is removed from the input. We propose a number of differentially-private rank aggregation algorithms: two are inspired by non-private approximate rank aggregators from the existing literature; another uses a novel rejection sampling method to sample privately from a complex distribution. For all the methods we propose, we quantify, both theoretically and empirically, the "cost" of privacy in terms of the quality of the rank aggregation computed.
CITATION STYLE
Hay, M., Elagina, L., & Miklau, G. (2017). Differentially private rank aggregation. In Proceedings of the 17th SIAM International Conference on Data Mining, SDM 2017 (pp. 669–677). Society for Industrial and Applied Mathematics Publications. https://doi.org/10.1137/1.9781611974973.75
Mendeley helps you to discover research relevant for your work.