Differential privacy in collaborative filtering recommender systems: a review

Peter Müllner*, Elisabeth Lex*, Markus Schedl, Dominik Kowald*

*Corresponding author for this work

Research output: Contribution to journalShort surveypeer-review

Abstract

State-of-the-art recommender systems produce high-quality recommendations to support users in finding relevant content. However, through the utilization of users' data for generating recommendations, recommender systems threaten users' privacy. To alleviate this threat, often, differential privacy is used to protect users' data via adding random noise. This, however, leads to a substantial drop in recommendation quality. Therefore, several approaches aim to improve this trade-off between accuracy and user privacy. In this work, we first overview threats to user privacy in recommender systems, followed by a brief introduction to the differential privacy framework that can protect users' privacy. Subsequently, we review recommendation approaches that apply differential privacy, and we highlight research that improves the trade-off between recommendation quality and user privacy. Finally, we discuss open issues, e.g., considering the relation between privacy and fairness, and the users' different needs for privacy. With this review, we hope to provide other researchers an overview of the ways in which differential privacy has been applied to state-of-the-art collaborative filtering recommender systems.

Original languageEnglish
Article number1249997
JournalFrontiers in Big Data
Volume6
DOIs
Publication statusPublished - 2023

Keywords

  • accuracy-privacy trade-off
  • collaborative filtering
  • differential privacy
  • recommender systems
  • review

ASJC Scopus subject areas

  • Computer Science (miscellaneous)
  • Information Systems
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Differential privacy in collaborative filtering recommender systems: a review'. Together they form a unique fingerprint.

Cite this