ABSTRACT
Recently, recommender systems have achieved promising performances and become one of the most widely used web applications. However, recommender systems are often trained on highly sensitive user data, thus potential data leakage from recommender systems may lead to severe privacy problems.
In this paper, we make the first attempt on quantifying the privacy leakage of recommender systems through the lens of membership inference. In contrast with traditional membership inference against machine learning classifiers, our attack faces two main differences. First, our attack is on the user-level but not on the data sample-level. Second, the adversary can only observe the ordered recommended items from a recommender system instead of prediction results in the form of posterior probabilities. To address the above challenges, we propose a novel method by representing users from relevant items. Moreover, a shadow recommender is established to derive the labeled training data for training the attack model. Extensive experimental results show that our attack framework achieves a strong performance. In addition, we design a defense mechanism to effectively mitigate the membership inference threat of recommender systems.
Supplemental Material
- Michael Backes, Mathias Humbert, Jun Pang, and Yang Zhang. walk2friends: Inferring Social Links from Mobility Profiles. In ACM SIGSAC Conference on Computer and Communications Security (CCS), pages 1943--1957. ACM, 2017.Google Scholar
- Ting Bai, Ji-Rong Wen, Jun Zhang, and Wayne Xin Zhao. A Neural Collaborative Filtering Model with Interaction-based Neighborhood. In ACM International Conference on Information and Knowledge Management (CIKM), pages 1979--1982. ACM, 2017.Google Scholar
- Battista Biggio, Igino Corona, Davide Maiorca, Blaine Nelson, Nedim Srndic, Pavel Laskov, Giorgio Giacinto, and Fabio Roli. Evasion Attacks against Machine Learning at Test Time. In European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML/PKDD), pages 387--402. Springer, 2013.Google Scholar
Digital Library
- Iván Cantador, Peter Brusilovsky, and Tsvi Kuflik. Second Workshop on Information Heterogeneity and Fusion in Recommender Systems (HetRec2011). In ACM Conference on Recommender Systems (RecSys), pages 387--388. ACM, 2011.Google Scholar
Digital Library
- Nicholas Carlini, Chang Liu, Úlfar Erlingsson, Jernej Kos, and Dawn Song. The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks. In USENIX Security Symposium (USENIX Security), pages 267--284. USENIX, 2019.Google Scholar
- Nicholas Carlini, Florian Tramè r, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom B. Brown, Dawn Song, Ú lfar Erlingsson, Alina Oprea, and Colin Raffel. Extracting Training Data from Large Language Models. CoRR abs/2012.07805, 2020.Google Scholar
- Nicholas Carlini and David Wagner. Towards Evaluating the Robustness of Neural Networks. In IEEE Symposium on Security and Privacy (S&P), pages 39--57. IEEE, 2017.Google Scholar
- Dingfan Chen, Ning Yu, Yang Zhang, and Mario Fritz. GAN-Leaks: A Taxonomy of Membership Inference Attacks against Generative Models. In ACM SIGSAC Conference on Computer and Communications Security (CCS), pages 343--362. ACM, 2020.Google Scholar
Digital Library
- Wanyu Chen, Fei Cai, Honghui Chen, and Maarten de Rijke. Joint Neural Collaborative Filtering for Recommender Systems. ACM Transactions on Information Systems, 2019.Google Scholar
Digital Library
- Christopher A. Choquette Choo, Florian Tramèr, Nicholas Carlini, and Nicolas Papernot. Label-Only Membership Inference Attacks. CoRR abs/2007.14321, 2020.Google Scholar
- Mukund Deshpande and George Karypis. Item-Based Top-N Recommendation Algorithms. ACM Transactions on Information Systems, 2004.Google Scholar
Digital Library
- Tianyu Gu, Brendan Dolan-Gavitt, and Siddharth Grag. Badnets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain. CoRR abs/1708.06733, 2017.Google Scholar
- F. Maxwell Harper and Joseph A Konstan. The MovieLens Datasets: History and Context. ACM Transactions on Interactive Intelligent Systems, 2015.Google Scholar
- Jamie Hayes, Luca Melis, George Danezis, and Emiliano De Cristofaro. LOGAN: Evaluating Privacy Leakage of Generative Models Using Generative Adversarial Networks. Symposium on Privacy Enhancing Technologies Symposium, 2019.Google Scholar
- Ruining He and Julian McAuley. Ups and Downs: Modeling the Visual Evolution of Fashion Trends with One-Class Collaborative Filtering. In The Web Conference (WWW), pages 507--517. ACM, 2016.Google Scholar
- Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. Neural Collaborative Filtering. In International Conference on World Wide Web (WWW), pages 173--182. ACM, 2017.Google Scholar
- Xiangnan He, Hanwang Zhang, Min-Yen Kan, and Tat-Seng Chua. Fast Matrix Factorization for Online Recommendation with Implicit Feedback. In International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), pages 549--558. ACM, 2016.Google Scholar
- Xinlei He, Jinyuan Jia, Michael Backes, Neil Zhenqiang Gong, and Yang Zhang. Stealing Links from Graph Neural Networks. In USENIX Security Symposium (USENIX Security). USENIX, 2021.Google Scholar
- Jonathan L. Herlocker, Joseph A. Konstan, and John Riedl. Explaining Collaborative Filtering Recommendations. In ACM Conference on Computer Supported Cooperative Work (CSCW), pages 241--250. ACM, 2000.Google Scholar
- Matthew Jagielski, Nicholas Carlini, David Berthelot, Alex Kurakin, and Nicolas Papernot. High Accuracy and High Fidelity Extraction of Neural Networks. In USENIX Security Symposium (USENIX Security), pages 1345--1362. USENIX, 2020.Google Scholar
- Matthew Jagielski, Alina Oprea, Battista Biggio, Chang Liu, Cristina Nita-Rotaru, and Bo Li. Manipulating Machine Learning: Poisoning Attacks and Countermeasures for Regression Learning. In IEEE Symposium on Security and Privacy (S&P), pages 19--35. IEEE, 2018.Google Scholar
- Jinyuan Jia, Ahmed Salem, Michael Backes, Yang Zhang, and Neil Zhenqiang Gong. MemGuard: Defending against Black-Box Membership Inference Attacks via Adversarial Examples. In ACM SIGSAC Conference on Computer and Communications Security (CCS), pages 259--274. ACM, 2019.Google Scholar
- George Karypis. Evaluation of Item-Based Top-N Recommendation Algorithms. In ACM International Conference on Information and Knowledge Management (CIKM), pages 247--254. ACM, 2001.Google Scholar
- Yehuda Koren. Factorization Meets the Neighborhood: a Multifaceted Collaborative Filtering Model. In ACM Conference on Knowledge Discovery and Data Mining (KDD), pages 426--434. ACM, 2008.Google Scholar
Digital Library
- Yehuda Koren. Collaborative Filtering with Temporal Dynamics. In ACM Conference on Knowledge Discovery and Data Mining (KDD), pages 447--456. ACM, 2009.Google Scholar
- Klas Leino and Matt Fredrikson. Stolen Memories: Leveraging Model Memorization for Calibrated White-Box Membership Inference. In USENIX Security Symposium (USENIX Security), pages 1605--1622. USENIX, 2020.Google Scholar
- Zheng Li, Chengyu Hu, Yang Zhang, and Shanqing Guo. How to Prove Your Model Belongs to You: A Blind-Watermark based Framework to Protect Intellectual Property of DNN. In Annual Computer Security Applications Conference (ACSAC), pages 126--137. ACM, 2019.Google Scholar
Digital Library
- Zheng Li and Yang Zhang. Membership Leakage in Label-Only Exposures. In ACM SIGSAC Conference on Computer and Communications Security (CCS). ACM, 2021.Google Scholar
- Luca Melis, Congzheng Song, Emiliano De Cristofaro, and Vitaly Shmatikov. Exploiting Unintended Feature Leakage in Collaborative Learning. In IEEE Symposium on Security and Privacy (S&P), pages 497--512. IEEE, 2019.Google Scholar
- Milad Nasr, Reza Shokri, and Amir Houmansadr. Machine Learning with Membership Privacy using Adversarial Regularization. In ACM SIGSAC Conference on Computer and Communications Security (CCS), pages 634--646. ACM, 2018.Google Scholar
Digital Library
- Milad Nasr, Reza Shokri, and Amir Houmansadr. Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning. In IEEE Symposium on Security and Privacy (S&P), pages 1021--1035. IEEE, 2019.Google Scholar
Cross Ref
- Milad Nasr, Shuang Song, Abhradeep Thakurta, Nicolas Papernot, and Nicholas Carlini. Adversary Instantiation: Lower Bounds for Differentially Private Machine Learning. In IEEE Symposium on Security and Privacy (S&P). IEEE, 2021.Google Scholar
- Nicolas Papernot, Patrick McDaniel, Arunesh Sinha, and Michael Wellman. SoK: Towards the Science of Security and Privacy in Machine Learning. In IEEE European Symposium on Security and Privacy (Euro S&P), pages 399--414. IEEE, 2018.Google Scholar
- Nicolas Papernot, Shuang Song, Ilya Mironov, Ananth Raghunathan, Kunal Talwar, and Ú lfar Erlingsson. Scalable Private Learning with PATE. In International Conference on Learning Representations (ICLR), 2018.Google Scholar
- Michael J. Pazzani and Daniel Billsus. Content-Based Recommendation Systems. In The Adaptive Web, Methods and Strategies of Web Personalization, pages 325--341. Springer, 2007.Google Scholar
- Huseyin Polat and Wenliang Du. SVD-based Collaborative Filtering with Privacy. In ACM Symposium on Applied Computing (SAC), pages 791--795. ACM, 2005.Google Scholar
- Alexandre Sablayrolles, Matthijs Douze, Cordelia Schmid, Yann Ollivier, and Hervé Jégou. White-box vs Black-box: Bayes Optimal Strategies for Membership Inference. In International Conference on Machine Learning (ICML), pages 5558--5567. PMLR, 2019.Google Scholar
- Ruslan Salakhutdinov and Andriy Mnih. Probabilistic Matrix Factorization. In Annual Conference on Neural Information Processing Systems (NIPS), pages 1257--1264. NIPS, 2007.Google Scholar
- Ahmed Salem, Yang Zhang, Mathias Humbert, Pascal Berrang, Mario Fritz, and Michael Backes. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models. In Network and Distributed System Security Symposium (NDSS). Internet Society, 2019.Google Scholar
- Badrul Munir Sarwar, George Karypis, Joseph A. Konstan, and John Riedl. Item-Based Collaborative Filtering Recommendation Algorithms. In International Conference on World Wide Web (WWW), pages 285--295. ACM, 2001.Google Scholar
- J. Ben Schafer, Dan Frankowski, Jon Herlocker, and Shilad Sen. Collaborative Filtering Recommender Systems. In The Adaptive Web, Methods and Strategies of Web Personalization, pages 291--324. Springer, 2007.Google Scholar
Cross Ref
- Virat Shejwalkar and Amir Houmansadr. Membership Privacy for Machine Learning Models Through Knowledge Transfer. In AAAI Conference on Artificial Intelligence (AAAI). AAAI, 2021.Google Scholar
- Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. Membership Inference Attacks Against Machine Learning Models. In IEEE Symposium on Security and Privacy (S&P), pages 3--18. IEEE, 2017.Google Scholar
- Reza Shokri, Georgios Theodorakopoulos, Jean-Yves Le Boudec, and Jean-Pierre Hubaux. Quantifying Location Privacy. In IEEE Symposium on Security and Privacy (S&P), pages 247--262. IEEE, 2011.Google Scholar
- Congzheng Song and Vitaly Shmatikov. Auditing Data Provenance in Text-Generation Models. In ACM Conference on Knowledge Discovery and Data Mining (KDD), pages 196--206. ACM, 2019.Google Scholar
- Peijie Sun, Le Wu, and Meng Wang. Attentive Recurrent Social Recommendation. In International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), pages 185--194. ACM, 2018.Google Scholar
- Peijie Sun, Le Wu, Kun Zhang, Yanjie Fu, Richang Hong, and Meng Wang. Dual Learning for Explainable Recommendation: Towards Unifying User Preference Prediction and Review Generation. In The Web Conference (WWW), pages 837--847. ACM, 2020.Google Scholar
- Florian Tramèr, Alexey Kurakin, Nicolas Papernot, Ian Goodfellow, Dan Boneh, and Patrick McDaniel. Ensemble Adversarial Training: Attacks and Defenses. In International Conference on Learning Representations (ICLR), 2017.Google Scholar
- Florian Tramèr, Fan Zhang, Ari Juels, Michael K. Reiter, and Thomas Ristenpart. Stealing Machine Learning Models via Prediction APIs. In USENIX Security Symposium (USENIX Security), pages 601--618. USENIX, 2016.Google Scholar
- Laurens van der Maaten and Geoffrey Hinton. Visualizing Data using t-SNE. Journal of Machine Learning Research, 2008.Google Scholar
- Bogdan Walek and Vladimir Fojtik. A Hybrid Recommender System for Recommending Relevant Movies Using An Expert System. Expert Systems with Applications, 2020.Google Scholar
Cross Ref
- Samuel Yeom, Irene Giacomelli, Matt Fredrikson, and Somesh Jha. Privacy Risk in Machine Learning: Analyzing the Connection to Overfitting. In IEEE Computer Security Foundations Symposium (CSF), pages 268--282. IEEE, 2018.Google Scholar
Cross Ref
Index Terms
Membership Inference Attacks Against Recommender Systems
Recommendations
Debiasing Learning for Membership Inference Attacks Against Recommender Systems
Learned recommender systems may inadvertently leak information about their training data, leading to privacy violations. We investigate privacy threats faced by recommender systems through the lens of membership inference. In such attacks, an adversary ...
Interaction-level Membership Inference Attack Against Federated Recommender Systems
The marriage of federated learning and recommender system (FedRec) has been widely used to address the growing data privacy concerns in personalized recommendation services. In FedRecs, users’ attribute information and behavior data (i.e., user-item ...
Membership Inference Attacks Against Sequential Recommender Systems
Recent studies have demonstrated the vulnerability of recommender systems to membership inference attacks, which determine whether a user’s historical data was utilized for model training, posing serious privacy leakage issues. Existing works assumed ...






Comments