Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3460120.3484770acmconferencesArticle/Chapter ViewAbstractPublication PagesccsConference Proceedingsconference-collections
research-article

Membership Inference Attacks Against Recommender Systems

Published: 13 November 2021 Publication History

Abstract

Recently, recommender systems have achieved promising performances and become one of the most widely used web applications. However, recommender systems are often trained on highly sensitive user data, thus potential data leakage from recommender systems may lead to severe privacy problems.
In this paper, we make the first attempt on quantifying the privacy leakage of recommender systems through the lens of membership inference. In contrast with traditional membership inference against machine learning classifiers, our attack faces two main differences. First, our attack is on the user-level but not on the data sample-level. Second, the adversary can only observe the ordered recommended items from a recommender system instead of prediction results in the form of posterior probabilities. To address the above challenges, we propose a novel method by representing users from relevant items. Moreover, a shadow recommender is established to derive the labeled training data for training the attack model. Extensive experimental results show that our attack framework achieves a strong performance. In addition, we design a defense mechanism to effectively mitigate the membership inference threat of recommender systems.

Supplementary Material

MP4 File (CCS21-fp288.mp4)
To investigate the privacy problem in recommender systems, we design various attack strategies of membership inference. To the best of our knowledge, ours is the first work on the membership inference attacks against recommender systems. Comparing to membership inference attacks on data sample-level classifiers, for recommender systems, our work focuses on the user-level membership status, which cannot be directly obtained from the system outputs. To address these challenges, we propose a novel membership inference attack scheme, the core of which is to obtain user-level feature vectors based on the interactions between users and the target recommender, and input these feature vectors into attack models. Extensive experiment results show the effectiveness and generalization ability of our attack. To remedy the situation, we further propose a defense mechanism, namely Popularity Randomization. Our empirical evaluations demonstrate that Popularity Randomization can largely mitigate the privacy risks.

References

[1]
Michael Backes, Mathias Humbert, Jun Pang, and Yang Zhang. walk2friends: Inferring Social Links from Mobility Profiles. In ACM SIGSAC Conference on Computer and Communications Security (CCS), pages 1943--1957. ACM, 2017.
[2]
Ting Bai, Ji-Rong Wen, Jun Zhang, and Wayne Xin Zhao. A Neural Collaborative Filtering Model with Interaction-based Neighborhood. In ACM International Conference on Information and Knowledge Management (CIKM), pages 1979--1982. ACM, 2017.
[3]
Battista Biggio, Igino Corona, Davide Maiorca, Blaine Nelson, Nedim Srndic, Pavel Laskov, Giorgio Giacinto, and Fabio Roli. Evasion Attacks against Machine Learning at Test Time. In European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML/PKDD), pages 387--402. Springer, 2013.
[4]
Iván Cantador, Peter Brusilovsky, and Tsvi Kuflik. Second Workshop on Information Heterogeneity and Fusion in Recommender Systems (HetRec2011). In ACM Conference on Recommender Systems (RecSys), pages 387--388. ACM, 2011.
[5]
Nicholas Carlini, Chang Liu, Úlfar Erlingsson, Jernej Kos, and Dawn Song. The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks. In USENIX Security Symposium (USENIX Security), pages 267--284. USENIX, 2019.
[6]
Nicholas Carlini, Florian Tramè r, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom B. Brown, Dawn Song, Ú lfar Erlingsson, Alina Oprea, and Colin Raffel. Extracting Training Data from Large Language Models. CoRR abs/2012.07805, 2020.
[7]
Nicholas Carlini and David Wagner. Towards Evaluating the Robustness of Neural Networks. In IEEE Symposium on Security and Privacy (S&P), pages 39--57. IEEE, 2017.
[8]
Dingfan Chen, Ning Yu, Yang Zhang, and Mario Fritz. GAN-Leaks: A Taxonomy of Membership Inference Attacks against Generative Models. In ACM SIGSAC Conference on Computer and Communications Security (CCS), pages 343--362. ACM, 2020.
[9]
Wanyu Chen, Fei Cai, Honghui Chen, and Maarten de Rijke. Joint Neural Collaborative Filtering for Recommender Systems. ACM Transactions on Information Systems, 2019.
[10]
Christopher A. Choquette Choo, Florian Tramèr, Nicholas Carlini, and Nicolas Papernot. Label-Only Membership Inference Attacks. CoRR abs/2007.14321, 2020.
[11]
Mukund Deshpande and George Karypis. Item-Based Top-N Recommendation Algorithms. ACM Transactions on Information Systems, 2004.
[12]
Tianyu Gu, Brendan Dolan-Gavitt, and Siddharth Grag. Badnets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain. CoRR abs/1708.06733, 2017.
[13]
F. Maxwell Harper and Joseph A Konstan. The MovieLens Datasets: History and Context. ACM Transactions on Interactive Intelligent Systems, 2015.
[14]
Jamie Hayes, Luca Melis, George Danezis, and Emiliano De Cristofaro. LOGAN: Evaluating Privacy Leakage of Generative Models Using Generative Adversarial Networks. Symposium on Privacy Enhancing Technologies Symposium, 2019.
[15]
Ruining He and Julian McAuley. Ups and Downs: Modeling the Visual Evolution of Fashion Trends with One-Class Collaborative Filtering. In The Web Conference (WWW), pages 507--517. ACM, 2016.
[16]
Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. Neural Collaborative Filtering. In International Conference on World Wide Web (WWW), pages 173--182. ACM, 2017.
[17]
Xiangnan He, Hanwang Zhang, Min-Yen Kan, and Tat-Seng Chua. Fast Matrix Factorization for Online Recommendation with Implicit Feedback. In International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), pages 549--558. ACM, 2016.
[18]
Xinlei He, Jinyuan Jia, Michael Backes, Neil Zhenqiang Gong, and Yang Zhang. Stealing Links from Graph Neural Networks. In USENIX Security Symposium (USENIX Security). USENIX, 2021.
[19]
Jonathan L. Herlocker, Joseph A. Konstan, and John Riedl. Explaining Collaborative Filtering Recommendations. In ACM Conference on Computer Supported Cooperative Work (CSCW), pages 241--250. ACM, 2000.
[20]
Matthew Jagielski, Nicholas Carlini, David Berthelot, Alex Kurakin, and Nicolas Papernot. High Accuracy and High Fidelity Extraction of Neural Networks. In USENIX Security Symposium (USENIX Security), pages 1345--1362. USENIX, 2020.
[21]
Matthew Jagielski, Alina Oprea, Battista Biggio, Chang Liu, Cristina Nita-Rotaru, and Bo Li. Manipulating Machine Learning: Poisoning Attacks and Countermeasures for Regression Learning. In IEEE Symposium on Security and Privacy (S&P), pages 19--35. IEEE, 2018.
[22]
Jinyuan Jia, Ahmed Salem, Michael Backes, Yang Zhang, and Neil Zhenqiang Gong. MemGuard: Defending against Black-Box Membership Inference Attacks via Adversarial Examples. In ACM SIGSAC Conference on Computer and Communications Security (CCS), pages 259--274. ACM, 2019.
[23]
George Karypis. Evaluation of Item-Based Top-N Recommendation Algorithms. In ACM International Conference on Information and Knowledge Management (CIKM), pages 247--254. ACM, 2001.
[24]
Yehuda Koren. Factorization Meets the Neighborhood: a Multifaceted Collaborative Filtering Model. In ACM Conference on Knowledge Discovery and Data Mining (KDD), pages 426--434. ACM, 2008.
[25]
Yehuda Koren. Collaborative Filtering with Temporal Dynamics. In ACM Conference on Knowledge Discovery and Data Mining (KDD), pages 447--456. ACM, 2009.
[26]
Klas Leino and Matt Fredrikson. Stolen Memories: Leveraging Model Memorization for Calibrated White-Box Membership Inference. In USENIX Security Symposium (USENIX Security), pages 1605--1622. USENIX, 2020.
[27]
Zheng Li, Chengyu Hu, Yang Zhang, and Shanqing Guo. How to Prove Your Model Belongs to You: A Blind-Watermark based Framework to Protect Intellectual Property of DNN. In Annual Computer Security Applications Conference (ACSAC), pages 126--137. ACM, 2019.
[28]
Zheng Li and Yang Zhang. Membership Leakage in Label-Only Exposures. In ACM SIGSAC Conference on Computer and Communications Security (CCS). ACM, 2021.
[29]
Luca Melis, Congzheng Song, Emiliano De Cristofaro, and Vitaly Shmatikov. Exploiting Unintended Feature Leakage in Collaborative Learning. In IEEE Symposium on Security and Privacy (S&P), pages 497--512. IEEE, 2019.
[30]
Milad Nasr, Reza Shokri, and Amir Houmansadr. Machine Learning with Membership Privacy using Adversarial Regularization. In ACM SIGSAC Conference on Computer and Communications Security (CCS), pages 634--646. ACM, 2018.
[31]
Milad Nasr, Reza Shokri, and Amir Houmansadr. Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning. In IEEE Symposium on Security and Privacy (S&P), pages 1021--1035. IEEE, 2019.
[32]
Milad Nasr, Shuang Song, Abhradeep Thakurta, Nicolas Papernot, and Nicholas Carlini. Adversary Instantiation: Lower Bounds for Differentially Private Machine Learning. In IEEE Symposium on Security and Privacy (S&P). IEEE, 2021.
[33]
Nicolas Papernot, Patrick McDaniel, Arunesh Sinha, and Michael Wellman. SoK: Towards the Science of Security and Privacy in Machine Learning. In IEEE European Symposium on Security and Privacy (Euro S&P), pages 399--414. IEEE, 2018.
[34]
Nicolas Papernot, Shuang Song, Ilya Mironov, Ananth Raghunathan, Kunal Talwar, and Ú lfar Erlingsson. Scalable Private Learning with PATE. In International Conference on Learning Representations (ICLR), 2018.
[35]
Michael J. Pazzani and Daniel Billsus. Content-Based Recommendation Systems. In The Adaptive Web, Methods and Strategies of Web Personalization, pages 325--341. Springer, 2007.
[36]
Huseyin Polat and Wenliang Du. SVD-based Collaborative Filtering with Privacy. In ACM Symposium on Applied Computing (SAC), pages 791--795. ACM, 2005.
[37]
Alexandre Sablayrolles, Matthijs Douze, Cordelia Schmid, Yann Ollivier, and Hervé Jégou. White-box vs Black-box: Bayes Optimal Strategies for Membership Inference. In International Conference on Machine Learning (ICML), pages 5558--5567. PMLR, 2019.
[38]
Ruslan Salakhutdinov and Andriy Mnih. Probabilistic Matrix Factorization. In Annual Conference on Neural Information Processing Systems (NIPS), pages 1257--1264. NIPS, 2007.
[39]
Ahmed Salem, Yang Zhang, Mathias Humbert, Pascal Berrang, Mario Fritz, and Michael Backes. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models. In Network and Distributed System Security Symposium (NDSS). Internet Society, 2019.
[40]
Badrul Munir Sarwar, George Karypis, Joseph A. Konstan, and John Riedl. Item-Based Collaborative Filtering Recommendation Algorithms. In International Conference on World Wide Web (WWW), pages 285--295. ACM, 2001.
[41]
J. Ben Schafer, Dan Frankowski, Jon Herlocker, and Shilad Sen. Collaborative Filtering Recommender Systems. In The Adaptive Web, Methods and Strategies of Web Personalization, pages 291--324. Springer, 2007.
[42]
Virat Shejwalkar and Amir Houmansadr. Membership Privacy for Machine Learning Models Through Knowledge Transfer. In AAAI Conference on Artificial Intelligence (AAAI). AAAI, 2021.
[43]
Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. Membership Inference Attacks Against Machine Learning Models. In IEEE Symposium on Security and Privacy (S&P), pages 3--18. IEEE, 2017.
[44]
Reza Shokri, Georgios Theodorakopoulos, Jean-Yves Le Boudec, and Jean-Pierre Hubaux. Quantifying Location Privacy. In IEEE Symposium on Security and Privacy (S&P), pages 247--262. IEEE, 2011.
[45]
Congzheng Song and Vitaly Shmatikov. Auditing Data Provenance in Text-Generation Models. In ACM Conference on Knowledge Discovery and Data Mining (KDD), pages 196--206. ACM, 2019.
[46]
Peijie Sun, Le Wu, and Meng Wang. Attentive Recurrent Social Recommendation. In International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), pages 185--194. ACM, 2018.
[47]
Peijie Sun, Le Wu, Kun Zhang, Yanjie Fu, Richang Hong, and Meng Wang. Dual Learning for Explainable Recommendation: Towards Unifying User Preference Prediction and Review Generation. In The Web Conference (WWW), pages 837--847. ACM, 2020.
[48]
Florian Tramèr, Alexey Kurakin, Nicolas Papernot, Ian Goodfellow, Dan Boneh, and Patrick McDaniel. Ensemble Adversarial Training: Attacks and Defenses. In International Conference on Learning Representations (ICLR), 2017.
[49]
Florian Tramèr, Fan Zhang, Ari Juels, Michael K. Reiter, and Thomas Ristenpart. Stealing Machine Learning Models via Prediction APIs. In USENIX Security Symposium (USENIX Security), pages 601--618. USENIX, 2016.
[50]
Laurens van der Maaten and Geoffrey Hinton. Visualizing Data using t-SNE. Journal of Machine Learning Research, 2008.
[51]
Bogdan Walek and Vladimir Fojtik. A Hybrid Recommender System for Recommending Relevant Movies Using An Expert System. Expert Systems with Applications, 2020.
[52]
Samuel Yeom, Irene Giacomelli, Matt Fredrikson, and Somesh Jha. Privacy Risk in Machine Learning: Analyzing the Connection to Overfitting. In IEEE Computer Security Foundations Symposium (CSF), pages 268--282. IEEE, 2018.

Cited By

View all
  • (2024)Machine learning security and privacy: a review of threats and countermeasuresEURASIP Journal on Information Security10.1186/s13635-024-00158-32024:1Online publication date: 23-Apr-2024
  • (2024)Defense Against Model Extraction Attacks on Recommender SystemsProceedings of the 17th ACM International Conference on Web Search and Data Mining10.1145/3616855.3635751(949-957)Online publication date: 4-Mar-2024
  • (2024)Generated Distributions Are All You Need for Membership Inference Attacks Against Generative Models2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)10.1109/WACV57701.2024.00477(4827-4837)Online publication date: 3-Jan-2024
  • Show More Cited By

Index Terms

  1. Membership Inference Attacks Against Recommender Systems

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      CCS '21: Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security
      November 2021
      3558 pages
      ISBN:9781450384544
      DOI:10.1145/3460120
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 13 November 2021

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. membership inference attack
      2. membership leakage
      3. recommender system

      Qualifiers

      • Research-article

      Funding Sources

      • Hybrid Intelligence Center
      • Natural Science Foundation of China
      • A 10-year program funded by the Dutch Ministry of Education, Culture and Science through the Netherlands Organisation for Scientific Research, https://hybrid-intelligence-centre.nl
      • Shandong University multidisciplinary research and innovation team of young scholars
      • National Key R&D Program of China
      • Tencent WeChat Rhino-Bird Focused Research Program
      • the Helmholtz Association within the project ``Trustworthy Federated Data Analytics' (TFDA)
      • Key Scientific and Technological Innovation Program of Shandong Province

      Conference

      CCS '21
      Sponsor:
      CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security
      November 15 - 19, 2021
      Virtual Event, Republic of Korea

      Acceptance Rates

      Overall Acceptance Rate 1,261 of 6,999 submissions, 18%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)269
      • Downloads (Last 6 weeks)30
      Reflects downloads up to 16 Oct 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Machine learning security and privacy: a review of threats and countermeasuresEURASIP Journal on Information Security10.1186/s13635-024-00158-32024:1Online publication date: 23-Apr-2024
      • (2024)Defense Against Model Extraction Attacks on Recommender SystemsProceedings of the 17th ACM International Conference on Web Search and Data Mining10.1145/3616855.3635751(949-957)Online publication date: 4-Mar-2024
      • (2024)Generated Distributions Are All You Need for Membership Inference Attacks Against Generative Models2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)10.1109/WACV57701.2024.00477(4827-4837)Online publication date: 3-Jan-2024
      • (2024)Data Provenance via Differential AuditingIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2023.333482136:10(5066-5079)Online publication date: Oct-2024
      • (2024)Please Tell Me More: Privacy Impact of Explainability through the Lens of Membership Inference Attack2024 IEEE Symposium on Security and Privacy (SP)10.1109/SP54263.2024.00120(4791-4809)Online publication date: 19-May-2024
      • (2024)Test-Time Poisoning Attacks Against Test-Time Adaptation Models2024 IEEE Symposium on Security and Privacy (SP)10.1109/SP54263.2024.00072(1306-1324)Online publication date: 19-May-2024
      • (2024)GradDiff: Gradient-based membership inference attacks against federated distillation with differential comparisonInformation Sciences10.1016/j.ins.2023.120068658(120068)Online publication date: Feb-2024
      • (2024)HAMIATCM: high-availability membership inference attack against text classification models under little knowledgeApplied Intelligence10.1007/s10489-024-05495-x54:17-18(7994-8019)Online publication date: 19-Jun-2024
      • (2024)The Impact of Differential Privacy on Recommendation Accuracy and Popularity BiasAdvances in Information Retrieval10.1007/978-3-031-56066-8_33(466-482)Online publication date: 24-Mar-2024
      • (2023)Differential testing of cross deep learning framework APIsProceedings of the 32nd USENIX Conference on Security Symposium10.5555/3620237.3620651(7393-7410)Online publication date: 9-Aug-2023
      • Show More Cited By

      View Options

      Get Access

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media