Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3133956.3134012acmconferencesArticle/Chapter ViewAbstractPublication PagesccsConference Proceedingsconference-collections
research-article

Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning

Published: 30 October 2017 Publication History

Abstract

Deep Learning has recently become hugely popular in machine learning for its ability to solve end-to-end learning systems, in which the features and the classifiers are learned simultaneously, providing significant improvements in classification accuracy in the presence of highly-structured and large databases.
Its success is due to a combination of recent algorithmic breakthroughs, increasingly powerful computers, and access to significant amounts of data.
Researchers have also considered privacy implications of deep learning. Models are typically trained in a centralized manner with all the data being processed by the same training algorithm. If the data is a collection of users' private data, including habits, personal pictures, geographical positions, interests, and more, the centralized server will have access to sensitive information that could potentially be mishandled. To tackle this problem, collaborative deep learning models have recently been proposed where parties locally train their deep learning structures and only share a subset of the parameters in the attempt to keep their respective training sets private. Parameters can also be obfuscated via differential privacy (DP) to make information extraction even more challenging, as proposed by Shokri and Shmatikov at CCS'15.
Unfortunately, we show that any privacy-preserving collaborative deep learning is susceptible to a powerful attack that we devise in this paper. In particular, we show that a distributed, federated, or decentralized deep learning approach is fundamentally broken and does not protect the training sets of honest participants. The attack we developed exploits the real-time nature of the learning process that allows the adversary to train a Generative Adversarial Network (GAN) that generates prototypical samples of the targeted training set that was meant to be private (the samples generated by the GAN are intended to come from the same distribution as the training data). Interestingly, we show that record-level differential privacy applied to the shared parameters of the model, as suggested in previous work, is ineffective (i.e., record-level DP is not designed to address our attack).

Supplemental Material

MP4 File

References

[1]
Martín Abadi, Andy Chu, Ian Goodfellow, H Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang 2016. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security. ACM, 308--318.
[2]
Ahmad Abdulkader, Aparna Lakshmiratan, and Joy Zhang. 2016. Introducing DeepText: Facebook's text understanding engine. (2016). https://tinyurl.com/jj359dv
[3]
Martin Arjovsky and Léon Bottou 2017. Towards principled methods for training generative adversarial networks 5th International Conference on Learning Representations (ICLR).
[4]
Giuseppe Ateniese, Luigi V Mancini, Angelo Spognardi, Antonio Villani, Domenico Vitali, and Giovanni Felici 2015. Hacking smart machines with smarter ones: How to extract meaningful data from machine learning classifiers. International Journal of Security and Networks, Vol. 10, 3 (2015), 137--150. https://arxiv.org/abs/1306.4447
[5]
Gilles Barthe, Noémie Fong, Marco Gaboardi, Benjamin Grégoire, Justin Hsu, and Pierre-Yves Strub 2016. Advanced probabilistic couplings for differential privacy Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security. ACM, 55--67.
[6]
Yoshua Bengio. 2009. Learning Deep Architectures for AI. Found. Trends Mach. Learn. Vol. 2, 1 (Jan. 2009), 1--127. 1109/CVPR.2014.220
[7]
Florian Tramèr, Fan Zhang, Ari Juels, Michael K Reiter, and Thomas Ristenpart 2016. Stealing Machine Learning Models via Prediction APIs USENIX Security.
[8]
Vladimir Naumovich Vapnik and Vlamimir Vapnik 1998. Statistical learning theory. Vol. Vol. 1. Wiley New York.
[9]
Martin J Wainwright, Michael I Jordan, and John C Duchi. 2012. Privacy aware learning. In Advances in Neural Information Processing Systems. 1430--1438.
[10]
Pengtao Xie, Misha Bilenko, Tom Finley, Ran Gilad-Bachrach, Kristin Lauter, and Michael Naehrig. 2014. Crypto-nets: Neural networks over encrypted data. arXiv preprint arXiv:1412.6181 (2014).
[11]
Weilin Xu, Yanjun Qi, and David Evans 2016. Automatically evading classifiers. In NDSS'16.
[12]
Jun Zhang, Zhenjie Zhang, Xiaokui Xiao, Yin Yang, and Marianne Winslett 2012. Functional mechanism: regression analysis under differential privacy. Proceedings of the VLDB Endowment Vol. 5, 11 (2012), 1364--1375.
[13]
Tong Zhang. 2004. Solving large scale linear prediction problems using stochastic gradient descent algorithms Proceedings of the twenty-first international conference on Machine learning. ACM, 116.
[14]
Xiang Zhang and Yann André LeCun 2016. Text Understanding from Scratch. arXiv preprint arXiv:1502.01710v5 (2016).
[15]
Martin Zinkevich, Markus Weimer, Lihong Li, and Alex J Smola. 2010. Parallelized stochastic gradient descent. In Advances in neural information processing systems. 2595--2603.

Cited By

View all
  • (2024)Efficient secure aggregation for privacy-preserving federated learning based on secret sharingJUSTC10.52396/JUSTC-2022-011654:1(0104)Online publication date: 2024
  • (2024)Secure AI Model Sharing: A Cryptographic Approach for Encrypted Model ExchangeInternational Journal of Artificial Intelligence and Machine Learning10.51483/IJAIML.4.1.2024.48-604:1(48-60)Online publication date: 5-Jan-2024
  • (2024)Human Resources Optimization for Public Space SecurityEnhancing Security in Public Spaces Through Generative Adversarial Networks (GANs)10.4018/979-8-3693-3597-0.ch019(274-295)Online publication date: 16-May-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
CCS '17: Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security
October 2017
2682 pages
ISBN:9781450349468
DOI:10.1145/3133956
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 30 October 2017

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. collaborative learning
  2. deep learning
  3. privacy
  4. security

Qualifiers

  • Research-article

Conference

CCS '17
Sponsor:

Acceptance Rates

CCS '17 Paper Acceptance Rate 151 of 836 submissions, 18%;
Overall Acceptance Rate 1,261 of 6,999 submissions, 18%

Upcoming Conference

CCS '24
ACM SIGSAC Conference on Computer and Communications Security
October 14 - 18, 2024
Salt Lake City , UT , USA

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)686
  • Downloads (Last 6 weeks)35
Reflects downloads up to 30 Aug 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Efficient secure aggregation for privacy-preserving federated learning based on secret sharingJUSTC10.52396/JUSTC-2022-011654:1(0104)Online publication date: 2024
  • (2024)Secure AI Model Sharing: A Cryptographic Approach for Encrypted Model ExchangeInternational Journal of Artificial Intelligence and Machine Learning10.51483/IJAIML.4.1.2024.48-604:1(48-60)Online publication date: 5-Jan-2024
  • (2024)Human Resources Optimization for Public Space SecurityEnhancing Security in Public Spaces Through Generative Adversarial Networks (GANs)10.4018/979-8-3693-3597-0.ch019(274-295)Online publication date: 16-May-2024
  • (2024)Enhancing Privacy and Security in Online Education Using Generative Adversarial NetworksEnhancing Security in Public Spaces Through Generative Adversarial Networks (GANs)10.4018/979-8-3693-3597-0.ch015(206-230)Online publication date: 16-May-2024
  • (2024)Enhancing Cyber Security Through Generative Adversarial NetworksEnhancing Security in Public Spaces Through Generative Adversarial Networks (GANs)10.4018/979-8-3693-3597-0.ch013(177-192)Online publication date: 16-May-2024
  • (2024)Revolutionizing Healthcare Harnessing IoT-Integrated Federated Learning for Early Disease Detection and Patient Privacy PreservationFederated Learning and Privacy-Preserving in Healthcare AI10.4018/979-8-3693-1874-4.ch013(195-216)Online publication date: 19-Apr-2024
  • (2024)Secure Aggregation Protocol Based on DC-Nets and Secret Sharing for Decentralized Federated LearningSensors10.3390/s2404129924:4(1299)Online publication date: 17-Feb-2024
  • (2024)EPFed: Achieving Optimal Balance between Privacy and Efficiency in Federated LearningElectronics10.3390/electronics1306102813:6(1028)Online publication date: 9-Mar-2024
  • (2024)Artificial Intelligence-Based Algorithms in Medical Image Scan Segmentation and Intelligent Visual Content Generation—A Concise OverviewElectronics10.3390/electronics1304074613:4(746)Online publication date: 13-Feb-2024
  • (2024)Improving Security in the Internet of Vehicles: A Blockchain-Based Data Sharing SchemeElectronics10.3390/electronics1304071413:4(714)Online publication date: 9-Feb-2024
  • Show More Cited By

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media