Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3422337.3447836acmconferencesArticle/Chapter ViewAbstractPublication PagescodaspyConference Proceedingsconference-collections
research-article
Open access

Membership Inference Attacks and Defenses in Classification Models

Published: 26 April 2021 Publication History

Abstract

We study the membership inference (MI) attack against classifiers, where the attacker's goal is to determine whether a data instance was used for training the classifier. Through systematic cataloging of existing MI attacks and extensive experimental evaluations of them, we find that a model's vulnerability to MI attacks is tightly related to the generalization gap---the difference between training accuracy and test accuracy. We then propose a defense against MI attacks that aims to close the gap by intentionally reduces the training accuracy. More specifically, the training process attempts to match the training and validation accuracies, by means of a new set regularizer using the Maximum Mean Discrepancy between the softmax output empirical distributions of the training and validation sets. Our experimental results show that combining this approach with another simple defense (mix-up training) significantly improves state-of-the-art defense against MI attacks, with minimal impact on testing accuracy.

Supplementary Material

MP4 File (codaspy-fp527.mp4)
This video is to present our paper "membership inference attacks and defenses in classification models". In this talk, we evaluated membership inference attacks in the literature and discuss what are the key features that make membership inference attacks successful. Based on the discussion, we proposed a new defense to reduce the generalization gap by reducing the training accuracy, and to match the output distribution of the training set with that of instances not used in the training. We summarized defenses in the literature and compared our defense with existing defenses. The experiments show that our defense is currently the most robust defense against existing attacks.

References

[1]
Martín Abadi, Andy Chu, Ian Goodfellow, H Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang. 2016. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security. ACM, 308--318.
[2]
Michael Backes, Pascal Berrang, Mathias Humbert, and Praveen Manoharan. 2016. Membership privacy in MicroRNA-based studies. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security. ACM, 319--330.
[3]
Karsten M Borgwardt, Arthur Gretton, Malte J Rasch, Hans-Peter Kriegel, Bernhard Schölkopf, and Alex J Smola. 2006. Integrating structured biological data by kernel maximum mean discrepancy. Bioinformatics, Vol. 22, 14 (2006), e49--e57.
[4]
Nicholas Carlini, Chang Liu, Jernej Kos, Ulfar Erlingsson, and Dawn Song. 2018. The Secret Sharer: Measuring Unintended Neural Network Memorization & Extracting Secrets. ArXiv e-prints, Vol. 1802.08232 (2018). https://arxiv.org/abs/1802.08232
[5]
Cynthia Dwork. 2006. Differential privacy. In ICALP. 1--12.
[6]
Cynthia Dwork. 2008. Differential privacy: A survey of results. In International Conference on Theory and Applications of Models of Computation. Springer, 1--19.
[7]
Cynthia Dwork, Frank McSherry, Kobbi Nissim, and Adam Smith. 2006. Calibrating noise to sensitivity in private data analysis. In Theory of cryptography conference. Springer, 265--284.
[8]
Robert Fortet and Edith Mourier. 1953. Convergence de la répartition empirique vers la répartition théorique. Annales scientifiques de l'École Normale Supérieure, Vol. 70, 3 (1953), 267--285.
[9]
Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in neural information processing systems. 2672--2680.
[10]
Arthur Gretton, Karsten M Borgwardt, Malte J Rasch, Bernhard Schölkopf, and Alexander Smola. 2012. A kernel two-sample test. Journal of Machine Learning Research, Vol. 13, Mar (2012), 723--773.
[11]
Jamie Hayes, Luca Melis, George Danezis, and Emiliano De Cristofaro. 2019. LOGAN: Membership inference attacks against generative models. Proceedings on Privacy Enhancing Technologies, Vol. 2019, 1 (2019), 133--152.
[12]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep Residual Learning for Image Recognition. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[13]
Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015).
[14]
Nils Homer, Szabolcs Szelinger, Margot Redman, David Duggan, Waibhav Tembe, Jill Muehling, John V Pearson, Dietrich A Stephan, Stanley F Nelson, and David W Craig. 2008. Resolving individuals contributing trace amounts of DNA to highly complex mixtures using high-density SNP genotyping microarrays. PLoS genetics, Vol. 4, 8 (2008), e1000167.
[15]
Gao Huang, Zhuang Liu, Laurens van der Maaten, and Kilian Q. Weinberger. 2017. Densely Connected Convolutional Networks. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[16]
Jinyuan Jia, Ahmed Salem, Michael Backes, Yang Zhang, and Neil Zhenqiang Gong. 2019. MemGuard: Defending against Black-Box Membership Inference Attacks via Adversarial Examples. In Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security. 259--274.
[17]
Yigitcan Kaya, Sanghyun Hong, and Tudor Dumitras. 2020. On the Effectiveness of Regularization Against Membership Inference Attacks. arXiv preprint arXiv:2006.05336 (2020).
[18]
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems. 1097--1105.
[19]
Ninghui Li, Wahbeh Qardaji, Dong Su, Yi Wu, and Weining Yang. 2013. Membership Privacy: A Unifying Framework for Privacy Definitions. In Proceedings of the 2013 ACM SIGSAC Conference on Computer & Communications Security (Berlin, Germany) (CCS '13). ACM, New York, NY, USA, 889--900. https://doi.org/10.1145/2508859.2516686
[20]
Mingsheng Long, Han Zhu, Jianmin Wang, and Michael I Jordan. 2017b. Deep transfer learning with joint adaptation networks. In Proceedings of the 34th International Conference on Machine Learning-Volume 70. JMLR. org, 2208--2217.
[21]
Yunhui Long, Vincent Bindschaedler, and Carl A Gunter. 2017a. Towards measuring membership privacy. arXiv preprint arXiv:1712.09136 (2017).
[22]
Yunhui Long, Vincent Bindschaedler, Lei Wang, Diyue Bu, Xiaofeng Wang, Haixu Tang, Carl A Gunter, and Kai Chen. 2018. Understanding membership inferences on well-generalized learning models. arXiv preprint arXiv:1802.04889 (2018).
[23]
Milad Nasr, Reza Shokri, and Amir Houmansadr. 2018. Machine learning with membership privacy using adversarial regularization. In Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security. ACM, 634--646.
[24]
Milad Nasr, Reza Shokri, and Amir Houmansadr. 2019. Comprehensive Privacy Analysis of Deep Learning: Stand-alone and Federated Learning under Passive and Active White-box Inference Attacks. In 2019 IEEE Symposium on Security and Privacy (SP).
[25]
Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. 2017. Automatic differentiation in PyTorch. In NIPS-W.
[26]
Ahmed Salem, Yang Zhang, Mathias Humbert, Mario Fritz, and Michael Backes. 2019. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models. In 25th Annual Network and Distributed System Security Symposium (NDSS).
[27]
Virat Shejwalkar and Amir Houmansadr. 2019. Reconciling Utility and Membership Privacy via Knowledge Distillation. arXiv preprint arXiv:1906.06589 (2019).
[28]
Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. 2017. Membership inference attacks against machine learning models. In Security and Privacy (SP), 2017 IEEE Symposium on. IEEE, 3--18.
[29]
K. Simonyan and A. Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. In International Conference on Learning Representations.
[30]
Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research, Vol. 15, 1 (2014), 1929--1958.
[31]
S. Truex, L. Liu, M. E. Gursoy, L. Yu, and W. Wei. 2019. Demystifying Membership Inference Attacks in Machine Learning as a Service. IEEE Transactions on Services Computing (2019), 1--1.
[32]
Jindong Wang et almbox. [n.d.]. Everything about Transfer Learning and Domain Adapation. http://transferlearning.xyz.
[33]
Rui Wang, Yong Fuga Li, XiaoFeng Wang, Haixu Tang, and Xiaoyong Zhou. 2009. Learning your identity and disease from research papers: information leaks in genome wide association study. In CCS. 534--544.
[34]
Samuel Yeom, Irene Giacomelli, Matt Fredrikson, and Somesh Jha. 2018. Privacy risk in machine learning: Analyzing the connection to overfitting. In 2018 IEEE 31st Computer Security Foundations Symposium (CSF). IEEE, 268--282.
[35]
Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin, and David Lopez-Paz. 2018. mixup: Beyond Empirical Risk Minimization. International Conference on Learning Representations (2018).

Cited By

View all
  • (2024)Preserving Privacy in GANs Against Membership Inference AttackIEEE Transactions on Information Forensics and Security10.1109/TIFS.2023.334265419(1728-1743)Online publication date: 2024
  • (2024)A Comprehensive Survey on Backdoor Attacks and Their Defenses in Face Recognition SystemsIEEE Access10.1109/ACCESS.2024.338258412(47433-47468)Online publication date: 2024
  • (2024)Dual Defense: Combining Preemptive Exclusion of Members and Knowledge Distillation to Mitigate Membership Inference AttacksJournal of Information and Intelligence10.1016/j.jiixd.2024.06.002Online publication date: Jun-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
CODASPY '21: Proceedings of the Eleventh ACM Conference on Data and Application Security and Privacy
April 2021
348 pages
ISBN:9781450381437
DOI:10.1145/3422337
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 26 April 2021

Permissions

Request permissions for this article.

Check for updates

Badges

  • Best Paper

Author Tags

  1. image classification
  2. membership inference
  3. neural networks

Qualifiers

  • Research-article

Funding Sources

  • United States National Science Foundation

Conference

CODASPY '21
Sponsor:

Acceptance Rates

Overall Acceptance Rate 149 of 789 submissions, 19%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)436
  • Downloads (Last 6 weeks)42
Reflects downloads up to 18 Sep 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Preserving Privacy in GANs Against Membership Inference AttackIEEE Transactions on Information Forensics and Security10.1109/TIFS.2023.334265419(1728-1743)Online publication date: 2024
  • (2024)A Comprehensive Survey on Backdoor Attacks and Their Defenses in Face Recognition SystemsIEEE Access10.1109/ACCESS.2024.338258412(47433-47468)Online publication date: 2024
  • (2024)Dual Defense: Combining Preemptive Exclusion of Members and Knowledge Distillation to Mitigate Membership Inference AttacksJournal of Information and Intelligence10.1016/j.jiixd.2024.06.002Online publication date: Jun-2024
  • (2024)Benchmarking robustness and privacy-preserving methods in federated learningFuture Generation Computer Systems10.1016/j.future.2024.01.009155:C(18-38)Online publication date: 1-Jun-2024
  • (2024)Re-ID-leak: Membership Inference Attacks Against Person Re-identificationInternational Journal of Computer Vision10.1007/s11263-024-02115-6Online publication date: 22-May-2024
  • (2024)Technical and Legal Aspects Relating to the (Re)Use of Health Data When Repurposing Machine Learning Models in the EUPrivacy Symposium 202310.1007/978-3-031-44939-0_3(33-48)Online publication date: 4-Jan-2024
  • (2023)FACE-AUDITORProceedings of the 32nd USENIX Conference on Security Symposium10.5555/3620237.3620640(7195-7212)Online publication date: 9-Aug-2023
  • (2023)Deep Neural Network Quantization Framework for Effective Defense against Membership Inference AttacksSensors10.3390/s2318772223:18(7722)Online publication date: 7-Sep-2023
  • (2023)Local Differential Privacy Based Membership-Privacy-Preserving Federated Learning for Deep-Learning-Driven Remote SensingRemote Sensing10.3390/rs1520505015:20(5050)Online publication date: 20-Oct-2023
  • (2023)RedactorProceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence and Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence and Thirteenth Symposium on Educational Advances in Artificial Intelligence10.1609/aaai.v37i12.26737(14874-14882)Online publication date: 7-Feb-2023
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media