Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1007/978-3-030-50309-3_32guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article

An Emerging Strategy for Privacy Preserving Databases: Differential Privacy

Published: 19 July 2020 Publication History

Abstract

Data De-identification and Differential Privacy are two possible approaches for providing data security and user privacy. Data de-identification is the process where the personal identifiable information of individuals is extracted to create anonymized databases. Data de-identification has been used for quite some time in industry to sanitize data before it is outsourced for data-mining purposes. Differential privacy attempts to protect sensitive data by adding an appropriate level of noise to the output of a query or to the primary database so that the presence or the absence of a single piece of information will not significantly alter the query output. Recent work in the literature has highlighted the risk of re-identification of information in a de-identified data set. In this paper, we provide a comprehensive comparison of these two privacy-preserving strategies. Our results show that the differentially private trained models produce highly accurate data, while preserving data privacy, making them a reliable alternative to the data de-identification models.

References

[1]
Dwork C, McSherry F, Nissim K, and Smith A Halevi S and Rabin T Calibrating noise to sensitivity in private data analysis Theory of Cryptography 2006 Heidelberg Springer 265-284
[2]
Dwork C and Pottenger R Toward practicing privacy J. Am. Med. Inform. Assoc. 2013 20 1 102-108
[3]
Dwork C, Roth A, et al. The algorithmic foundations of differential privacy Found. Trends® Theor. Comput. Sci. 2014 9 3–4 211-407
[4]
Holohan, N.: Welcome to the IBM differential privacy library. https://diffprivlib.readthedocs.io/en/latest/. Accessed 21 Dec 2019
[5]
Information, of Ontario, P.C.: De-identification guidelines for structured data. White Paper, pp. 1–28 (2016)
[6]
Jain P, Gyanchandani M, and Khare N Differential privacy: its technological prescriptive using big data J. Big Data 2018 5 1 1-24
[7]
Kim J and Winkler W Multiplicative noise for masking continuous data Statistics 2003 1 9
[8]
McSherry, F., Talwar, K.: Mechanism design via differential privacy. In: Proceedings of the 48th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2007), pp. 94–103. IEEE (2007)
[9]
McSherry, F.D.: Privacy integrated queries: an extensible platform for privacy-preserving data analysis. In: Proceedings of the 2009 ACM SIGMOD International Conference on Management of Data, pp. 19–30. ACM (2009)
[10]
Mohan, P., Thakurta, A., Shi, E., Song, D., Culler, D.: GUPT: privacy preserving data analysis made easy. In: Proceedings of the 2012 ACM SIGMOD International Conference on Management of Data, pp. 349–360. ACM (2012)
[11]
Office, N.: MIT and Harvard release de-identified learning data from open online courses. http://news.mit.edu/2014/mit-and-harvard-release-de-identified-learning-data-open-online-courses. Accessed 15 May 2019
[12]
Organization, H.: Hitrust de-identification framework. https://hitrustalliance.net/de-identification/. Accessed 30 Jan 2020
[13]
Registry, B.O., Spafford, O.: Requesting data. https://www.bornontario.ca/en/data/requesting-data.aspx. Accessed 12 Aug 2019
[14]
Rocher L, Hendrickx JM, and De Montjoye YA Estimating the success of re-identifications in incomplete datasets using generative models Nat. Commun. 2019 10 1 1-9
[15]
Roy, I., Setty, S.T., Kilzer, A., Shmatikov, V., Witchel, E.: Airavat: Security and privacy for mapreduce. In: Proceedings of the 7th USENIX Symposium on Networked Systems Design and Implementation, NSDI 2010, vol. 10, pp. 297–312 (2010)
[16]
Sarathy R and Muralidhar K Evaluating laplace noise addition to satisfy differential privacy for numeric data Trans. Data Priv. 2011 4 1 1-17
[17]
Spafford, K.: Will my personal information be safe? http://cpcssn.ca/faq-posts/will-my-personal-information-be-safe/. Accessed 10 Aug 2019
[18]
UCI: Center for machine learning and intelligent systems. https://cml.ics.uci.edu/. Accessed Aug 2020

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Guide Proceedings
HCI for Cybersecurity, Privacy and Trust: Second International Conference, HCI-CPT 2020, Held as Part of the 22nd HCI International Conference, HCII 2020, Copenhagen, Denmark, July 19–24, 2020, Proceedings
Jul 2020
695 pages
ISBN:978-3-030-50308-6
DOI:10.1007/978-3-030-50309-3
  • Editor:
  • Abbas Moallem

Publisher

Springer-Verlag

Berlin, Heidelberg

Publication History

Published: 19 July 2020

Author Tags

  1. Differential privacy
  2. Data De-identification
  3. Separate architecture
  4. Privacy preserving databases

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 0
    Total Downloads
  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 26 Sep 2024

Other Metrics

Citations

View Options

View options

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media