skip to main content
10.1145/2020408.2020598acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
poster

Personal privacy vs population privacy: learning to attack anonymization

Published: 21 August 2011 Publication History

Abstract

Over the last decade great strides have been made in developing techniques to compute functions privately. In particular, Differential Privacy gives strong promises about conclusions that can be drawn about an individual. In contrast, various syntactic methods for providing privacy (criteria such as k-anonymity and l-diversity) have been criticized for still allowing private information of an individual to be inferred. In this paper, we consider the ability of an attacker to use data meeting privacy definitions to build an accurate classifier. We demonstrate that even under Differential Privacy, such classifiers can be used to infer "private" attributes accurately in realistic data. We compare this to similar approaches for inference-based attacks on other forms of anonymized data. We show how the efficacy of all these attacks can be measured on the same scale, based on the probability of successfully inferring a private attribute. We observe that the accuracy of inference of private attributes for differentially private data and $l$-diverse data can be quite similar.

References

[1]
A. Asuncion and D. Newman. UCI ML repository, 2007.
[2]
B. Barak, K. Chaudhuri, C. Dwork, S. Kale, F. McSherry, and K. Talwar. Privacy, accuracy, and consistency too: a holistic solution to contingency table release. In ACM Principles of Database Systems, 2007.
[3]
J. Brickell and V. Shmatikov. The cost of privacy: Destruction of data-mining utility in anonymized data publishing. In ACM SIGKDD, 2008.
[4]
G. Cormode, N. Li, T. Li, and D. Srivastava. Minimizing minimality and maximizing utility: Analyzing method-based attacks on anonymized data. In International Conference on Very Large Data Bases, 2010.
[5]
T. Dalenius. Towards a methodology for statistical disclosure control. Statistik Tidskrift, 15:429--444, 1977.
[6]
C. Dwork. Differential privacy. In ICALP, pages 1--12, 2006.
[7]
C. Dwork. Differential privacy: A survey of results. In Theory and Applications of Models of Computation, 2008.
[8]
A. Friedman and A. Schuster. Data mining with differential privacy. In ACM SIGKDD, 2010.
[9]
S. R. Ganta, S. P. Kasiviswanathan, and A. Smith. Composition attacks and auxiliary information in data privacy. In ACM SIGKDD, 2008.
[10]
A. Ghosh, T. Roughgarden, and M. Sundararajan. Universally utility-maximizing privacy mechanisms. In ACM Symposium on Theory of Computing, 2009.
[11]
P. Golle. Revisiting the uniqueness of simple demographics in the us population. In Workshop On Privacy In The Electronic Society, 2006.
[12]
S. P. Kasiviswanathan, H. K. Lee, K. Nissim, S. Raskhodnikova, and A. Smith. What can we learn privately? In IEEE Conference on Foundations of Computer Science, 2008.
[13]
S. P. Kasiviswanathan, M. Rudelson, A. Smith, and J. Ullman. The price of privately releasing contingency tables and the spectra of random matrices with correlated rows. In ACM Symposium on Theory of Computing, 2010.
[14]
D. Kifer. Attacks on privacy and de Finetti's theorem. In ACM SIGMOD International Conference on Management of Data, 2009.
[15]
D. Kifer and A. Machanavajjhala. No free lunch in data privacy. In ACM SIGMOD International Conference on Management of Data, 2011.
[16]
N. Li, T. Li, and S. Venkatasubramanian. t-closeness: Privacy beyond k-anonymity and $l$-diversity. In IEEE International Conference on Data Engineering, 2007.
[17]
A. Machanavajjhala, J. Gehrke, D. Kifer, and M. Venkitasubramaniam. l-diversity: Privacy beyond k-anonymity. In IEEE International Conference on Data Engineering, 2006.
[18]
F. McSherry and I. Mironov. Differentially private recommender systems: Building privacy into the netflix prize contenders. In ACM SIGKDD, 2009.
[19]
P. Samarati and L. Sweeney. Protecting privacy when disclosing information: k-anonymity and its enforcement through generalization and suppression. Technical Report SRI-CSL-98-04, SRI, 1998.
[20]
L. Sweeney. k-anonymity: a model for protecting privacy. International Journal on Uncertainty, Fuzziness and Knowledge-based systems, 10(5):557--570, 2002.
[21]
R. C.-W. Wong, A. W.-C. Fu, K. Wang, and J. Pei. Minimality attack in privacy preserving data publishing. In International Conference on Very Large Data Bases, pages 543--554, 2007.
[22]
X. Xiao and Y. Tao. Anatomy: Simple and effective privacy preservation. In International Conference on Very Large Data Bases, 2006.

Cited By

View all
  • (2024)On Data Distribution Leakage in Cross-Silo Federated LearningIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2023.334932336:7(3312-3328)Online publication date: 3-Jan-2024
  • (2024)Practical Attribute Reconstruction Attack Against Federated LearningIEEE Transactions on Big Data10.1109/TBDATA.2022.315923610:6(851-863)Online publication date: Dec-2024
  • (2024)Unraveling Attacks to Machine-Learning-Based IoT Systems: A Survey and the Open Libraries Behind ThemIEEE Internet of Things Journal10.1109/JIOT.2024.337773011:11(19232-19255)Online publication date: 1-Jun-2024
  • Show More Cited By

Index Terms

  1. Personal privacy vs population privacy: learning to attack anonymization

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    KDD '11: Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining
    August 2011
    1446 pages
    ISBN:9781450308137
    DOI:10.1145/2020408
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 21 August 2011

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. anonymization
    2. differential privacy

    Qualifiers

    • Poster

    Conference

    KDD '11
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

    Upcoming Conference

    KDD '25

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)42
    • Downloads (Last 6 weeks)6
    Reflects downloads up to 07 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)On Data Distribution Leakage in Cross-Silo Federated LearningIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2023.334932336:7(3312-3328)Online publication date: 3-Jan-2024
    • (2024)Practical Attribute Reconstruction Attack Against Federated LearningIEEE Transactions on Big Data10.1109/TBDATA.2022.315923610:6(851-863)Online publication date: Dec-2024
    • (2024)Unraveling Attacks to Machine-Learning-Based IoT Systems: A Survey and the Open Libraries Behind ThemIEEE Internet of Things Journal10.1109/JIOT.2024.337773011:11(19232-19255)Online publication date: 1-Jun-2024
    • (2024)An overview of proposals towards the privacy-preserving publication of trajectory dataInternational Journal of Information Security10.1007/s10207-024-00894-023:6(3711-3747)Online publication date: 4-Sep-2024
    • (2024)On the Alignment of Group Fairness with Attribute PrivacyWeb Information Systems Engineering – WISE 202410.1007/978-981-96-0567-5_24(333-348)Online publication date: 3-Dec-2024
    • (2024)Evaluating Differential Privacy on Correlated Datasets Using Pointwise Maximal LeakagePrivacy Technologies and Policy10.1007/978-3-031-68024-3_4(73-86)Online publication date: 1-Aug-2024
    • (2023)Synthesizing Realistic Trajectory Data With Differential PrivacyIEEE Transactions on Intelligent Transportation Systems10.1109/TITS.2023.324129024:5(5502-5515)Online publication date: May-2023
    • (2023)Toward Compliance Implications and Security Objectives: A Qualitative Study2023 IEEE 39th International Conference on Data Engineering Workshops (ICDEW)10.1109/ICDEW58674.2023.00028(138-145)Online publication date: Apr-2023
    • (2023)Towards Measuring Fairness for Local Differential PrivacyData Privacy Management, Cryptocurrencies and Blockchain Technology10.1007/978-3-031-25734-6_2(19-34)Online publication date: 24-Feb-2023
    • (2022)A Novel Privacy Paradigm for Improving Serial Data PrivacySensors10.3390/s2207281122:7(2811)Online publication date: 6-Apr-2022
    • Show More Cited By

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media