Years and Authors of Summarized Original Work
-
2006; Cynthia Dwork, Frank McSherry, Kobbi Nissim, Adam Smith
-
2007; Kobbi Nissim, Sofya Raskhodnikova, Adam Smith
-
2009; Cynthia Dwork, Jing Lei
-
2013; Adam Smith, Abhradeep Thakurta
Problem Definition
Over the last few years, differential privacy [5, 6] has emerged as one of the most accepted notions of statistical data privacy. At a high level differential privacy ensures that from the output of an algorithm executed on a data set of potentially sensitive records, an adversary learns “almost” the same thing about an individual irrespective of his presence or absence in the data set. Formally, differential privacy is defined below (Definition 1). Setting the privacy parameters ε < 1 and \(\delta \ll \frac{1} {n^{2}}\) ensures semantically meaningful privacy guarantees. For a detailed survey on the semantics of differential privacy, see [2, 3, 8, 9].
Definition 1 ((ε, δ)-differential privacy [5, 6])
We call two data sets D and D′ (with n...
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Recommended Reading
Anthony M, Bartlett PL (2009) Neural network learning: theoretical foundations. Cambridge University Press, Cambridge
Dwork C (2006) Differential privacy. In: 33rd international colloquium on automata, languages and programming, Venice, LNCS pp 1–12
Dwork C (2011) A firm foundation for private data analysis. Commun ACM 54(1):86–95
Dwork C, Lei J (2009) Differential privacy and robust statistics. In: Symposium on theory of computing (STOC), Bethesda, pp 371–380
Dwork C, Kenthapadi K, Mcsherry F, Mironov I, Naor M (2006) Our data, ourselves: privacy via distributed noise generation. In: EUROCRYPT. Springer, pp 486–503
Dwork C, McSherry F, Nissim K, Smith A (2006) Calibrating noise to sensitivity in private data analysis. In: Theory of cryptography conference. Springer, New York, pp 265–284
Hardt M, Roth A (2013) Beyond worst-case analysis in private singular vector computation. In: STOC, Palo Alto
Kasiviswanathan SP, Smith A (2008) A note on differential privacy: defining resistance to arbitrary side information. CoRR arXiv:0803.39461 [cs.CR]
Kifer D, Machanavajjhala A (2012) A rigorous and customizable framework for privacy. In: Principles of database systems (PODS 2012), Scottsdale
Kifer D, Smith A, Thakurta A (2012) Private convex empirical risk minimization and high-dimensional regression. In: Conference on learning theory, Edinburgh, pp 25.1–25.40
Nissim K, Raskhodnikova S, Smith A (2007) Smooth sensitivity and sampling in private data analysis. In: Symposium on theory of computing (STOC), San Diego, ACM, pp 75–84. Full paper: http://www.cse.psu.edu/~asmith/pubs/NRS07
Smith A (2011) Privacy-preserving statistical estimation with optimal convergence rates. In: Proceedings of the forty-third annual ACM symposium on theory of computing, San Jose, pp 813–822
Smith AD, Thakurta A (2013) Differentially private model selection via stability arguments and the robustness of the lasso. J Mach Learn Res Proc Track 30:819–850
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer Science+Business Media New York
About this entry
Cite this entry
Thakurta, A. (2016). Beyond Worst Case Sensitivity in Private Data Analysis. In: Kao, MY. (eds) Encyclopedia of Algorithms. Springer, New York, NY. https://doi.org/10.1007/978-1-4939-2864-4_547
Download citation
DOI: https://doi.org/10.1007/978-1-4939-2864-4_547
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4939-2863-7
Online ISBN: 978-1-4939-2864-4
eBook Packages: Computer ScienceReference Module Computer Science and Engineering