Attacks on statistical databases: The highly noisy case

Alexander Kantor, Kobbi Nissim

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

A formal investigation of the utility-privacy tradeoff in statistical databases has proved essential for the rigorous discussion of privacy of recent years. Initial results in this direction dealt with databases that answer (all) subset-sum queries to within some fixed distortion [Dinur and Nissim, PODC 2003]. Subsequent work has extended these results to the case where a constant portion of the queries are answered arbitrarily [Dwork, McSherry, and Talwar, STOC 2007], and furthermore to the case where up to almost half the queries are answered arbitrarily [Dwork and Yekhanin, CRYPTO 2008]. All these results demonstrate how an efficient attacker may learn the underlying database (exactly or approximately), and hence bear consequences to tasks such as private sanitization of data. We give the first efficient attack for the case where the queries that are answered within the fixed distortion form only a polynomially small fraction of the queries (the rest are answered arbitrarily). Our techniques borrow from program correction and learning in the presence of noise.

Original languageEnglish
Pages (from-to)409-413
Number of pages5
JournalInformation Processing Letters
Volume113
Issue number12
DOIs
StatePublished - 8 Apr 2013

Keywords

  • Privacy Statistical databases Databases Learning with noise

Fingerprint

Dive into the research topics of 'Attacks on statistical databases: The highly noisy case'. Together they form a unique fingerprint.

Cite this