Differential Privacy (DP) is the current gold-standard for ensuring privacy for statistical queries. Estimation problems under DP constraints appearing in the literature have largely focused on providing equal privacy to all users. We consider the problems of empirical mean estimation for univariate data and frequency estimation for categorical data, both subject to heterogeneous privacy constraints. Each user, contributing a sample to the dataset, is allowed to have a different privacy demand. The dataset itself is assumed to be worst-case and we study both problems under two different formulations - first, where privacy demands and data may be correlated, and second, where correlations are weakened by random permutation of the dataset. We establish theoretical performance guarantees for our proposed algorithms, under both PAC error and mean-squared error. These performance guarantees translate to minimax optimality in several instances, and experiments confirm superior performance of our algorithms over other baseline techniques.
@InProceedings{chaudhuri_et_al:LIPIcs.FORC.2025.3, author = {Chaudhuri, Syomantak and Courtade, Thomas A.}, title = {{Private Estimation When Data and Privacy Demands Are Correlated}}, booktitle = {6th Symposium on Foundations of Responsible Computing (FORC 2025)}, pages = {3:1--3:20}, series = {Leibniz International Proceedings in Informatics (LIPIcs)}, ISBN = {978-3-95977-367-6}, ISSN = {1868-8969}, year = {2025}, volume = {329}, editor = {Bun, Mark}, publisher = {Schloss Dagstuhl -- Leibniz-Zentrum f{\"u}r Informatik}, address = {Dagstuhl, Germany}, URL = {https://drops.dagstuhl.de/entities/document/10.4230/LIPIcs.FORC.2025.3}, URN = {urn:nbn:de:0030-drops-231305}, doi = {10.4230/LIPIcs.FORC.2025.3}, annote = {Keywords: Differential Privacy, Personalized Privacy, Heterogeneous Privacy, Correlations in Privacy} }
Feedback for Dagstuhl Publishing