Conference
On the Sample Complexity of Privately Learning Unbounded High-Dimensional Gaussians
Abstract
We provide sample complexity upper bounds for agnostically learning multivariate Gaussians under the constraint of approximate differential privacy. These are the first finite sample upper bounds for general Gaussians which do not impose restrictions on the parameters of the distribution. Our bounds are near-optimal in the case when the covariance is known to be the identity, and conjectured to be near-optimal in the general case. From a …
Authors
Aden-Ali I; Ashtiani H; Kamath G
Volume
132
Pagination
pp. 185-216
Publication Date
January 1, 2021
Conference proceedings
Proceedings of Machine Learning Research