Conference
Nearly tight sample complexity bounds for learning mixtures of Gaussians via sample compression schemes
Abstract
We prove that Θ(e kd2/ε2) samples are necessary and sufficient for learning a mixture of k Gaussians in Rd, up to error ε in total variation distance. This improves both the known upper bounds and lower bounds for this problem. For mixtures of axis-aligned Gaussians, we show that Oe(kd/ε2) samples suffice, matching a known lower bound. The upper bound is based on a novel technique for distribution learning based on a notion of sample …
Authors
Ashtiani H; Ben-David S; Harvey NJA; Liaw C; Mehrabian A; Plan Y
Volume
2018-December
Pagination
pp. 3412-3421
Publication Date
January 1, 2018
Conference proceedings
Advances in Neural Information Processing Systems
ISSN
1049-5258