Spotlight Poster
Sample-Efficient Private Learning of Mixtures of Gaussians
Hassan Ashtiani · Mahbod Majid · Shyam Narayanan
West Ballroom A-D #6202
Abstract:
We study the problem of learning mixtures of Gaussians with approximate differential privacy. We prove that roughly samples suffice to learn a mixture of arbitrary -dimensional Gaussians up to low total variation distance, with differential privacy. Our work improves over the previous best result (which required roughly samples) and is provably optimal when is much larger than . Moreover, we give the first optimal bound for privately learning mixtures of univariate (i.e., -dimensional) Gaussians. Importantly, we show that the sample complexity for learning mixtures of univariate Gaussians is linear in the number of components , whereas the previous best sample complexity was quadratic in . Our algorithms utilize various techniques, including the inverse sensitivity mechanism, sample compression for distributions, and methods for bounding volumes of sumsets.
Chat is not available.