List-Decodable Robust Mean Estimation and Learning Mixtures of Spherical Gaussians

11/20/2017
by   Ilias Diakonikolas, et al.
0

We study the problem of list-decodable Gaussian mean estimation and the related problem of learning mixtures of separated spherical Gaussians. We develop a set of techniques that yield new efficient algorithms with significantly improved guarantees for these problems. List-Decodable Mean Estimation. Fix any d ∈Z_+ and 0< α <1/2. We design an algorithm with runtime O (poly(n/α)^d) that outputs a list of O(1/α) many candidate vectors such that with high probability one of the candidates is within ℓ_2-distance O(α^-1/(2d)) from the true mean. The only previous algorithm for this problem achieved error Õ(α^-1/2) under second moment conditions. For d = O(1/ϵ), our algorithm runs in polynomial time and achieves error O(α^ϵ). We also give a Statistical Query lower bound suggesting that the complexity of our algorithm is qualitatively close to best possible. Learning Mixtures of Spherical Gaussians. We give a learning algorithm for mixtures of spherical Gaussians that succeeds under significantly weaker separation assumptions compared to prior work. For the prototypical case of a uniform mixture of k identity covariance Gaussians we obtain: For any ϵ>0, if the pairwise separation between the means is at least Ω(k^ϵ+√((1/δ))), our algorithm learns the unknown parameters within accuracy δ with sample complexity and running time poly (n, 1/δ, (k/ϵ)^1/ϵ). The previously best known polynomial time algorithm required separation at least k^1/4polylog(k/δ). Our main technical contribution is a new technique, using degree-d multivariate polynomials, to remove outliers from high-dimensional datasets where the majority of the points are corrupted.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset