pith. machine review for the scientific record. sign in

arxiv: 1704.03866 · v2 · submitted 2017-04-12 · 💻 cs.DS · cs.IT· cs.LG· math.IT· math.ST· stat.ML· stat.TH

Recognition: unknown

Robustly Learning a Gaussian: Getting Optimal Error, Efficiently

Authors on Pith no claims yet
classification 💻 cs.DS cs.ITcs.LGmath.ITmath.STstat.MLstat.TH
keywords erroroptimalpolynomialvarepsilonalgorithmsgaussianhigh-dimensionallearning
0
0 comments X
read the original abstract

We study the fundamental problem of learning the parameters of a high-dimensional Gaussian in the presence of noise -- where an $\varepsilon$-fraction of our samples were chosen by an adversary. We give robust estimators that achieve estimation error $O(\varepsilon)$ in the total variation distance, which is optimal up to a universal constant that is independent of the dimension. In the case where just the mean is unknown, our robustness guarantee is optimal up to a factor of $\sqrt{2}$ and the running time is polynomial in $d$ and $1/\epsilon$. When both the mean and covariance are unknown, the running time is polynomial in $d$ and quasipolynomial in $1/\varepsilon$. Moreover all of our algorithms require only a polynomial number of samples. Our work shows that the same sorts of error guarantees that were established over fifty years ago in the one-dimensional setting can also be achieved by efficient algorithms in high-dimensional settings.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.