Robustly Learning a Gaussian: Getting Optimal Error, Efficiently

We study the fundamental problem of learning the parameters of a high-dimensional Gaussian in the presence of noise | where an "-fraction of our samples were chosen by an adversary. We give robust estimators that achieve estimation error O(ϵ) in the total variation distance, which is optimal u...

Full description

Bibliographic Details
Main Authors: Stewart, Alistair, Diakonikolas, Ilias, Kamath, Gautam Chetan, Kane, Daniel M, Li, Jerry Zheng, Moitra, Ankur
Other Authors: Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
Format: Article
Published: Society for Industrial and Applied Mathematics 2018
Online Access:http://hdl.handle.net/1721.1/116214
https://orcid.org/0000-0003-0048-2559
https://orcid.org/0000-0002-9937-0049
https://orcid.org/0000-0001-7047-0495
_version_ 1826193440087474176
author Stewart, Alistair
Diakonikolas, Ilias
Kamath, Gautam Chetan
Kane, Daniel M
Li, Jerry Zheng
Moitra, Ankur
author2 Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
author_facet Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
Stewart, Alistair
Diakonikolas, Ilias
Kamath, Gautam Chetan
Kane, Daniel M
Li, Jerry Zheng
Moitra, Ankur
author_sort Stewart, Alistair
collection MIT
description We study the fundamental problem of learning the parameters of a high-dimensional Gaussian in the presence of noise | where an "-fraction of our samples were chosen by an adversary. We give robust estimators that achieve estimation error O(ϵ) in the total variation distance, which is optimal up to a universal constant that is independent of the dimension. In the case where just the mean is unknown, our robustness guarantee is optimal up to a factor of p 2 and the running time is polynomial in d and 1/ϵ. When both the mean and covariance are unknown, the running time is polynomial in d and quasipolynomial in 1/ϵ. Moreover all of our algorithms require only a polynomial number of samples. Our work shows that the same sorts of error guarantees that were established over fifty years ago in the one-dimensional setting can also be achieved by efficient algorithms in high-dimensional settings.
first_indexed 2024-09-23T09:39:16Z
format Article
id mit-1721.1/116214
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T09:39:16Z
publishDate 2018
publisher Society for Industrial and Applied Mathematics
record_format dspace
spelling mit-1721.1/1162142022-09-30T15:59:06Z Robustly Learning a Gaussian: Getting Optimal Error, Efficiently Stewart, Alistair Diakonikolas, Ilias Kamath, Gautam Chetan Kane, Daniel M Li, Jerry Zheng Moitra, Ankur Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology. Department of Mathematics Massachusetts Institute of Technology. Department of Physics Diakonikolas, Ilias Kamath, Gautam Chetan Kane, Daniel M Li, Jerry Zheng Moitra, Ankur We study the fundamental problem of learning the parameters of a high-dimensional Gaussian in the presence of noise | where an "-fraction of our samples were chosen by an adversary. We give robust estimators that achieve estimation error O(ϵ) in the total variation distance, which is optimal up to a universal constant that is independent of the dimension. In the case where just the mean is unknown, our robustness guarantee is optimal up to a factor of p 2 and the running time is polynomial in d and 1/ϵ. When both the mean and covariance are unknown, the running time is polynomial in d and quasipolynomial in 1/ϵ. Moreover all of our algorithms require only a polynomial number of samples. Our work shows that the same sorts of error guarantees that were established over fifty years ago in the one-dimensional setting can also be achieved by efficient algorithms in high-dimensional settings. 2018-06-11T17:27:24Z 2018-06-11T17:27:24Z 2018-01 2018-05-29T13:19:50Z Article http://purl.org/eprint/type/JournalArticle 0368-4245 http://hdl.handle.net/1721.1/116214 Diakonikolas, Ilias, Gautam Kamath, Daniel M. Kane, Jerry Li, Ankur Moitra, and Alistair Stewart. “Robustly Learning a Gaussian: Getting Optimal Error, Efficiently.” Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms (January 2018): 2683–2702. https://orcid.org/0000-0003-0048-2559 https://orcid.org/0000-0002-9937-0049 https://orcid.org/0000-0001-7047-0495 http://dx.doi.org/10.1137/1.9781611975031.171 Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use. application/pdf Society for Industrial and Applied Mathematics SIAM
spellingShingle Stewart, Alistair
Diakonikolas, Ilias
Kamath, Gautam Chetan
Kane, Daniel M
Li, Jerry Zheng
Moitra, Ankur
Robustly Learning a Gaussian: Getting Optimal Error, Efficiently
title Robustly Learning a Gaussian: Getting Optimal Error, Efficiently
title_full Robustly Learning a Gaussian: Getting Optimal Error, Efficiently
title_fullStr Robustly Learning a Gaussian: Getting Optimal Error, Efficiently
title_full_unstemmed Robustly Learning a Gaussian: Getting Optimal Error, Efficiently
title_short Robustly Learning a Gaussian: Getting Optimal Error, Efficiently
title_sort robustly learning a gaussian getting optimal error efficiently
url http://hdl.handle.net/1721.1/116214
https://orcid.org/0000-0003-0048-2559
https://orcid.org/0000-0002-9937-0049
https://orcid.org/0000-0001-7047-0495
work_keys_str_mv AT stewartalistair robustlylearningagaussiangettingoptimalerrorefficiently
AT diakonikolasilias robustlylearningagaussiangettingoptimalerrorefficiently
AT kamathgautamchetan robustlylearningagaussiangettingoptimalerrorefficiently
AT kanedanielm robustlylearningagaussiangettingoptimalerrorefficiently
AT lijerryzheng robustlylearningagaussiangettingoptimalerrorefficiently
AT moitraankur robustlylearningagaussiangettingoptimalerrorefficiently