Statistical Query Lower Bounds for Robust Estimation of High-Dimensional Gaussians and Gaussian Mixtures (Extended Abstract)

被引:126
|
作者
Diakonikolas, Ilias [1 ]
Kane, Daniel M. [2 ]
Stewart, Alistair [1 ]
机构
[1] Univ Southern Calif, CS, Los Angeles, CA 90007 USA
[2] Univ Calif San Diego, CSE & Math, Los Angeles, CA USA
基金
美国国家科学基金会;
关键词
unsupervised learning; statistical learning; statistical queries; robust algorithm; LEARNING MIXTURES; PRINCIPAL-COMPONENTS; DISTRIBUTIONS; COMPLEXITY; PCA;
D O I
10.1109/FOCS.2017.16
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
We describe a general technique that yields the first Statistical Query lower bounds for a range of fundamental high-dimensional learning problems involving Gaussian distributions. Our main results are for the problems of (1) learning Gaussian mixture models (GMMs), and (2) robust (agnostic) learning of a single unknown Gaussian distribution. For each of these problems, we show a super-polynomial gap between the (information-theoretic) sample complexity and the computational complexity of any Statistical Query algorithm for the problem. Statistical Query (SQ) algorithms are a class of algorithms that are only allowed to query expectations of functions of the distribution rather than directly access samples. This class of algorithms is quite broad: a wide range of known algorithmic techniques in machine learning are known to be implementable using SQs. Moreover, for the unsupervised learning problems studied in this paper, all known algorithms with non-trivial performance guarantees are SQ or are easily implementable using SQs. Our SQ lower bound for Problem (1) is qualitatively matched by known learning algorithms for GMMs. At a conceptual level, this result implies that-as far as SQ algorithms are concerned-the computational complexity of learning GMMs is inherently exponential in the dimension of the latent space-even though there is no such information-theoretic barrier. Our lower bound for Problem (2) implies that the accuracy of the robust learning algorithm in [29] is essentially best possible among all polynomial-time SQ algorithms. On the positive side, we also give a new (SQ) learning algorithm for Problem (2) achieving the information-theoretically optimal accuracy, up to a constant factor, whose running time essentially matches our lower bound. Our algorithm relies on a filtering technique generalizing [29] that removes outliers based on higher-order tensors. Our SQ lower bounds are attained via a unified moment-matching technique that is useful in other contexts and may be of broader interest. Our technique yields nearly-tight lower bounds for a number of related unsupervised estimation problems. Specifically, for the problems of (3) robust covariance estimation in spectral norm, and (4) robust sparse mean estimation, we establish a quadratic statistical-computational tradeoff for SQ algorithms, matching known upper bounds. Finally, our technique can be used to obtain tight sample complexity lower bounds for high-dimensional testing problems. Specifically, for the classical problem of robustly testing an unknown mean (known covariance) Gaussian, our technique implies an information-theoretic sample lower bound that scales linearly in the dimension. Our sample lower bound matches the sample complexity of the corresponding robust learning problem and separates the sample complexity of robust testing from standard (non-robust) testing. This separation is surprising because such a gap does not exist for the corresponding learning problem.
引用
收藏
页码:73 / 84
页数:12
相关论文
共 50 条
  • [1] OPTIMAL ESTIMATION OF HIGH-DIMENSIONAL GAUSSIAN LOCATION MIXTURES
    Doss, Natalie
    Wu, Yihong
    Yang, Pengkun
    Zhou, Harrison H.
    ANNALS OF STATISTICS, 2023, 51 (01): : 62 - 95
  • [2] HIGH-DIMENSIONAL GAUSSIAN COPULA REGRESSION: ADAPTIVE ESTIMATION AND STATISTICAL INFERENCE
    Cai, T. Tony
    Zhang, Linjun
    STATISTICA SINICA, 2018, 28 (02) : 963 - 993
  • [3] Unified lower bounds for interactive high-dimensional estimation under information constraints
    Acharya, Jayadev
    Canonne, Clement L.
    Sun, Ziteng
    Tyagi, Himanshu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [4] An algorithmic argument for nonadaptive query complexity lower bounds on advised quantum computation (Extended abstract)
    Nishimura, H
    Yamakami, T
    MATHEMATICAL FOUNDATIONS OF COMPUTER SCIENCE 2004, PROCEEDINGS, 2004, 3153 : 827 - 838
  • [5] Interpolating discriminant functions in high-dimensional Gaussian latent mixtures
    Bing, Xin
    Wegkamp, Marten
    BIOMETRIKA, 2024, 111 (01) : 291 - 308
  • [6] Distributed Statistical Estimation of High-Dimensional and Nonparametric Distributions
    Han, Yanjun
    Mukherjee, Pritam
    Ozgur, Ayfer
    Weissman, Tsachy
    2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 506 - 510
  • [7] Robust estimation of a high-dimensional integrated covariance matrix
    Morimoto, Takayuki
    Nagata, Shuichi
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2017, 46 (02) : 1102 - 1112
  • [8] A Robust High-Dimensional Estimation of Multinomial Mixture Models
    Azam Sabbaghi
    Farzad Eskandari
    Hamid Reza Navabpoor
    Journal of Statistical Theory and Applications, 2021, 20 : 21 - 32
  • [9] Robust estimation of high-dimensional covariance and precision matrices
    Avella-Medina, Marco
    Battey, Heather S.
    Fan, Jianqing
    Li, Quefeng
    BIOMETRIKA, 2018, 105 (02) : 271 - 284
  • [10] Robust Estimation of High-Dimensional Linear Regression With Changepoints
    Cui, Xiaolong
    Geng, Haoyu
    Wang, Zhaojun
    Zou, Changliang
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2024, 70 (10) : 7297 - 7319