Nonasymptotic support recovery for high-dimensional sparse covariance matrices

被引:2
|
作者
Kashlak, Adam B. [1 ]
Kong, Linglong [1 ]
机构
[1] Univ Alberta, Math & Stat Sci, Edmonton, AB T6G 2G1, Canada
来源
STAT | 2021年 / 10卷 / 01期
关键词
concentration inequality; genomics; random matrix; Schatten norm; REGULARIZATION; ESTIMATORS;
D O I
10.1002/sta4.316
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
For high-dimensional data, the standard empirical estimator for the covariance matrix is very poor, and thus many methods have been proposed to more accurately estimate the covariance structure of high-dimensional data. In this article, we consider estimation under the assumption of sparsity but regularize with respect to the individual false-positive rate for incorrectly including a matrix entry in the support of the final estimator. The two benefits of this approach are (1) an interpretable regularization parameter removing the need for computationally expensive tuning and (2) extremely fast computation time arising from use of a binary search algorithm implemented to find the best estimator within a carefully constructed operator norm ball. We compare our approach to universal thresholding estimators and lasso-style penalized estimators on both simulated data and data from gene expression for cancerous tumours.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] SPARSE HIGH-DIMENSIONAL VARYING COEFFICIENT MODEL: NONASYMPTOTIC MINIMAX STUDY
    Klopp, Olga
    Pensky, Marianna
    ANNALS OF STATISTICS, 2015, 43 (03): : 1273 - 1299
  • [22] Sparse estimation of high-dimensional correlation matrices
    Cui, Ying
    Leng, Chenlei
    Sun, Defeng
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2016, 93 : 390 - 403
  • [23] Sparse covariance matrix estimation in high-dimensional deconvolution
    Belomestny, Denis
    Trabs, Mathias
    Tsybakov, Alexandre B.
    BERNOULLI, 2019, 25 (03) : 1901 - 1938
  • [24] A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data
    Kshitij Khare
    Sang-Yun Oh
    Syed Rahman
    Bala Rajaratnam
    Machine Learning, 2019, 108 : 2061 - 2086
  • [25] SPARSE COVARIANCE THRESHOLDING FOR HIGH-DIMENSIONAL VARIABLE SELECTION
    Jeng, X. Jessie
    Daye, Z. John
    STATISTICA SINICA, 2011, 21 (02) : 625 - 657
  • [26] A scalable sparse Cholesky based approach for learning high-dimensional covariance matrices in ordered data
    Khare, Kshitij
    Oh, Sang-Yun
    Rahman, Syed
    Rajaratnam, Bala
    MACHINE LEARNING, 2019, 108 (12) : 2061 - 2086
  • [27] Forecasting High-Dimensional Covariance Matrices Using High-Dimensional Principal Component Analysis
    Shigemoto, Hideto
    Morimoto, Takayuki
    AXIOMS, 2022, 11 (12)
  • [28] HIGH-DIMENSIONAL SPARSE COVARIANCE ESTIMATION FOR RANDOM SIGNALS
    Nasif, Ahmed O.
    Tian, Zhi
    Ling, Qing
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 4658 - 4662
  • [29] A Class of Structured High-Dimensional Dynamic Covariance Matrices
    Yang, Jin
    Lian, Heng
    Zhang, Wenyang
    COMMUNICATIONS IN MATHEMATICS AND STATISTICS, 2023,
  • [30] A robust test for sphericity of high-dimensional covariance matrices
    Tian, Xintao
    Lu, Yuting
    Li, Weiming
    JOURNAL OF MULTIVARIATE ANALYSIS, 2015, 141 : 217 - 227