Robust High-Dimensional Low-Rank Matrix Estimation: Optimal Rate and Data-Adaptive Tuning

被引:0
|
作者
Cui, Xiaolong [1 ]
Shi, Lei [2 ]
Zhong, Wei [3 ,4 ]
Zou, Changliang [5 ]
机构
[1] Nankai Univ, Sch Stat & Data Sci, Tianjin, Peoples R China
[2] Univ Calif Berkeley, Dept Biostat, Berkeley, CA USA
[3] Xiamen Univ, WISE, Xiamen, Peoples R China
[4] Xiamen Univ, Dept Stat & Data Sci, SOE, Xiamen, Peoples R China
[5] Nankai Univ, Sch Stat & Data Sci, LPMC KLMDASR & LEBPS, Tianjin, Peoples R China
关键词
heavy-tailed error; high dimension; low-rank matrix; non-asymptotic bounds; robustness; tuning parameter selection; PROXIMAL GRADIENT ALGORITHM; REGRESSION; COMPLETION; SELECTION; RECOVERY; MINIMIZATION; CONVERGENCE;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The matrix lasso, which minimizes a least-squared loss function with the nuclear-norm regularization, offers a generally applicable paradigm for high-dimensional low-rank matrix estimation, but its efficiency is adversely affected by heavy-tailed distributions. This paper introduces a robust procedure by incorporating a Wilcoxon-type rank-based loss function with the nuclear-norm penalty for a unified high-dimensional low-rank matrix estimation framework. It includes matrix regression, multivariate regression and matrix completion as special examples. This procedure enjoys several appealing features. First, it relaxes the distributional conditions on random errors from sub-exponential or sub-Gaussian to more general distributions and thus it is robust with substantial efficiency gain for heavy-tailed random errors. Second, as the gradient function of the rank-based loss function is completely pivotal, it overcomes the challenge of tuning parameter selection and substantially saves the computation time by using an easily simulated tuning parameter. Third, we theoretically establish non-asymptotic error bounds with a nearly-oracle rate for the new estimator. Numerical results indicate that the new estimator can be highly competitive among existing methods, especially for heavy-tailed or skewed errors.
引用
收藏
页数:57
相关论文
共 50 条
  • [1] Reconstruction of a high-dimensional low-rank matrix
    Yata, Kazuyoshi
    Aoshima, Makoto
    ELECTRONIC JOURNAL OF STATISTICS, 2016, 10 (01): : 895 - 917
  • [2] ESTIMATION OF HIGH-DIMENSIONAL LOW-RANK MATRICES
    Rohde, Angelika
    Tsybakov, Alexandre B.
    ANNALS OF STATISTICS, 2011, 39 (02): : 887 - 930
  • [3] High-dimensional covariance matrix estimation using a low-rank and diagonal decomposition
    Wu, Yilei
    Qin, Yingli
    Zhu, Mu
    CANADIAN JOURNAL OF STATISTICS-REVUE CANADIENNE DE STATISTIQUE, 2020, 48 (02): : 308 - 337
  • [4] A SHRINKAGE PRINCIPLE FOR HEAVY-TAILED DATA: HIGH-DIMENSIONAL ROBUST LOW-RANK MATRIX RECOVERY
    Fan, Jianqing
    Wang, Weichen
    Zhu, Ziwei
    ANNALS OF STATISTICS, 2021, 49 (03): : 1239 - 1266
  • [5] ROBUST LOW-RANK MATRIX ESTIMATION
    Elsener, Andreas
    van de Geer, Sara
    ANNALS OF STATISTICS, 2018, 46 (6B): : 3481 - 3509
  • [6] Robust Classification of High-Dimensional Data Using Data-Adaptive Energy Distance
    Choudhury, Jyotishka Ray
    Saha, Aytijhya
    Roy, Sarbojit
    Dutta, Subhajit
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT V, 2023, 14173 : 86 - 101
  • [7] Low-rank diffusion matrix estimation for high-dimensional time-changed Levy processes
    Belomestny, Denis
    Trabs, Mathias
    ANNALES DE L INSTITUT HENRI POINCARE-PROBABILITES ET STATISTIQUES, 2018, 54 (03): : 1583 - 1621
  • [8] ESTIMATION OF (NEAR) LOW-RANK MATRICES WITH NOISE AND HIGH-DIMENSIONAL SCALING
    Negahban, Sahand
    Wainwright, Martin J.
    ANNALS OF STATISTICS, 2011, 39 (02): : 1069 - 1097
  • [9] Robust low-rank data matrix approximations
    XingDong Feng
    XuMing He
    Science China Mathematics, 2017, 60 : 189 - 200
  • [10] Robust low-rank data matrix approximations
    Feng XingDong
    He XuMing
    SCIENCE CHINA-MATHEMATICS, 2017, 60 (02) : 189 - 200