A Unifying Framework of High-Dimensional Sparse Estimation with Difference-of-Convex (DC) Regularizations

被引:0
|
作者
Cao, Shanshan [1 ]
Huo, Xiaoming [1 ]
Pang, Jong-Shi [2 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Univ Southern Calif, Los Angeles, CA 90089 USA
关键词
(Generalized) linear regression; high-dimensional sparse estimation; nonconvex regularization; difference of convex (DC) functions; DC algorithms; asymptotic optimality; model selection consistency; NONCONCAVE PENALIZED LIKELIHOOD; CONFIDENCE-INTERVALS; VARIABLE SELECTION; MODEL SELECTION; ADAPTIVE LASSO; REGRESSION; OPTIMIZATION; CONVERGENCE; RELAXATION; OPTIMALITY;
D O I
10.1214/21-STS832
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Under the linear regression framework, we study the variable selection problem when the underlying model is assumed to have a small number of nonzero coefficients. Nonconvex penalties in specific forms are well studied in the literature for sparse estimation. Recent work pointed out that nearly all existing nonconvex penalties can be represented as differenceof-convex (DC) functions, which are the difference of two convex functions, while itself may not be convex. There is a large existing literature on optimization problems when their objectives and/or constraints involve DC functions. Efficient numerical solutions have been proposed. Under the DC framework, directional-stationary (d-stationary) solutions are considered, and they are usually not unique. In this paper, we show that under some mild conditions, a certain subset of d-stationary solutions in an optimization problem (with a DC objective) has some ideal statistical properties: namely, asymptotic estimation consistency, asymptotic model selection consistency, asymptotic efficiency. Our assumptions are either weaker than or comparable with those conditions that have been adopted in other existing works. This work shows that DC is a nice framework to offer a unified approach to these existing works where nonconvex penalties are involved. Our work bridges the communities of optimization and statistics.
引用
收藏
页码:411 / 424
页数:14
相关论文
共 50 条
  • [31] A Software Tool For Sparse Estimation Of A General Class Of High-dimensional GLMs
    Pazira, Hassan
    Augugliaro, Luigi
    Wit, Ernst C.
    R JOURNAL, 2022, 14 (01): : 34 - 53
  • [32] Efficient Minimax Estimation of a Class of High-Dimensional Sparse Precision Matrices
    Chen, Xiaohui
    Kim, Young-Heon
    Wang, Z. Jane
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2012, 60 (06) : 2899 - 2912
  • [33] A General Theory of Concave Regularization for High-Dimensional Sparse Estimation Problems
    Zhang, Cun-Hui
    Zhang, Tong
    STATISTICAL SCIENCE, 2012, 27 (04) : 576 - 593
  • [34] ACTIVE SET STRATEGY FOR HIGH-DIMENSIONAL NON-CONVEX SPARSE OPTIMIZATION PROBLEMS
    Boisbunon, Aurelie
    Flamary, Remi
    Rakotomamonjy, Alain
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [35] Sparse High-Dimensional Isotonic Regression
    Gamarnik, David
    Gaudio, Julia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [36] Classification of sparse high-dimensional vectors
    Ingster, Yuri I.
    Pouet, Christophe
    Tsybakov, Alexandre B.
    PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2009, 367 (1906): : 4427 - 4448
  • [37] High-dimensional sparse Fourier algorithms
    Bosu Choi
    Andrew Christlieb
    Yang Wang
    Numerical Algorithms, 2021, 87 : 161 - 186
  • [38] High-Dimensional Computing with Sparse Vectors
    Laiho, Mika
    Poikonen, Jussi H.
    Kanerva, Pentti
    Lehtonen, Eero
    2015 IEEE BIOMEDICAL CIRCUITS AND SYSTEMS CONFERENCE (BIOCAS), 2015, : 515 - 518
  • [39] On the anonymization of sparse high-dimensional data
    Ghinita, Gabriel
    Tao, Yufei
    Kalnis, Panos
    2008 IEEE 24TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING, VOLS 1-3, 2008, : 715 - +
  • [40] The sparse structure of high-dimensional integrands
    Verlinden, P
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2001, 132 (01) : 33 - 49