l1-penalization for mixture regression models

被引:131
|
作者
Staedler, Nicolas [1 ]
Buehlmann, Peter [1 ]
van de Geer, Sara [1 ]
机构
[1] ETH, Seminar Stat, CH-8092 Zurich, Switzerland
关键词
Adaptive Lasso; Finite mixture models; Generalized EM algorithm; High-dimensional estimation; Lasso; Oracle inequality; VARIABLE SELECTION; DANTZIG SELECTOR; FINITE MIXTURES; ADAPTIVE LASSO; SPARSITY; RECOVERY; CONVERGENCE; LIKELIHOOD;
D O I
10.1007/s11749-010-0197-z
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We consider a finite mixture of regressions (FMR) model for high-dimensional inhomogeneous data where the number of covariates may be much larger than sample size. We propose an l(1)-penalized maximum likelihood estimator in an appropriate parameterization. This kind of estimation belongs to a class of problems where optimization and theory for non-convex functions is needed. This distinguishes itself very clearly from high-dimensional estimation with convex loss-or objective functions as, for example, with the Lasso in linear or generalized linear models. Mixture models represent a prime and important example where non-convexity arises. For FMR models, we develop an efficient EM algorithm for numerical optimization with provable convergence properties. Our penalized estimator is numerically better posed (e.g., boundedness of the criterion function) than unpenalized maximum likelihood estimation, and it allows for effective statistical regularization including variable selection. We also present some asymptotic theory and oracle inequalities: due to non-convexity of the negative log-likelihood function, different mathematical arguments are needed than for problems with convex losses. Finally, we apply the new method to both simulated and real data.
引用
收藏
页码:209 / 256
页数:48
相关论文
共 50 条
  • [1] Comments on: l1-penalization for mixture regression models
    Antoniadis, Anestis
    TEST, 2010, 19 (02) : 257 - 258
  • [2] Comment on: l1-penalization for mixture regression models
    Lugosi, Gabor
    TEST, 2010, 19 (02) : 259 - 263
  • [3] ℓ1-penalization for mixture regression models
    Nicolas Städler
    Peter Bühlmann
    Sara van de Geer
    TEST, 2010, 19 : 209 - 256
  • [4] Comments on: ℓ1-penalization for mixture regression models
    Jianqing Fan
    Jinchi Lv
    TEST, 2010, 19 : 264 - 269
  • [5] Comment on: ℓ1-penalization for mixture regression models
    Gábor Lugosi
    TEST, 2010, 19 : 259 - 263
  • [6] GLMMLasso: An Algorithm for High-Dimensional Generalized Linear Mixed Models Using l1-Penalization
    Schelldorfer, Juerg
    Meier, Lukas
    Buehlmann, Peter
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2014, 23 (02) : 460 - 477
  • [7] Estimation for High-Dimensional Linear Mixed-Effects Models Using l1-Penalization
    Schelldorfer, Juerg
    Buehlmann, Peter
    Van De Geer, Sara
    SCANDINAVIAN JOURNAL OF STATISTICS, 2011, 38 (02) : 197 - 214
  • [8] Sparse N-way partial least squares by L1-penalization
    Hervas, D.
    Prats-Montalban, J. M.
    Garcia-Canaveras, J. C.
    Lahoz, A.
    Ferrer, A.
    CHEMOMETRICS AND INTELLIGENT LABORATORY SYSTEMS, 2019, 185 : 85 - 91
  • [9] Adjusting for high-dimensional covariates in sparse precision matrix estimation by l1-penalization
    Yin, Jianxin
    Li, Hongzhe
    JOURNAL OF MULTIVARIATE ANALYSIS, 2013, 116 : 365 - 381
  • [10] Honest variable selection in linear and logistic regression models via l1 and l1 + l2 penalization
    Bunea, Florentina
    ELECTRONIC JOURNAL OF STATISTICS, 2008, 2 : 1153 - 1194