A NON ASYMPTOTIC PENALIZED CRITERION FOR GAUSSIAN MIXTURE MODEL SELECTION

被引:30
|
作者
Maugis, Cathy [1 ]
Michel, Bertrand [2 ]
机构
[1] Univ Toulouse, Inst Math Toulouse, INSA Toulouse, F-31077 Toulouse 4, France
[2] Univ Paris 06, Lab Stat Theor & Appl, F-75013 Paris, France
关键词
Model-based clustering; variable selection; penalized likelihood criterion; bracketing entropy; MAXIMUM-LIKELIHOOD; CONVERGENCE; RATES;
D O I
10.1051/ps/2009004
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Specific Gaussian mixtures are considered to solve simultaneously variable selection and clustering problems. A non asymptotic penalized criterion is proposed to choose the number of mixture components and the relevant variable subset. Because of the non linearity of the associated Kullback-Leibler contrast on Gaussian mixtures, a general model selection theorem for maximum likelihood estimation proposed by [Massart Concentration inequalities and model selection Springer, Berlin (2007). Lectures from the 33rd Summer School on Probability Theory held in Saint-Flour, July 6-23 (2003)] is used to obtain the penalty function form. This theorem requires to control the bracketing entropy of Gaussian mixture families. The ordered and non-ordered variable selection cases are both addressed in this paper.
引用
收藏
页码:41 / 68
页数:28
相关论文
共 50 条
  • [31] Bayesian feature and model selection for Gaussian mixture models
    Constantinopoulos, C
    Titsias, MK
    Likas, A
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2006, 28 (06) : 1013 - U1
  • [32] Iterative feature selection in Gaussian mixture clustering with automatic model selection
    Zeng, Hong
    Cheung, Yiu-Ming
    2007 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-6, 2007, : 2277 - 2282
  • [33] A robust generalization and asymptotic properties of the model selection criterion family
    Kurata, Sumito
    Hamada, Etsuo
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2018, 47 (03) : 532 - 547
  • [34] Pleiotropic mapping and annotation selection in genome-wide association studies with penalized Gaussian mixture models
    Zeng, Ping
    Hao, Xingjie
    Zhou, Xiang
    BIOINFORMATICS, 2018, 34 (16) : 2797 - 2807
  • [35] A Cooperative and Penalized Competitive Learning Approach to Gaussian Mixture Clustering
    Cheung, Yiu-ming
    Jia, Hong
    ARTIFICIAL NEURAL NETWORKS (ICANN 2010), PT III, 2010, 6354 : 435 - 440
  • [36] Penalized multinomial mixture logit model
    Shaheena Bashir
    Edward M. Carter
    Computational Statistics, 2010, 25 : 121 - 141
  • [37] Penalized multinomial mixture logit model
    Bashir, Shaheena
    Carter, Edward M.
    COMPUTATIONAL STATISTICS, 2010, 25 (01) : 121 - 141
  • [38] ASYMPTOTIC THEORY OF GENERALIZED INFORMATION CRITERION FOR GEOSTATISTICAL REGRESSION MODEL SELECTION
    Chang, Chih-Hao
    Huang, Hsin-Cheng
    Ing, Ching-Kang
    ANNALS OF STATISTICS, 2014, 42 (06): : 2441 - 2468
  • [39] A non-asymptotic theory for model selection
    Massart, P
    European Congress of Mathematics, 2005, : 309 - 323
  • [40] Model selection via Bayesian information criterion for divide-and-conquer penalized quantile regression
    Kang, Jongkyeong
    Han, Seokwon
    Bang, Sungwan
    KOREAN JOURNAL OF APPLIED STATISTICS, 2022, 35 (02) : 217 - 227