Robust Information Criterion for Model Selection in Sparse High-Dimensional Linear Regression Models

被引:3
|
作者
Gohain, Prakash Borpatra [1 ]
Jansson, Magnus [1 ]
机构
[1] KTH Royal Inst Technol, Div Informat Sci & Engn, SE-10044 Stockholm, Sweden
基金
欧洲研究理事会;
关键词
High-dimension; linear regression; data scaling; statistical model selection; subset selection; sparse estimation; scale-invariant; variable selection; CROSS-VALIDATION; MDL;
D O I
10.1109/TSP.2023.3284365
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Model selection in linear regression models is a major challenge when dealing with high-dimensional data where the number of available measurements (sample size) is much smaller than the dimension of the parameter space. Traditional methods for model selection such as Akaike information criterion, Bayesian information criterion (BIC), and minimum description length are heavily prone to overfitting in the high-dimensional setting. In this regard, extended BIC (EBIC), which is an extended version of the original BIC, and extended Fisher information criterion (EFIC), which is a combination of EBIC and Fisher information criterion, are consistent estimators of the true model as the number of measurements grows very large. However, EBIC is not consistent in high signal-to-noise-ratio (SNR) scenarios where the sample size is fixed and EFIC is not invariant to data scaling resulting in unstable behaviour. In this article, we propose a new form of the EBIC criterion called EBIC-Robust, which is invariant to data scaling and consistent in both large sample sizes and high-SNR scenarios. Analytical proofs are presented to guarantee its consistency. Simulation results indicate that the performance of EBIC-Robust is quite superior to that of both EBIC and EFIC.
引用
收藏
页码:2251 / 2266
页数:16
相关论文
共 50 条
  • [31] Variational Bayes for High-Dimensional Linear Regression With Sparse Priors
    Ray, Kolyan
    Szabo, Botond
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2022, 117 (539) : 1270 - 1281
  • [32] Empirical Priors for Prediction in Sparse High-dimensional Linear Regression
    Martin, Ryan
    Tang, Yiqi
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [33] Empirical priors for prediction in sparse high-dimensional linear regression
    Martin, Ryan
    Tang, Yiqi
    Journal of Machine Learning Research, 2020, 21
  • [34] Consistent group selection in high-dimensional linear regression
    Wei, Fengrong
    Huang, Jian
    BERNOULLI, 2010, 16 (04) : 1369 - 1384
  • [35] Robust transfer learning for high-dimensional quantile regression model with linear constraints
    Longjie Cao
    Yunquan Song
    Applied Intelligence, 2024, 54 : 1263 - 1274
  • [36] Robust transfer learning for high-dimensional quantile regression model with linear constraints
    Cao, Longjie
    Song, Yunquan
    APPLIED INTELLIGENCE, 2024, 54 (02) : 1263 - 1274
  • [37] Valid Post-Selection Inference in High-Dimensional Approximately Sparse Quantile Regression Models
    Belloni, Alexandre
    Chernozhukov, Victor
    Kato, Kengo
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2019, 114 (526) : 749 - 758
  • [38] TESTING HIGH-DIMENSIONAL REGRESSION COEFFICIENTS IN LINEAR MODELS
    Zhao, Alex
    Li, Changcheng
    Li, Runze
    Zhang, Zhe
    ANNALS OF STATISTICS, 2024, 52 (05): : 2034 - 2058
  • [39] A generalized information criterion for high-dimensional PCA rank selection
    Hung Hung
    Su-Yun Huang
    Ching-Kang Ing
    Statistical Papers, 2022, 63 : 1295 - 1321
  • [40] Tests for high-dimensional partially linear regression models
    Shi, Hongwei
    Yang, Weichao
    Sun, Bowen
    Guo, Xu
    STATISTICAL PAPERS, 2025, 66 (03)