Bayesian grouping-Gibbs sampling estimation of high-dimensional linear model with non-sparsity

被引:0
|
作者
Qin, Shanshan [1 ]
Zhang, Guanlin [2 ]
Wu, Yuehua [2 ]
Zhu, Zhongyi [3 ]
机构
[1] Tianjin Univ Finance & Econ, Sch Stat, Tianjin 300222, Peoples R China
[2] York Univ, Dept Math & Stat, Toronto, ON M3J 1P3, Canada
[3] Fudan Univ, Dept Stat & Data Sci, Shanghai 200437, Peoples R China
基金
中国国家自然科学基金;
关键词
Bayesian grouping-Gibbs sampling; Non-sparse; High-dimensional; Linear regression; VARIABLE SELECTION; REGRESSION; CONSISTENT; MIXTURES;
D O I
10.1016/j.csda.2024.108072
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In high-dimensional linear regression models, common assumptions typically entail sparsity of regression coefficients beta is an element of R-p. However, these assumptions may not hold when the majority, if not all, of regression coefficients are non-zeros. Statistical methods designed for sparse models may lead to substantial bias in model estimation. Therefore, this article proposes a novel Bayesian Grouping-Gibbs Sampling (BGGS) method, which departs from the common sparse assumptions in high-dimensional problems. The BGGS method leverages a grouping strategy that partitions beta into distinct groups, facilitating rapid sampling in high-dimensional space. The grouping number (k) can be determined using the 'Elbow plot', which operates efficiently and is robust against the initial value. Theoretical analysis, under some regular conditions, guarantees model selection and parameter estimation consistency, and bound for the prediction error. Furthermore, three finite simulations are conducted to assess the competitive advantages of the proposed method in terms of parameter estimation and prediction accuracy. Finally, the BGGS method is applied to a financial dataset to explore its practical utility.
引用
收藏
页数:21
相关论文
共 50 条
  • [41] Simultaneous estimation of cluster number and feature sparsity in high-dimensional cluster analysis
    Li, Yujia
    Zeng, Xiangrui
    Lin, Chien-Wei
    Tseng, George C.
    BIOMETRICS, 2022, 78 (02) : 574 - 585
  • [42] Shrinkage and Sparse Estimation for High-Dimensional Linear Models
    Asl, M. Noori
    Bevrani, H.
    Belaghi, R. Arabi
    Ahmed, Syed Ejaz
    PROCEEDINGS OF THE THIRTEENTH INTERNATIONAL CONFERENCE ON MANAGEMENT SCIENCE AND ENGINEERING MANAGEMENT, VOL 1, 2020, 1001 : 147 - 156
  • [43] Noise Level Estimation in High-Dimensional Linear Models
    G. K. Golubev
    E. A. Krymova
    Problems of Information Transmission, 2018, 54 : 351 - 371
  • [44] Robust Estimation of High-Dimensional Linear Regression With Changepoints
    Cui, Xiaolong
    Geng, Haoyu
    Wang, Zhaojun
    Zou, Changliang
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2024, 70 (10) : 7297 - 7319
  • [45] Estimation of high-dimensional change-points under a group sparsity structure
    Cai, Hanqing
    Wang, Tengyao
    ELECTRONIC JOURNAL OF STATISTICS, 2023, 17 (01): : 858 - 894
  • [46] Noise Level Estimation in High-Dimensional Linear Models
    Golubev, G. K.
    Krymova, E. A.
    PROBLEMS OF INFORMATION TRANSMISSION, 2018, 54 (04) : 351 - 371
  • [47] Estimation in High-Dimensional Analysis and Multivariate Linear Models
    Kollo, Tonu
    Von Rosen, Tatjana
    Von Rosen, Dietrich
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2011, 40 (07) : 1241 - 1253
  • [48] Thompson Sampling for High-Dimensional Sparse Linear Contextual Bandits
    Chakraborty, Sunrit
    Roy, Saptarshi
    Tewari, Ambuj
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [49] High-Dimensional MVAR Model Identification Based on Structured Sparsity Penalization
    Fang, Z.
    Albera, L.
    Kachenoura, A.
    Shu, H.
    Kang, Y.
    Jeannes, R. Le Bouquin
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1975 - 1979
  • [50] Intrinsic Bayesian model for high-dimensional unsupervised reduction
    Jin, Longcun
    Wan, Wanggen
    Wu, Yongliang
    Cui, Bin
    Yu, Xiaoqing
    NEUROCOMPUTING, 2012, 98 : 143 - 150