Bayesian regression based on principal components for high-dimensional data

被引:8
|
作者
Lee, Jaeyong [1 ]
Oh, Hee-Seok [1 ]
机构
[1] Seoul Natl Univ, Seoul 151, South Korea
基金
新加坡国家研究基金会;
关键词
D O I
10.1016/j.jmva.2013.02.002
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
The Gaussian sequence model can be obtained from the high-dimensional regression model through principal component analysis. It is shown that the Gaussian sequence model is equivalent to the original high-dimensional regression model in terms of prediction. Under a sparsity condition, we investigate the posterior consistency and convergence rates of the Gaussian sequence model. In particular, we examine two different modeling strategies: Bayesian inference with and without covariate selection. For Bayesian inferences without covariate selection, we obtain the consistency results of the estimators and posteriors with normal priors with constant and decreasing variances, and the James Stein estimator; for Bayesian inference with covariate selection, we obtain convergence rates of Bayesian model averaging (BMA) and median probability model (MPM) estimators, and the posterior with variable selection prior. Based on these results, we conclude that variable selection is essential in high-dimensional Bayesian regression. A simulation study also confirms the conclusion. The methodologies are applied to a climate prediction problem. (C) 2013 Elsevier Inc. All rights reserved.
引用
收藏
页码:175 / 192
页数:18
相关论文
共 50 条
  • [21] Unconditional quantile regression with high-dimensional data
    Sasaki, Yuya
    Ura, Takuya
    Zhang, Yichong
    QUANTITATIVE ECONOMICS, 2022, 13 (03) : 955 - 978
  • [22] Robust Ridge Regression for High-Dimensional Data
    Maronna, Ricardo A.
    TECHNOMETRICS, 2011, 53 (01) : 44 - 53
  • [23] Bayesian Regression Trees for High-Dimensional Prediction and Variable Selection
    Linero, Antonio R.
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2018, 113 (522) : 626 - 636
  • [24] A comparative study on high-dimensional bayesian regression with binary predictors
    Slanzi, Debora
    Mameli, Valentina
    Brown, Philip J.
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2023, 52 (05) : 1979 - 1999
  • [25] A comparison study of Bayesian high-dimensional linear regression models
    Shin, Ju-Won
    Lee, Kyoungjae
    KOREAN JOURNAL OF APPLIED STATISTICS, 2021, 34 (03) : 491 - 505
  • [26] Principal component analysis for sparse high-dimensional data
    Raiko, Tapani
    Ilin, Alexander
    Karhunen, Juha
    NEURAL INFORMATION PROCESSING, PART I, 2008, 4984 : 566 - 575
  • [27] Visualization of high-dimensional data on the probabilistic principal surface
    Chang, KY
    Ghosh, J
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON INDUSTRIAL ENGINEERING AND ENGINEERING MANAGEMENT, VOLS 1 AND 2: INDUSTRIAL ENGINEERING AND ENGINEERING MANAGEMENT IN THE GLOBAL ECONOMY, 2005, : 1315 - 1319
  • [28] Applying stability selection to consistently estimate sparse principal components in high-dimensional molecular data
    Sill, Martin
    Saadati, Maral
    Benner, Axel
    BIOINFORMATICS, 2015, 31 (16) : 2683 - 2690
  • [29] Regression-Based Network Estimation for High-Dimensional Genetic Data
    Lee, Kyu Min
    Lee, Minhyeok
    Seok, Junhee
    Han, Sung Won
    JOURNAL OF COMPUTATIONAL BIOLOGY, 2019, 26 (04) : 336 - 349
  • [30] Ensemble-Based Local Learning for High-Dimensional Data Regression
    Raytchev, B.
    Katamoto, Y.
    Koujiba, M.
    Tamaki, T.
    Kaneda, K.
    2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 2640 - 2645