AMAGOLD: Amortized Metropolis Adjustment for Efficient Stochastic Gradient MCMC

被引:0
|
作者
Zhang, Ruqi [1 ]
Cooper, A. Feder [1 ]
De Sa, Christopher [1 ]
机构
[1] Cornell Univ, Ithaca, NY 14853 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic gradient Hamiltonian Monte Carlo (SGHMC) is an efficient method for sampling from continuous distributions. It is a faster alternative to HMC: instead of using the whole dataset at each iteration, SGHMC uses only a subsample. This improves performance, but introduces bias that can cause SGHMC to converge to the wrong distribution. One can prevent this using a step size that decays to zero, but such a step size schedule can drastically slow down convergence. To address this tension, we propose a novel second-order SGMCMC algorithm-AMAGOLD-that infrequently uses Metropolis-Hastings (M-H) corrections to remove bias. The infrequency of corrections amortizes their cost. We prove AMAGOLD converges to the target distribution with a fixed, rather than a diminishing, step size, and that its convergence rate is at most a constant factor slower than a full-batch baseline. We empirically demonstrate AMAGOLD's effectiveness on synthetic distributions, Bayesian logistic regression, and Bayesian neural networks.
引用
收藏
页码:2142 / 2151
页数:10
相关论文
共 50 条
  • [1] Efficient and generalizable tuning strategies for stochastic gradient MCMC
    Coullon, Jeremie
    South, Leah
    Nemeth, Christopher
    STATISTICS AND COMPUTING, 2023, 33 (03)
  • [2] Efficient and generalizable tuning strategies for stochastic gradient MCMC
    Jeremie Coullon
    Leah South
    Christopher Nemeth
    Statistics and Computing, 2023, 33 (3)
  • [3] Communication-Efficient Stochastic Gradient MCMC for Neural Networks
    Li, Chunyuan
    Chen, Changyou
    Pu, Yunchen
    Henao, Ricardo
    Carin, Lawrence
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 4173 - 4180
  • [4] Distributed Stochastic Gradient MCMC
    Ahn, Sungjin
    Shahbaba, Babak
    Welling, Max
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 1044 - 1052
  • [5] Structured Stochastic Gradient MCMC
    Alexos, Antonios
    Boyd, Alex
    Mandt, Stephan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 414 - 434
  • [6] STOCHASTIC THERMODYNAMIC INTEGRATION: EFFICIENT BAYESIAN MODEL SELECTION VIA STOCHASTIC GRADIENT MCMC
    Simsekli, Umut
    Badeau, Roland
    Richard, Gael
    Cemgil, Ali Taylan
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 2574 - 2578
  • [7] A Complete Recipe for Stochastic Gradient MCMC
    Ma, Yi-An
    Chen, Tianqi
    Fox, Emily B.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [8] Control variates for stochastic gradient MCMC
    Baker, Jack
    Fearnhead, Paul
    Fox, Emily B.
    Nemeth, Christopher
    STATISTICS AND COMPUTING, 2019, 29 (03) : 599 - 615
  • [9] Control variates for stochastic gradient MCMC
    Jack Baker
    Paul Fearnhead
    Emily B. Fox
    Christopher Nemeth
    Statistics and Computing, 2019, 29 : 599 - 615
  • [10] Stochastic Gradient MCMC with Stale Gradients
    Chen, Changyou
    Ding, Nan
    Li, Chunyuan
    Zhang, Yizhe
    Carin, Lawrence
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29