AMAGOLD: Amortized Metropolis Adjustment for Efficient Stochastic Gradient MCMC

被引:0
|
作者
Zhang, Ruqi [1 ]
Cooper, A. Feder [1 ]
De Sa, Christopher [1 ]
机构
[1] Cornell Univ, Ithaca, NY 14853 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic gradient Hamiltonian Monte Carlo (SGHMC) is an efficient method for sampling from continuous distributions. It is a faster alternative to HMC: instead of using the whole dataset at each iteration, SGHMC uses only a subsample. This improves performance, but introduces bias that can cause SGHMC to converge to the wrong distribution. One can prevent this using a step size that decays to zero, but such a step size schedule can drastically slow down convergence. To address this tension, we propose a novel second-order SGMCMC algorithm-AMAGOLD-that infrequently uses Metropolis-Hastings (M-H) corrections to remove bias. The infrequency of corrections amortizes their cost. We prove AMAGOLD converges to the target distribution with a fixed, rather than a diminishing, step size, and that its convergence rate is at most a constant factor slower than a full-batch baseline. We empirically demonstrate AMAGOLD's effectiveness on synthetic distributions, Bayesian logistic regression, and Bayesian neural networks.
引用
收藏
页码:2142 / 2151
页数:10
相关论文
共 50 条
  • [21] An adaptive Hessian approximated stochastic gradient MCMC method
    Wang, Yating
    Deng, Wei
    Lin, Guang
    JOURNAL OF COMPUTATIONAL PHYSICS, 2021, 432
  • [22] CPSG-MCMC: Clustering-Based Preprocessing method for Stochastic Gradient MCMC
    Fu, Tianfan
    Zhang, Zhihua
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 54, 2017, 54 : 841 - 850
  • [23] Learning Weight Uncertainty with Stochastic Gradient MCMC for Shape Classification
    Li, Chunyuan
    Stevens, Andrew
    Chen, Changyou
    Pu, Yunchen
    Gan, Zhe
    Carin, Lawrence
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 5666 - 5675
  • [24] Learning Deep Generative Models With Doubly Stochastic Gradient MCMC
    Du, Chao
    Zhu, Jun
    Zhang, Bo
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (07) : 3084 - 3096
  • [25] Bayesian sparse learning with preconditioned stochastic gradient MCMC and its applications
    Wang, Yating
    Deng, Wei
    Lin, Guang
    JOURNAL OF COMPUTATIONAL PHYSICS, 2021, 432
  • [26] On the Convergence of Stochastic Gradient MCMC Algorithms with High-Order Integrators
    Chen, Changyou
    Ding, Nan
    Carin, Lawrence
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [27] A convergence analysis for a class of practical variance-reduction stochastic gradient MCMC
    Changyou CHEN
    Wenlin WANG
    Yizhe ZHANG
    Qinliang SU
    Lawrence CARIN
    Science China(Information Sciences), 2019, 62 (01) : 67 - 79
  • [28] Non-convex Learning via Replica Exchange Stochastic Gradient MCMC
    Deng, Wei
    Feng, Qi
    Gao, Liyao
    Liang, Faming
    Lin, Guang
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [29] Non-convex Learning via Replica Exchange Stochastic Gradient MCMC
    Deng, Wei
    Feng, Qi
    Gao, Liyao
    Liang, Faming
    Lin, Guang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [30] A convergence analysis for a class of practical variance-reduction stochastic gradient MCMC
    Changyou Chen
    Wenlin Wang
    Yizhe Zhang
    Qinliang Su
    Lawrence Carin
    Science China Information Sciences, 2019, 62