AMAGOLD: Amortized Metropolis Adjustment for Efficient Stochastic Gradient MCMC

被引:0
|
作者
Zhang, Ruqi [1 ]
Cooper, A. Feder [1 ]
De Sa, Christopher [1 ]
机构
[1] Cornell Univ, Ithaca, NY 14853 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic gradient Hamiltonian Monte Carlo (SGHMC) is an efficient method for sampling from continuous distributions. It is a faster alternative to HMC: instead of using the whole dataset at each iteration, SGHMC uses only a subsample. This improves performance, but introduces bias that can cause SGHMC to converge to the wrong distribution. One can prevent this using a step size that decays to zero, but such a step size schedule can drastically slow down convergence. To address this tension, we propose a novel second-order SGMCMC algorithm-AMAGOLD-that infrequently uses Metropolis-Hastings (M-H) corrections to remove bias. The infrequency of corrections amortizes their cost. We prove AMAGOLD converges to the target distribution with a fixed, rather than a diminishing, step size, and that its convergence rate is at most a constant factor slower than a full-batch baseline. We empirically demonstrate AMAGOLD's effectiveness on synthetic distributions, Bayesian logistic regression, and Bayesian neural networks.
引用
收藏
页码:2142 / 2151
页数:10
相关论文
共 50 条
  • [31] LEARNING SPARSE STRUCTURED ENSEMBLES WITH STOCHASTIC GRADIENT MCMC SAMPLING AND NETWORK PRUNING
    Zhang, Yichi
    Ou, Zhijian
    2018 IEEE 28TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2018,
  • [32] A convergence analysis for a class of practical variance-reduction stochastic gradient MCMC
    Chen, Changyou
    Wang, Wenlin
    Zhang, Yizhe
    Su, Qinliang
    Carin, Lawrence
    SCIENCE CHINA-INFORMATION SCIENCES, 2019, 62 (01)
  • [33] Efficient Stochastic Gradient Hard Thresholding
    Zhou, Pan
    Yuan, Xiao-Tong
    Feng, Jiashi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [34] An adaptively weighted stochastic gradient MCMC algorithm for Monte Carlo simulation and global optimization
    Wei Deng
    Guang Lin
    Faming Liang
    Statistics and Computing, 2022, 32
  • [35] An adaptively weighted stochastic gradient MCMC algorithm for Monte Carlo simulation and global optimization
    Deng, Wei
    Lin, Guang
    Liang, Faming
    STATISTICS AND COMPUTING, 2022, 32 (04)
  • [36] Large-Scale Distributed Bayesian Matrix Factorization using Stochastic Gradient MCMC
    Ahn, Sungjin
    Korattikara, Anoop
    Liu, Nathan
    Rajan, Suju
    Welling, Max
    KDD'15: PROCEEDINGS OF THE 21ST ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2015, : 9 - 18
  • [37] Inference of Stochastic Disease Transmission Models Using Particle-MCMC and a Gradient Based Proposal
    Rosato, Conor
    Harris, John
    Panovska-Griffiths, Jasmina
    Maskell, Simon
    2022 25TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION 2022), 2022,
  • [38] Deep Latent Dirichlet Allocation with Topic-Layer-Adaptive Stochastic Gradient Riemannian MCMC
    Cong, Yulai
    Chen, Bo
    Liu, Hongwei
    Zhou, Mingyuan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [39] Randomized reduced forward models for efficient Metropolis–Hastings MCMC, with application to subsurface fluid flow and capacitance tomography
    Colin Fox
    Tiangang Cui
    Markus Neumayer
    GEM - International Journal on Geomathematics, 2020, 11
  • [40] Quantifying model uncertainty for semantic segmentation of Fluorine-19 MRI using stochastic gradient MCMC
    Javanbakhat, Masoumeh
    Starke, Ludger
    Waiczies, Sonia
    Lippert, Christoph
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 241