AMAGOLD: Amortized Metropolis Adjustment for Efficient Stochastic Gradient MCMC

被引:0
|
作者
Zhang, Ruqi [1 ]
Cooper, A. Feder [1 ]
De Sa, Christopher [1 ]
机构
[1] Cornell Univ, Ithaca, NY 14853 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic gradient Hamiltonian Monte Carlo (SGHMC) is an efficient method for sampling from continuous distributions. It is a faster alternative to HMC: instead of using the whole dataset at each iteration, SGHMC uses only a subsample. This improves performance, but introduces bias that can cause SGHMC to converge to the wrong distribution. One can prevent this using a step size that decays to zero, but such a step size schedule can drastically slow down convergence. To address this tension, we propose a novel second-order SGMCMC algorithm-AMAGOLD-that infrequently uses Metropolis-Hastings (M-H) corrections to remove bias. The infrequency of corrections amortizes their cost. We prove AMAGOLD converges to the target distribution with a fixed, rather than a diminishing, step size, and that its convergence rate is at most a constant factor slower than a full-batch baseline. We empirically demonstrate AMAGOLD's effectiveness on synthetic distributions, Bayesian logistic regression, and Bayesian neural networks.
引用
收藏
页码:2142 / 2151
页数:10
相关论文
共 50 条
  • [41] IMPROVING SAMPLING ACCURACY OF STOCHASTIC GRADIENT MCMC METHODS VIA NON-UNIFORM SUBSAMPLING OF GRADIENTS
    Li, Ruilin
    Wang, Xin
    Zha, Hongyuan
    Tao, Molei
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS, 2021,
  • [42] IMPROVING SAMPLING ACCURACY OF STOCHASTIC GRADIENT MCMC METHODS VIA NON-UNIFORM SUBSAMPLING OF GRADIENTS
    Li, Ruilin
    Wang, Xin
    Zha, Hongyuan
    Tao, Molei
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS-SERIES S, 2023, 16 (02): : 329 - 360
  • [43] Randomized reduced forward models for efficient Metropolis-Hastings MCMC, with application to subsurface fluid flow and capacitance tomography
    Fox, Colin
    Cui, Tiangang
    Neumayer, Markus
    GEM-INTERNATIONAL JOURNAL ON GEOMATHEMATICS, 2020, 11 (01)
  • [45] AN EFFICIENT GRADIENT PROJECTION METHOD FOR STOCHASTIC OPTIMAL CONTROL PROBLEMS
    Gong, Bo
    Liu, Wenbin
    Tang, Tao
    Zhao, Weidong
    Zhou, Tao
    SIAM JOURNAL ON NUMERICAL ANALYSIS, 2017, 55 (06) : 2982 - 3005
  • [46] A STOCHASTIC VERSION OF STEIN VARIATIONAL GRADIENT DESCENT FOR EFFICIENT SAMPLING
    Li, Lei
    Li, Yingzhou
    Liu, Jian-Guo
    Liu, Zibu
    Lu, Jianfeng
    COMMUNICATIONS IN APPLIED MATHEMATICS AND COMPUTATIONAL SCIENCE, 2020, 15 (01) : 37 - 63
  • [47] An Efficient Preconditioner for Stochastic Gradient Descent Optimization of Image Registration
    Qiao, Yuchuan
    Lelieveldt, Boudewijn P. F.
    Staring, Marius
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2019, 38 (10) : 2314 - 2325
  • [48] Stochastic Gradient Descent-like relaxation is equivalent to Metropolis dynamics in discrete optimization and inference problems
    Angelini, Maria Chiara
    Cavaliere, Angelo Giorgio
    Marino, Raffaele
    Ricci-Tersenghi, Federico
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [49] Highly efficient online stochastic gradient and sliding window stochastic gradient signal modelling methods for multi-frequency signals
    Song, Guanglei
    Xu, Ling
    INTERNATIONAL JOURNAL OF MODELLING IDENTIFICATION AND CONTROL, 2024, 44 (01) : 14 - 22
  • [50] Efficient preconditioned stochastic gradient descent for estimation in latent variable models
    Baey, Charlotte
    Delattre, Maud
    Kuhn, Estelle
    Leger, Jean-Benoist
    Lemler, Sarah
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202