Wide Bayesian neural networks have a simple weight posterior: theory and accelerated sampling

被引:0
|
作者
Hron, Jiri [1 ,2 ]
Novak, Roman [1 ]
Pennington, Jeffrey [1 ]
Sohl-Dickstein, Jascha [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
[2] Univ Cambridge, Cambridge, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce repriorisation, a data-dependent reparameterisation which transforms a Bayesian neural network (BNN) posterior to a distribution whose KL divergence to the BNN prior vanishes as layer widths grow. The repriorisation map acts directly on parameters, and its analytic simplicity complements the known neural network Gaussian process (NNGP) behaviour of wide BNNs in function space. Exploiting the repriorisation, we develop a Markov chain Monte Carlo (MCMC) posterior sampling algorithm which mixes faster the wider the BNN. This contrasts with the typically poor performance of MCMC in high dimensions. We observe up to 50x higher effective sample size relative to no reparametrisation for both fully-connected and residual networks. Improvements are achieved at all widths, with the margin between reparametrised and standard BNNs growing with layer width.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Gibbs sampling the posterior of neural networks
    Piccioli, Giovanni
    Troiani, Emanuele
    Zdeborova, Lenka
    JOURNAL OF PHYSICS A-MATHEMATICAL AND THEORETICAL, 2024, 57 (12)
  • [2] On MCMC sampling in Bayesian MLP neural networks
    Vehtari, A
    Särkkä, S
    Lampinen, J
    IJCNN 2000: PROCEEDINGS OF THE IEEE-INNS-ENNS INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOL I, 2000, : 317 - 322
  • [3] Improving Bayesian Neural Networks by Adversarial Sampling
    Zhang, Jiaru
    Hua, Yang
    Song, Tao
    Wang, Hao
    Xue, Zhengui
    Ma, Ruhui
    Guan, Haibing
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10110 - 10117
  • [4] Bayesian inference with finitely wide neural networks
    Lu, Chi-Ken
    PHYSICAL REVIEW E, 2023, 108 (01)
  • [5] Accelerated Bayesian SED Modeling Using Amortized Neural Posterior Estimation
    Hahn, ChangHoon
    Melchior, Peter
    ASTROPHYSICAL JOURNAL, 2022, 938 (01):
  • [6] Approximate blocked Gibbs sampling for Bayesian neural networks
    Theodore Papamarkou
    Statistics and Computing, 2023, 33
  • [7] Approximate blocked Gibbs sampling for Bayesian neural networks
    Papamarkou, Theodore
    STATISTICS AND COMPUTING, 2023, 33 (05)
  • [8] Bayesian Graph Neural Networks with Adaptive Connection Sampling
    Hasanzadeh, Arman
    Hajiramezanali, Ehsan
    Boluki, Shahin
    Zhou, Mingyuan
    Duffield, Nick
    Narayanan, Krishna
    Qian, Xiaoning
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [9] Accelerated Physical Emulation of Bayesian Inference in Spiking Neural Networks
    Kungl, Akos F.
    Schmitt, Sebastian
    Klaehn, Johann
    Mueller, Paul
    Baumbach, Andreas
    Dold, Dominik
    Kugele, Alexander
    Mueller, Eric
    Koke, Christoph
    Kleider, Mitja
    Mauch, Christian
    Breitwieser, Oliver
    Leng, Luziwei
    Guertler, Nico
    Guettler, Maurice
    Husmann, Dan
    Husmann, Kai
    Hartel, Andreas
    Karasenko, Vitali
    Gruebl, Andreas
    Schemmel, Johannes
    Meier, Karlheinz
    Petrovici, Mihai A.
    FRONTIERS IN NEUROSCIENCE, 2019, 13
  • [10] Simple Neural Networks and Bayesian Inference: A Computational Account
    Dawson, Michael R. W.
    Dupuis, Brian A.
    Jans, Sheldon
    CANADIAN JOURNAL OF EXPERIMENTAL PSYCHOLOGY-REVUE CANADIENNE DE PSYCHOLOGIE EXPERIMENTALE, 2013, 67 (04): : 299 - 299