Adaptive Monte Carlo augmented with normalizing flows

被引:76
作者
Gabrie, Marylou [1 ,2 ]
Rotskoff, Grant M. [3 ]
Vanden-Eijnden, Eric [4 ]
机构
[1] Flatiron Inst, Ctr Computat Math, New York, NY 10010 USA
[2] NYU, Ctr Data Sci, New York, NY 10011 USA
[3] Stanford Univ, Dept Chem, Stanford, CA 94305 USA
[4] NYU, Courant Inst Math Sci, 251 Mercer St, New York, NY 10012 USA
关键词
Monte Carlo; normalizing flows; free energy calculations; phase transitions;
D O I
10.1073/pnas.2109420119
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Many problems in the physical sciences, machine learning, and statistical inference necessitate sampling from a high-dimensional, multimodal probability distribution. Markov Chain Monte Carlo (MCMC) algorithms, the ubiquitous tool for this task, typically rely on random local updates to propagate configurations of a given system in a way that ensures that generated configurations will be distributed according to a target probability distribution asymptotically. In high-dimensional settings with multiple relevant metastable basins, local approaches require either immense computational effort or intricately designed importance sampling strategies to capture information about, for example, the relative populations of such basins. Here, we analyze an adaptive MCMC, which augments MCMC sampling with nonlocal transition kernels parameterized with generative models known as normalizing flows. We focus on a setting where there are no preexisting data, as is commonly the case for problems in which MCMC is used. Our method uses 1) an MCMC strategy that blends local moves obtained from any standard transition kernel with those from a generative model to accelerate the sampling and 2) the data generated this way to adapt the generative model and improve its efficacy in the MCMC algorithm. We provide a theoretical analysis of the convergence properties of this algorithm and investigate numerically its efficiency, in particular in terms of its propensity to equilibrate fast between metastable modes whose rough location is known a priori but respective probability weight is not. We show that our algorithm can sample effectively across large free energy barriers, providing dramatic accelerations relative to traditional MCMC algorithms.
引用
收藏
页数:9
相关论文
共 45 条
[1]   Flow-based generative models for Markov chain Monte Carlo in lattice field theory [J].
Albergo, M. S. ;
Kanwar, G. ;
Shanahan, P. E. .
PHYSICAL REVIEW D, 2019, 100 (03)
[2]   On the ergodicity properties of some adaptive MCMC algorithms [J].
Andrieu, Christophe ;
Moulines, Eric .
ANNALS OF APPLIED PROBABILITY, 2006, 16 (03) :1462-1505
[3]   On nonlinear Markov chain Monte Carlo [J].
Andrieu, Christophe ;
Jasra, Ajay ;
Doucet, Arnaud ;
Del Moral, Pierre .
BERNOULLI, 2011, 17 (03) :987-1014
[4]  
[Anonymous], 2018, INT C LEARN REPR ICL
[5]   An Eyring-Kramers law for the stochastic Allen-Cahn equation in dimension two [J].
Berglund, Nils ;
Di Gesu, Giacomo ;
Weber, Hendrik .
ELECTRONIC JOURNAL OF PROBABILITY, 2017, 22
[6]   Transition path sampling: Throwing ropes over rough mountain passes, in the dark [J].
Bolhuis, PG ;
Chandler, D ;
Dellago, C ;
Geissler, PL .
ANNUAL REVIEW OF PHYSICAL CHEMISTRY, 2002, 53 :291-318
[7]   Efficient modeling of trivializing maps for lattice φ4 theory using normalizing flows: A first look at scalability [J].
Del Debbio, Luigi ;
Rossney, Joe Marsh ;
Wilson, Michael .
PHYSICAL REVIEW D, 2021, 104 (09)
[8]  
Dibak M., 2020, 3 WORKSH MACH LEARN
[9]  
Dinh L., 3 INT C LEARN REPR I
[10]  
Dinh L., 2017, INT C LEARN REPR, P32