Bounding the Bias of Contrastive Divergence Learning

被引:37
|
作者
Fischer, Asja [1 ]
Igel, Christian [2 ]
机构
[1] Ruhr Univ Bochum, Inst Neuroinformat, D-44780 Bochum, Germany
[2] Univ Copenhagen, Dept Comp Sci, DK-2100 Copenhagen O, Denmark
关键词
D O I
10.1162/NECO_a_00085
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Optimization based on k-step contrastive divergence (CD) has become a common way to train restricted Boltzmann machines (RBMs). The k-step CD is a biased estimator of the log-likelihood gradient relying on Gibbs sampling. We derive a new upper bound for this bias. Its magnitude depends on k, the number of variables in the RBM, and the maximum change in energy that can be produced by changing a single variable. The last reflects the dependence on the absolute values of the RBM parameters. The magnitude of the bias is also affected by the distance in variation between the modeled distribution and the starting distribution of the Gibbs chain.
引用
收藏
页码:664 / 673
页数:10
相关论文
共 50 条
  • [41] GBVSSL: Contrastive Semi-Supervised Learning Based on Generalized Bias-Variance Decomposition
    Li, Shu
    Han, Lixin
    Wang, Yang
    Zhu, Jun
    SYMMETRY-BASEL, 2024, 16 (06):
  • [42] Neighborhood-Based Stopping Criterion for Contrastive Divergence
    Romero Merino, Enrique
    Mazzanti Castrillejo, Ferran
    Delgado Pin, Jordi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (07) : 2695 - 2704
  • [43] Why do We Need Large Batchsizes in Contrastive Learning? A Gradient-Bias Perspective
    Chen, Changyou
    Zhang, Jianyi
    Xu, Yi
    Chen, Liqun
    Duan, Jiali
    Chen, Yiran
    Tran, Son Dinh
    Zeng, Belinda
    Chilimbi, Trishul
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [44] Bayesian Pseudo-Coresets via Contrastive Divergence
    Tiwary, Piyush
    Shubham, Kumar
    Kashyap, Vivek V.
    Prathosh, A. P.
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2024, 244 : 3368 - 3390
  • [45] Average Contrastive Divergence for Training Restricted Boltzmann Machines
    Ma, Xuesi
    Wang, Xiaojie
    ENTROPY, 2016, 18 (01):
  • [46] Active learning for reducing bias and variance of a classifier using Jensen-Shannon divergence
    Aminian, M
    ICMLA 2005: FOURTH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, PROCEEDINGS, 2005, : 43 - 48
  • [47] Decoupled Contrastive Learning
    Yeh, Chun-Hsiao
    Hong, Cheng-Yao
    Hsu, Yen-Chi
    Liu, Tyng-Luh
    Chen, Yubei
    LeCun, Yann
    COMPUTER VISION, ECCV 2022, PT XXVI, 2022, 13686 : 668 - 684
  • [48] Geometric Contrastive Learning
    Koishekenov, Yeskendir
    Vadgama, Sharvaree
    Valperga, Riccardo
    Bekkers, Erik J.
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 206 - 215
  • [49] Supervised Contrastive Learning
    Khosla, Prannay
    Teterwak, Piotr
    Wang, Chen
    Sarna, Aaron
    Tian, Yonglong
    Isola, Phillip
    Maschinot, Aaron
    Liu, Ce
    Krishnan, Dilip
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [50] Parametric Contrastive Learning
    Cui, Jiequan
    Zhong, Zhisheng
    Liu, Shu
    Yu, Bei
    Jia, Jiaya
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 695 - 704