Sharpened Generalization Bounds based on Conditional Mutual Information and an Application to Noisy, Iterative Algorithms

被引:0
|
作者
Haghifam, Mahdi [1 ]
Negrea, Jeffrey [1 ]
Khisti, Ashish [2 ]
Roy, Daniel M. [1 ]
Dziugaite, Gintare Karolina [3 ]
机构
[1] Univ Toronto, Vector Inst, Toronto, ON, Canada
[2] Univ Toronto, Toronto, ON, Canada
[3] Element AI, Toronto, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The information-theoretic framework of Russo and Zou (2016) and Xu and Raginsky (2017) provides bounds on the generalization error of a learning algorithm in terms of the mutual information between the algorithm's output and the training sample. In this work, we study the proposal by Steinke and Zakynthinou (2020) to reason about the generalization error of a learning algorithm by introducing a super sample that contains the training sample as a random subset and computing mutual information conditional on the super sample. We first show that these new bounds based on the conditional mutual information are tighter than those based on the unconditional mutual information. We then introduce yet tighter bounds, building on the "individual sample" idea of Bu et al. (2019) and the "data dependent" ideas of Negrea et al. (2019), using disintegrated mutual information. Finally, we apply these bounds to the study of the Langevin dynamics algorithm, showing that conditioning on the super sample allows us to exploit information in the optimization trajectory to obtain tighter bounds based on hypothesis tests.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Generalization Error Bounds for Noisy, Iterative Algorithms
    Pensia, Ankit
    Jog, Varun
    Loh, Po-Ling
    2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 546 - 550
  • [2] Generalization Error Bounds for Noisy, Iterative Algorithms via Maximal Leakage
    Issa, Ibrahim
    Esposito, Amedeo Roberto
    Gastpar, Michael
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [3] Generalization Bounds for Noisy Iterative Algorithms Using Properties of Additive Noise Channels
    Wang, Hao
    Gao, Rui
    Calmon, Flavio P.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [4] Tightening Mutual Information Based Bounds on Generalization Error
    Bu, Yuheng
    Zou, Shaofeng
    Veeravalli, Venugopal V.
    2019 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2019, : 587 - 591
  • [5] Tightening mutual information-based bounds on generalization error
    Bu Y.
    Zou S.
    Veeravalli V.V.
    IEEE Journal on Selected Areas in Information Theory, 2020, 1 (01): : 121 - 130
  • [6] Chaining Mutual Information and Tightening Generalization Bounds
    Asadi, Amir R.
    Abbe, Emmanuel
    Verdu, Sergio
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [7] Conditional Mutual Information-Based Generalization Bound for Meta Learning
    Rezazadeh, Arezou
    Jose, Sharu Theresa
    Durisi, Giuseppe
    Simeone, Osvaldo
    2021 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2021, : 1176 - 1181
  • [8] Methods and Algorithms of Information Generalization in Noisy Databases
    Vagin, Vadim
    Fomina, Marina
    ADVANCES IN SOFT COMPUTING - MICAI 2010, PT II, 2010, 6438 : 44 - 55
  • [9] Generalization bounds via information density and conditional information density
    Hellström F.
    Durisi G.
    IEEE Journal on Selected Areas in Information Theory, 2020, 1 (03): : 824 - 839
  • [10] Universal Spreading of Conditional Mutual Information in Noisy Random Circuits
    Lee, Su-un
    Oh, Changhun
    Wong, Yat
    Chen, Senrui
    Jiang, Liang
    PHYSICAL REVIEW LETTERS, 2024, 133 (20)