(f, Γ)-Diyergences: Interpolating between f-Divergences and Integral Probability Metrics

被引:0
|
作者
Birrell, Jeremiah [1 ]
Dupuis, Paul [2 ]
Katsoulakis, Markos A. [3 ]
Pantazis, Yannis [4 ]
Rey-Bellet, Luc [3 ]
机构
[1] Univ Massachusetts, TRIPODS Inst Theoret Fdn Data Sci, Amherst, MA 01003 USA
[2] Brown Univ, Div Appl Math, Providence, RI 02912 USA
[3] Univ Massachusetts, Dept Math & Stat, Amherst, MA 01003 USA
[4] Fdn Res & Technol Hellas, Inst Appl & Computat Math, GR-70013 Iraklion, Greece
关键词
f-divergences; Integral probability metrics; Wasserstein metric; Variational representations; GANs; SENSITIVITY-ANALYSIS; RISK; INFORMATION; FUNCTIONALS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We develop a rigorous and general framework for constructing information-theoretic divergences that subsume both f-divergences and integral probability metrics (IPMs), such as the 1-Wasserstein distance. We prove under which assumptions these divergences, hereafter referred to as (f, Gamma)-divergences, provide a notion of 'distance' between probability measures and show that they can be expressed as a two-stage mass-redistribution/mass-transport process. The (f, Gamma)-divergences inherit features from IPMs, such as the ability to compare distributions which are not absolutely continuous, as well as from f-divergences, namely the strict concavity of their variational representations and the ability to control heavy-tailed distributions for particular choices of f . When combined, these features establish a divergence with improved properties for estimation, statistical learning, and uncertainty quantification applications. Using statistical learning as an example, we demonstrate their advantage in training generative adversarial networks (GANs) for heavy-tailed, not-absolutely continuous sample distributions. We also show improved performance and stability over gradient-penalized Wasserstein GAN in image generation.
引用
收藏
页码:1 / 70
页数:70
相关论文
共 50 条
  • [1] (f,Γ)-Divergences: Interpolating between f-Divergences and Integral Probability Metrics
    Birrell, Jeremiah
    Dupuis, Paul
    Katsoulakis, Markos A.
    Pantazis, Yannis
    Rey-Bellet, Luc
    Journal of Machine Learning Research, 2022, 23
  • [2] Optimal Bounds between f-Divergences and Integral Probability Metrics
    Agrawal, Rohit
    Horel, Thibaut
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [3] Optimal bounds between f-divergences and integral probability metrics
    Agrawal, Rohit
    Horel, Thibaut
    Journal of Machine Learning Research, 2021, 22
  • [4] Optimal Bounds between f-Divergences and Integral Probability Metrics
    Agrawal, Rohit
    Horel, Thibaut
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [5] On f-divergences Between Cauchy Distributions
    Nielsen, Frank
    Okamura, Kazuki
    GEOMETRIC SCIENCE OF INFORMATION (GSI 2021), 2021, 12829 : 799 - 807
  • [6] On f-Divergences Between Cauchy Distributions
    Nielsen, Frank
    Okamura, Kazuki
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2023, 69 (05) : 3150 - 3171
  • [7] On Relativistic f-Divergences
    Jolicoeur-Martineau, Alexia
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [8] On f-Divergences: Integral Representations, Local Behavior, and Inequalities
    Sason, Igal
    ENTROPY, 2018, 20 (05)
  • [9] Quantum Renyi and f-Divergences from Integral Representations
    Hirche, Christoph
    Tomamichel, Marco
    COMMUNICATIONS IN MATHEMATICAL PHYSICS, 2024, 405 (09)
  • [10] Zipf–Mandelbrot law, f-divergences and the Jensen-type interpolating inequalities
    Neda Lovričević
    Ðilda Pečarić
    Josip Pečarić
    Journal of Inequalities and Applications, 2018