Information-Theoretic GAN Compression with Variational Energy-based Model

被引:0
|
作者
Kang, Minsoo [1 ]
Yoo, Hyewon [2 ]
Kang, Eunhee [3 ]
Ki, Sehwan [3 ]
Lee, Hyong-Euk [3 ]
Han, Bohyung [1 ,2 ]
机构
[1] Seoul Natl Univ, ECE, Seoul, South Korea
[2] Seoul Natl Univ, IPAI, Seoul, South Korea
[3] Samsung Adv Inst Technol SAIT, Suwon, South Korea
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose an information-theoretic knowledge distillation approach for the compression of generative adversarial networks, which aims to maximize the mutual information between teacher and student networks via a variational optimization based on an energy-based model. Because the direct computation of the mutual information in continuous domains is intractable, our approach alternatively optimizes the student network by maximizing the variational lower bound of the mutual information. To achieve a tight lower bound, we introduce an energy-based model relying on a deep neural network to represent a flexible variational distribution that deals with high-dimensional images and consider spatial dependencies between pixels, effectively. Since the proposed method is a generic optimization algorithm, it can be conveniently incorporated into arbitrary generative adversarial networks and even dense prediction networks, e.g., image enhancement models. We demonstrate that the proposed algorithm achieves outstanding performance in model compression of generative adversarial networks consistently when combined with several existing models.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] PROBABILISTIC IMAGE MODEL AND ITS INFORMATION-THEORETIC PROPERTIES
    ZHANG, YQ
    LOEW, MH
    PICKHOLTZ, RL
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1993, 41 (01) : 508 - 510
  • [42] Infomod: information-theoretic machine learning model diagnostics
    Esmaelizadeh, Armin
    Cotterill, Sunil
    Hebert, Liam
    Golab, Lukasz
    Taghva, Kazem
    DISTRIBUTED AND PARALLEL DATABASES, 2025, 43 (01)
  • [43] Learning to Explain: An Information-Theoretic Perspective on Model Interpretation
    Chen, Jianbo
    Song, Le
    Wainwright, Martin J.
    Jordan, Michael I.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [44] An information-theoretic learning model based on importance sampling with application in face verification
    Zhang, Jiangshe
    Ji, Lizhen
    Gao, Fei
    Li, Mengyao
    Zhang, Chunxia
    Cui, Yukun
    PATTERN RECOGNITION LETTERS, 2025, 188 : 81 - 87
  • [45] Comment on “Zero and negative energy dissipation at information-theoretic erasure”
    Neal G. Anderson
    Journal of Computational Electronics, 2016, 15 : 340 - 342
  • [47] Potential energy landscapes identify the information-theoretic nature of the epigenome
    Garrett Jenkinson
    Elisabet Pujadas
    John Goutsias
    Andrew P Feinberg
    Nature Genetics, 2017, 49 : 719 - 729
  • [48] Information-theoretic lower bound on energy cost of stochastic computation
    Wiesner, Karoline
    Gu, Mile
    Rieper, Elisabeth
    Vedral, Vlatko
    PROCEEDINGS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2012, 468 (2148): : 4058 - 4066
  • [49] ENERGY-BASED MODEL OF A LIQUID PISTON GAS COMPRESSION SYSTEM
    Kuznyetsov, Oleksiy
    Atamaniuk, Vitalii
    REVUE ROUMAINE DES SCIENCES TECHNIQUES-SERIE ELECTROTECHNIQUE ET ENERGETIQUE, 2025, 70 (01): : 145 - 150
  • [50] Comment on "Zero and negative energy dissipation at information-theoretic erasure"
    Anderson, Neal G.
    JOURNAL OF COMPUTATIONAL ELECTRONICS, 2016, 15 (01) : 340 - 342