Bayesian Distillation of Deep Learning Models

被引:1
|
作者
Grabovoy, A. V. [1 ]
Strijov, V. V. [2 ]
机构
[1] Moscow Inst Phys & Technol, Dolgoprudnyi 141701, Russia
[2] Russian Acad Sci, Dorodnicyn Comp Ctr, Moscow 119333, Russia
基金
俄罗斯基础研究基金会;
关键词
model selection; Bayesian inference; model distillation; local transformation; probability space transformation;
D O I
10.1134/S0005117921110023
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We study the problem of reducing the complexity of approximating models and consider methods based on distillation of deep learning models. The concepts of trainer and student are introduced. It is assumed that the student model has fewer parameters than the trainer model. A Bayesian approach to the student model selection is suggested. A method is proposed for assigning an a priori distribution of student parameters based on the a posteriori distribution of trainer model parameters. Since the trainer and student parameter spaces do not coincide, we propose a mechanism for the reduction of the trainer model parameter space to the student model parameter space by changing the trainer model structure. A theoretical analysis of the proposed reduction mechanism is carried out. A computational experiment was carried out on synthesized and real data. The FashionMNIST sample was used as real data.
引用
收藏
页码:1846 / 1856
页数:11
相关论文
共 50 条
  • [21] Bayesian Deep Reinforcement Learning via Deep Kernel Learning
    Xuan, Junyu
    Lu, Jie
    Yan, Zheng
    Zhang, Guangquan
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2019, 12 (01) : 164 - 171
  • [22] Generalized Bayesian Posterior Expectation Distillation for Deep Neural Networks
    Vadera, Meet P.
    Jalain, Brian
    Marlin, Benjamin M.
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI 2020), 2020, 124 : 719 - 728
  • [23] Cutting Weights of Deep Learning Models for Heart Sound Classification: Introducing a Knowledge Distillation Approach
    Song, Zikai
    Zhu, Lixian
    Wang, Yiyan
    Sun, Mengkai
    Qian, Kun
    Hu, Bin
    Yamamoto, Yoshiharu
    Schuller, Bjorn W.
    2023 45TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY, EMBC, 2023,
  • [24] Neuron Manifold Distillation for Edge Deep Learning
    Tao, Zeyi
    Xia, Qi
    Li, Qun
    2021 IEEE/ACM 29TH INTERNATIONAL SYMPOSIUM ON QUALITY OF SERVICE (IWQOS), 2021,
  • [25] Knowledge distillation in deep learning and its applications
    Alkhulaifi, Abdolmaged
    Alsahli, Fahad
    Ahmad, Irfan
    PEERJ COMPUTER SCIENCE, 2021, PeerJ Inc. (07) : 1 - 24
  • [26] An Evaluation of Distillation Deep Learning Network Architecture
    Fujii, Yoshitaka
    Ichimura, Takumi
    2017 IEEE 10TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL INTELLIGENCE AND APPLICATIONS (IWCIA), 2017, : 103 - 108
  • [27] The origin and evolution of open habitats in North America inferred by Bayesian deep learning models
    Andermann, Tobias
    Stromberg, Caroline A. E.
    Antonelli, Alexandre
    Silvestro, Daniele
    NATURE COMMUNICATIONS, 2022, 13 (01)
  • [28] The origin and evolution of open habitats in North America inferred by Bayesian deep learning models
    Tobias Andermann
    Caroline A. E. Strömberg
    Alexandre Antonelli
    Daniele Silvestro
    Nature Communications, 13
  • [29] Class Incremental Learning With Deep Contrastive Learning and Attention Distillation
    Zhu, Jitao
    Luo, Guibo
    Duan, Baishan
    Zhu, Yuesheng
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 1224 - 1228
  • [30] Estimation of agent-based models using Bayesian deep learning approach of BayesFlow
    Shiono, Takashi
    JOURNAL OF ECONOMIC DYNAMICS & CONTROL, 2021, 125