Bayesian Distillation of Deep Learning Models

被引:1
|
作者
Grabovoy, A. V. [1 ]
Strijov, V. V. [2 ]
机构
[1] Moscow Inst Phys & Technol, Dolgoprudnyi 141701, Russia
[2] Russian Acad Sci, Dorodnicyn Comp Ctr, Moscow 119333, Russia
基金
俄罗斯基础研究基金会;
关键词
model selection; Bayesian inference; model distillation; local transformation; probability space transformation;
D O I
10.1134/S0005117921110023
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We study the problem of reducing the complexity of approximating models and consider methods based on distillation of deep learning models. The concepts of trainer and student are introduced. It is assumed that the student model has fewer parameters than the trainer model. A Bayesian approach to the student model selection is suggested. A method is proposed for assigning an a priori distribution of student parameters based on the a posteriori distribution of trainer model parameters. Since the trainer and student parameter spaces do not coincide, we propose a mechanism for the reduction of the trainer model parameter space to the student model parameter space by changing the trainer model structure. A theoretical analysis of the proposed reduction mechanism is carried out. A computational experiment was carried out on synthesized and real data. The FashionMNIST sample was used as real data.
引用
收藏
页码:1846 / 1856
页数:11
相关论文
共 50 条
  • [1] Bayesian Distillation of Deep Learning Models
    A. V. Grabovoy
    V. V. Strijov
    Automation and Remote Control, 2021, 82 : 1846 - 1856
  • [2] A Bayesian Data Augmentation Approach for Learning Deep Models
    Toan Tran
    Trung Pham
    Carneiro, Gustavo
    Palmer, Lyle
    Reid, Ian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [3] A Deep Learning Method for Comparing Bayesian Hierarchical Models
    Elsemueller, Lasse
    Schnuerch, Martin
    Buerkner, Paul-Christian
    Radev, Stefan T.
    PSYCHOLOGICAL METHODS, 2024,
  • [4] Federated Bayesian Deep Learning: The Application of Statistical Aggregation Methods to Bayesian Models
    Fischer, John
    Orescanin, Marko
    Loomis, Justin
    Mcclure, Patrick
    IEEE ACCESS, 2024, 12 : 185790 - 185806
  • [5] Ensemble Learning of Lightweight Deep Learning Models Using Knowledge Distillation for Image Classification
    Kang, Jaeyong
    Gwak, Jeonghwan
    MATHEMATICS, 2020, 8 (10)
  • [6] Bayesian Active Learning for Choice Models With Deep Gaussian Processes
    Yang, Jie
    Klabjan, Diego
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (02) : 1080 - 1092
  • [7] Explaining Deep Learning Models - A Bayesian Non-parametric Approach
    Guo, Wenbo
    Huang, Sui
    Tao, Yunzhe
    Xing, Xinyu
    Lin, Lin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [8] Rapid Risk Minimization with Bayesian Models Through Deep Learning Approximation
    Lowe, Mathias
    Lunnemann, Per
    Risi, Sebastian
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [9] Verification of integrity of deployed deep learning models using Bayesian Optimization
    Gupta, Sunil
    Nguyen, Dang
    Rana, Santu
    Venkatesh, Svetha
    Kuttichira, Deepthi Praveenlal
    KNOWLEDGE-BASED SYSTEMS, 2022, 241
  • [10] A Review of Dataset Distillation for Deep Learning
    Thi-Thu-Huong Le
    Larasati, Harashta Tatimma
    Prihatno, Aji Teguh
    Kim, Howon
    2022 INTERNATIONAL CONFERENCE ON PLATFORM TECHNOLOGY AND SERVICE (PLATCON22), 2022, : 34 - 37