Invariance Learning in Deep Neural Networks with Differentiable Laplace Approximations

被引:0
|
作者
Immer, Alexander [1 ,2 ]
van der Ouderaa, Tycho F. A. [3 ]
Ratsch, Gunnar [1 ]
Fortuin, Vincent [1 ,4 ]
van der Wilk, Mark [3 ]
机构
[1] Swiss Fed Inst Technol, Dept Comp Sci, Zurich, Switzerland
[2] Max Planck Inst Intelligent Syst, Tubingen, Germany
[3] Imperial Coll London, Dept Comp, London, England
[4] Univ Cambridge, Dept Engn, Cambridge, England
基金
瑞士国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data augmentation is commonly applied to improve performance of deep learning by enforcing the knowledge that certain transformations on the input preserve the output. Currently, the data augmentation parameters are chosen by human effort and costly cross-validation, which makes it cumbersome to apply to new datasets. We develop a convenient gradient-based method for selecting the data augmentation without validation data during training of a deep neural network. Our approach relies on phrasing data augmentation as an invariance in the prior distribution on the functions of a neural network, which allows us to learn it using Bayesian model selection. This has been shown to work in Gaussian processes, but not yet for deep neural networks. We propose a differentiable Kronecker-factored Laplace approximation to the marginal likelihood as our objective, which can be optimised without human supervision or validation data. We show that our method can successfully recover invariances present in the data, and that this improves generalisation and data efficiency on image datasets.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] Deep learning with coherent VCSEL neural networks
    Chen, Zaijun
    Sludds, Alexander
    Davis III, Ronald
    Christen, Ian
    Bernstein, Liane
    Ateshian, Lamia
    Heuser, Tobias
    Heermeier, Niels
    Lott, James A.
    Reitzenstein, Stephan
    Hamerly, Ryan
    Englund, Dirk
    NATURE PHOTONICS, 2023, 17 (08) : 723 - +
  • [42] Learning hidden chemistry with deep neural networks
    Nguyen, Tien-Cuong
    Nguyen, Van-Quyen
    Ngo, Van-Linh
    Than, Quang-Khoat
    Pham, Tien-Lam
    COMPUTATIONAL MATERIALS SCIENCE, 2021, 200
  • [43] Learning Structured Sparsity in Deep Neural Networks
    Wen, Wei
    Wu, Chunpeng
    Wang, Yandan
    Chen, Yiran
    Li, Hai
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [44] Evolving Deep Neural Networks for Continuous Learning
    Atamanczuk, Bruna
    Karadas, Kurt Arve Skipenes
    Agrawal, Bikash
    Chakravorty, Antorweep
    FRONTIERS OF ARTIFICIAL INTELLIGENCE, ETHICS, AND MULTIDISCIPLINARY APPLICATIONS, FAIEMA 2023, 2024, : 3 - 16
  • [45] Neural networks and deep learning: a brief introduction
    Georgevici, Adrian Iustin
    Terblanche, Marius
    INTENSIVE CARE MEDICINE, 2019, 45 (05) : 712 - 714
  • [46] Representational Distance Learning for Deep Neural Networks
    McClure, Patrick
    Kriegeskorte, Nikolaus
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2016, 10
  • [47] Learning hidden elasticity with deep neural networks
    Chen, Chun-Teh
    Gu, Grace X.
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2021, 118 (31)
  • [48] Evolutionary neural networks for deep learning: a review
    Ma, Yongjie
    Xie, Yirong
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2022, 13 (10) : 3001 - 3018
  • [49] Deep Learning for Epidemiologists: An Introduction to Neural Networks
    Serghiou, Stylianos
    Rough, Kathryn
    AMERICAN JOURNAL OF EPIDEMIOLOGY, 2023, 192 (11) : 1904 - 1916
  • [50] Learning Sparse Patterns in Deep Neural Networks
    Wen, Weijing
    Yang, Fan
    Su, Yangfeng
    Zhou, Dian
    Zeng, Xuan
    2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,