Invariance Learning in Deep Neural Networks with Differentiable Laplace Approximations

被引:0
|
作者
Immer, Alexander [1 ,2 ]
van der Ouderaa, Tycho F. A. [3 ]
Ratsch, Gunnar [1 ]
Fortuin, Vincent [1 ,4 ]
van der Wilk, Mark [3 ]
机构
[1] Swiss Fed Inst Technol, Dept Comp Sci, Zurich, Switzerland
[2] Max Planck Inst Intelligent Syst, Tubingen, Germany
[3] Imperial Coll London, Dept Comp, London, England
[4] Univ Cambridge, Dept Engn, Cambridge, England
基金
瑞士国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data augmentation is commonly applied to improve performance of deep learning by enforcing the knowledge that certain transformations on the input preserve the output. Currently, the data augmentation parameters are chosen by human effort and costly cross-validation, which makes it cumbersome to apply to new datasets. We develop a convenient gradient-based method for selecting the data augmentation without validation data during training of a deep neural network. Our approach relies on phrasing data augmentation as an invariance in the prior distribution on the functions of a neural network, which allows us to learn it using Bayesian model selection. This has been shown to work in Gaussian processes, but not yet for deep neural networks. We propose a differentiable Kronecker-factored Laplace approximation to the marginal likelihood as our objective, which can be optimised without human supervision or validation data. We show that our method can successfully recover invariances present in the data, and that this improves generalisation and data efficiency on image datasets.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Riemannian Laplace approximations for Bayesian neural networks
    Bergamin, Federico
    Moreno-Munoz, Pablo
    Hauberg, Soren
    Arvanitidis, Georgios
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [2] Deep neural networks for rotation-invariance approximation and learning
    Chui, Charles K.
    Lin, Shao-Bo
    Zhou, Ding-Xuan
    ANALYSIS AND APPLICATIONS, 2019, 17 (05) : 737 - 772
  • [3] Contrast Invariance in Deep Neural Networks
    Akbarinia, Arash
    Gegenfurtner, Karl
    PERCEPTION, 2019, 48 : 51 - 51
  • [4] Differentiable neural architecture learning for efficient neural networks
    Guo, Qingbei
    Wu, Xiao-Jun
    Kittler, Josef
    Feng, Zhiquan
    PATTERN RECOGNITION, 2022, 126
  • [5] Exploiting Invariance in Training Deep Neural Networks
    Ye, Chengxi
    Zhou, Xiong
    McKinney, Tristan
    Liu, Yanfeng
    Zhou, Qinggang
    Zhdanov, Fedor
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8849 - 8856
  • [6] Permutation Invariance of Deep Neural Networks with ReLUs
    Mukhopadhyay, Diganta
    Madhukar, Kumar
    Srivas, Mandayam
    NASA FORMAL METHODS (NFM 2022), 2022, 13260 : 318 - 337
  • [7] Invariance of object detection in untrained deep neural networks
    Cheon, Jeonghwan
    Baek, Seungdae
    Paik, Se-Bum
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2022, 16
  • [8] Multifingered Grasp Planning via Inference in Deep Neural Networks: Outperforming Sampling by Learning Differentiable Models
    Lu, Qingkai
    Van der Merwe, Mark
    Sundaralingam, Balakumar
    Hermans, Tucker
    IEEE ROBOTICS & AUTOMATION MAGAZINE, 2020, 27 (02) : 55 - 65
  • [9] Approximations with deep neural networks in Sobolev time-space
    Abdeljawad, Ahmed
    Grohs, Philipp
    ANALYSIS AND APPLICATIONS, 2022, 20 (03) : 499 - 541
  • [10] Online Deep Learning: Learning Deep Neural Networks on the Fly
    Sahoo, Doyen
    Pham, Quang
    Lu, Jing
    Hoi, Steven C. H.
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2660 - 2666