Invariance Learning in Deep Neural Networks with Differentiable Laplace Approximations

被引:0
|
作者
Immer, Alexander [1 ,2 ]
van der Ouderaa, Tycho F. A. [3 ]
Ratsch, Gunnar [1 ]
Fortuin, Vincent [1 ,4 ]
van der Wilk, Mark [3 ]
机构
[1] Swiss Fed Inst Technol, Dept Comp Sci, Zurich, Switzerland
[2] Max Planck Inst Intelligent Syst, Tubingen, Germany
[3] Imperial Coll London, Dept Comp, London, England
[4] Univ Cambridge, Dept Engn, Cambridge, England
基金
瑞士国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data augmentation is commonly applied to improve performance of deep learning by enforcing the knowledge that certain transformations on the input preserve the output. Currently, the data augmentation parameters are chosen by human effort and costly cross-validation, which makes it cumbersome to apply to new datasets. We develop a convenient gradient-based method for selecting the data augmentation without validation data during training of a deep neural network. Our approach relies on phrasing data augmentation as an invariance in the prior distribution on the functions of a neural network, which allows us to learn it using Bayesian model selection. This has been shown to work in Gaussian processes, but not yet for deep neural networks. We propose a differentiable Kronecker-factored Laplace approximation to the marginal likelihood as our objective, which can be optimised without human supervision or validation data. We show that our method can successfully recover invariances present in the data, and that this improves generalisation and data efficiency on image datasets.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Differentiable Dynamic Wirings for Neural Networks
    Yuan, Kun
    Li, Quanquan
    Guo, Shaopeng
    Chen, Dapeng
    Zhou, Aojun
    Yu, Fengwei
    Liu, Ziwei
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 317 - 326
  • [32] Deep Differentiable Logic Gate Networks
    Petersen, Felix
    Borgelt, Christian
    Kuehne, Hilde
    Deussen, Oliver
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [33] Differentiable Learning-to-Group Channels via Groupable Convolutional Neural Networks
    Zhang, Zhaoyang
    Li, Jingyu
    Shao, Wenqi
    Peng, Zhanglin
    Zhang, Ruimao
    Wang, Xiaogang
    Luo, Ping
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 3541 - 3550
  • [34] Inspecting the behaviour of Deep Learning Neural Networks
    Duer, Alexander
    Filzmoser, Peter
    Rauber, Andreas
    ERCIM NEWS, 2019, (116): : 18 - 19
  • [35] Piecewise linear neural networks and deep learning
    Qinghua Tao
    Li Li
    Xiaolin Huang
    Xiangming Xi
    Shuning Wang
    Johan A. K. Suykens
    Nature Reviews Methods Primers, 2
  • [36] Learning deep neural networks for node classification
    Li, Bentian
    Pi, Dechang
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 137 : 324 - 334
  • [37] Abstraction Hierarchy in Deep Learning Neural Networks
    Ilin, Roman
    Watson, Thomas
    Kozma, Robert
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 768 - 774
  • [38] Deep Neural Networks for Learning Graph Representations
    Cao, Shaosheng
    Lu, Wei
    Xu, Qiongkai
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 1145 - 1152
  • [39] Advances in Machine Learning and Deep Neural Networks
    Chellappa, Rama
    Theodoridis, Sergios
    van Schaik, Andre
    PROCEEDINGS OF THE IEEE, 2021, 109 (05) : 607 - 611
  • [40] Deep learning with coherent VCSEL neural networks
    Zaijun Chen
    Alexander Sludds
    Ronald Davis
    Ian Christen
    Liane Bernstein
    Lamia Ateshian
    Tobias Heuser
    Niels Heermeier
    James A. Lott
    Stephan Reitzenstein
    Ryan Hamerly
    Dirk Englund
    Nature Photonics, 2023, 17 : 723 - 730