Learning Layer-wise Equivariances Automatically using Gradients

被引:0
|
作者
van der Ouderaa, Tycho F. A. [1 ]
Immer, Alexander [2 ,3 ]
van der Wilk, Mark [1 ,4 ]
机构
[1] Imperial Coll London, Dept Comp, London, England
[2] Swiss Fed Inst Technol, Dept Comp Sci, Zurich, Switzerland
[3] Max Planck Inst Intelligent Syst, Tubingen, Germany
[4] Univ Oxford, Dept Comp Sci, Oxford, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutions encode equivariance symmetries into neural networks leading to better generalisation performance. However, symmetries provide fixed hard constraints on the functions a network can represent, need to be specified in advance, and can not be adapted. Our goal is to allow flexible symmetry constraints that can automatically be learned from data using gradients. Learning symmetry and associated weight connectivity structures from scratch is difficult for two reasons. First, it requires efficient and flexible parameterisations of layer-wise equivariances. Secondly, symmetries act as constraints and are therefore not encouraged by training losses measuring data fit. To overcome these challenges, we improve parameterisations of soft equivariance and learn the amount of equivariance in layers by optimising the marginal likelihood, estimated using differentiable Laplace approximations. The objective balances data fit and model complexity enabling layer-wise symmetry discovery in deep networks. We demonstrate the ability to automatically learn layer-wise equivariances on image classification tasks, achieving equivalent or improved performance over baselines with hard-coded symmetry.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Deep Learning Layer-Wise Learning of Feature Hierarchies
    Schulz, Hannes
    Behnke, Sven
    KUNSTLICHE INTELLIGENZ, 2012, 26 (04): : 357 - 363
  • [2] FedLF: Layer-Wise Fair Federated Learning
    Pan, Zibin
    Li, Chi
    Yu, Fangchen
    Wang, Shuyi
    Wang, Haijin
    Tang, Xiaoying
    Zhao, Junhua
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14527 - 14535
  • [3] Layer-Wise Personalized Federated Learning with Hypernetwork
    Suxia Zhu
    Tianyu Liu
    Guanglu Sun
    Neural Processing Letters, 2023, 55 (9) : 12273 - 12287
  • [4] Layer-Wise Personalized Federated Learning with Hypernetwork
    Zhu, Suxia
    Liu, Tianyu
    Sun, Guanglu
    NEURAL PROCESSING LETTERS, 2023, 55 (09) : 12273 - 12287
  • [5] Explaining Deep Neural Network using Layer-wise Relevance Propagation and Integrated Gradients
    Cik, Ivan
    Rasamoelina, Andrindrasana David
    Mach, Marian
    Sincak, Peter
    2021 IEEE 19TH WORLD SYMPOSIUM ON APPLIED MACHINE INTELLIGENCE AND INFORMATICS (SAMI 2021), 2021, : 381 - 386
  • [6] Efficient Federated Learning Using Layer-Wise Regulation and Momentum Aggregation*
    Zhang, Fan
    Fang, Zekuan
    Li, Yiming
    Chen, Mingsong
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2022, 31 (18)
  • [7] Guided Layer-Wise Learning for Deep Models Using Side Information
    Sulimov, Pavel
    Sukmanova, Elena
    Chereshnev, Roman
    Kertesz-Farkas, Attila
    ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS (AIST 2019), 2020, 1086 : 50 - 61
  • [8] Learning sparse reparameterization with layer-wise continuous sparsification
    Wang, Xiaodong
    Huang, Yaxiang
    Zeng, Xianxian
    Guo, Jianlan
    Chen, Yuqiang
    KNOWLEDGE-BASED SYSTEMS, 2023, 276
  • [9] LSDDL: Layer-Wise Sparsification for Distributed Deep Learning
    Hong, Yuxi
    Han, Peng
    BIG DATA RESEARCH, 2021, 26
  • [10] Dynamic layer-wise sparsification for distributed deep learning
    Zhang, Hao
    Wu, Tingting
    Ma, Zhifeng
    Li, Feng
    Liu, Jie
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 147 : 1 - 15