Data Symmetries and Learning in Fully Connected Neural Networks

被引:1
|
作者
Anselmi, Fabio [1 ,2 ]
Manzoni, Luca [1 ]
D'onofrio, Alberto [1 ]
Rodriguez, Alex [1 ]
Caravagna, Giulio [1 ]
Bortolussi, Luca [1 ]
Cairoli, Francesca [1 ]
机构
[1] Univ Trieste, Dept Math & Geosci, I-34127 Trieste, Italy
[2] MIT, McGovern Inst, Ctr Brains Minds & Machines, Cambridge, MA 02139 USA
关键词
Orbits; Finite element analysis; Reflection; Task analysis; Complexity theory; Artificial neural networks; Machine learning; symmetry invariance; equivariance; INVARIANT OBJECT RECOGNITION; PATTERN-RECOGNITION; SIZE-INVARIANT; SHIFT;
D O I
10.1109/ACCESS.2023.3274938
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Symmetries in the data and how they constrain the learned weights of modern deep networks is still an open problem. In this work we study the simple case of fully connected shallow non-linear neural networks and consider two types of symmetries: full dataset symmetries where the dataset X is mapped into itself by any transformation g, i.e. gX = X or single data point symmetries where gx = x, x ? X. We prove and experimentally confirm that symmetries in the data are directly inherited at the level of the network's learned weights and relate these findings with the common practice of data augmentation in modern machine learning. Finally, we show how symmetry constraints have a profound impact on the spectrum of the learned weights, an aspect of the so-called network implicit bias.
引用
收藏
页码:47282 / 47290
页数:9
相关论文
共 50 条
  • [41] Differentiable homotopy methods for gradually reinforcing the training of fully connected neural networks
    Li, Peixuan
    Li, Yuanbo
    NEUROCOMPUTING, 2024, 605
  • [42] Gradient rectified parameter unit of the fully connected layer in convolutional neural networks
    Zheng, Tianyou
    Wang, Qiang
    Shen, Yue
    Lin, Xiaotian
    KNOWLEDGE-BASED SYSTEMS, 2022, 248
  • [43] FULLY CONNECTED NEURAL NETWORKS WITH SELF-CONTROL OF NOISE-LEVELS
    LEWENSTEIN, M
    NOWAK, A
    PHYSICAL REVIEW LETTERS, 1989, 62 (02) : 225 - 228
  • [44] Open loop stability criterion for layered and fully-connected neural networks
    Snyder, M.M.
    Ferry, D.K.
    Neural Networks, 1988, 1 (1 SUPPL)
  • [45] Compressing fully connected layers of deep neural networks using permuted features
    Nagaraju, Dara
    Chandrachoodan, Nitin
    IET COMPUTERS AND DIGITAL TECHNIQUES, 2023, 17 (3-4): : 149 - 161
  • [46] Kurtosis-Based CRTRL Algorithms for Fully Connected Recurrent Neural Networks
    Menguc, Engin Cemal
    Acir, Nurettin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (12) : 6123 - 6131
  • [47] Impact of fully connected layers on performance of convolutional neural networks for image classification
    Basha, S. H. Shabbeer
    Dubey, Shiv Ram
    Pulabaigari, Viswanath
    Mukherjee, Snehasis
    NEUROCOMPUTING, 2020, 378 (378) : 112 - 119
  • [48] Structure and performance of fully connected neural networks: Emerging complex network properties
    Scabini, Leonardo F. S.
    Bruno, Odemir M.
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2023, 615
  • [49] Detecting symmetries with neural networks
    Krippendorf, Sven
    Syvaeri, Marc
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2021, 2 (01):
  • [50] On the BP Training Algorithm of Fuzzy Neural Networks (FNNs) via Its Equivalent Fully Connected Neural Networks (FFNNs)
    Wang, Jing
    Wang, Chi-Hsu
    Chen, C. L. Philip
    2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 1376 - 1381