Neural (Tangent Kernel) Collapse

被引:0
|
作者
Seleznova, Mariia [1 ]
Weitzner, Dana [2 ]
Giryes, Raja [2 ]
Kutyniok, Gitta [1 ]
Chou, Hung-Hsu [1 ]
机构
[1] Ludwig Maximilians Univ Munchen, Munich, Germany
[2] Tel Aviv Univ, Tel Aviv, Israel
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work bridges two important concepts: the Neural Tangent Kernel (NTK), which captures the evolution of deep neural networks (DNNs) during training, and the Neural Collapse (NC) phenomenon, which refers to the emergence of symmetry and structure in the last-layer features of well-trained classification DNNs. We adopt the natural assumption that the empirical NTK develops a block structure aligned with the class labels, i.e., samples within the same class have stronger correlations than samples from different classes. Under this assumption, we derive the dynamics of DNNs trained with mean squared (MSE) loss and break them into interpretable phases. Moreover, we identify an invariant that captures the essence of the dynamics, and use it to prove the emergence of NC in DNNs with block-structured NTK. We provide large-scale numerical experiments on three common DNN architectures and three benchmark datasets to support our theory.
引用
收藏
页数:31
相关论文
共 50 条
  • [31] Benefits of Jointly Training Autoencoders: An Improved Neural Tangent Kernel Analysis
    Nguyen, Thanh V.
    Wong, Raymond K. W.
    Hegde, Chinmay
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2021, 67 (07) : 4669 - 4692
  • [32] GRAPH CONVOLUTIONAL NETWORKS FROM THE PERSPECTIVE OF SHEAVES AND THE NEURAL TANGENT KERNEL
    Gebhart, Thomas
    TOPOLOGICAL, ALGEBRAIC AND GEOMETRIC LEARNING WORKSHOPS 2022, VOL 196, 2022, 196
  • [33] Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel
    Kobayashi, Seijin
    Aceituno, Pau Vilimelis
    von Oswald, Johannes
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [34] When and why PINNs fail to train: A neural tangent kernel perspective
    Wang, Sifan
    Yu, Xinling
    Perdikaris, Paris
    JOURNAL OF COMPUTATIONAL PHYSICS, 2022, 449
  • [35] Quantum tangent kernel
    Shirai, Norihito
    Kubo, Kenji
    Mitarai, Kosuke
    Fujii, Keisuke
    PHYSICAL REVIEW RESEARCH, 2024, 6 (03):
  • [36] "Lossless" Compression of Deep Neural Networks: A High-dimensional Neural Tangent Kernel Approach
    Gu, Lingyu
    Du, Yongqi
    Zhang, Yuan
    Xie, Di
    Pu, Shiliang
    Qiu, Robert C.
    Liao, Zhenyu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [37] Unsupervised Shape Completion via Deep Prior in the Neural Tangent Kernel Perspective
    Chu, Lei
    Pan, Hao
    Wang, Wenping
    ACM TRANSACTIONS ON GRAPHICS, 2021, 40 (03):
  • [38] DoA Estimation Using Neural Tangent Kernel under Electromagnetic Mutual Coupling
    Wang, Qifeng
    Hu, Xiaolin
    Deng, Xiaobao
    Buris, Nicholas E.
    ELECTRONICS, 2021, 10 (09)
  • [39] On the Generalization Power of the Overfitted Three-Layer Neural Tangent Kernel Model
    Ju, Peizhong
    Lin, Xiaojun
    Shroff, Ness B.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [40] Beyond Double Ascent via Recurrent Neural Tangent Kernel in Sequential Recommendation
    Qiu, Ruihong
    Huang, Zi
    Yin, Hongzhi
    2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2022, : 428 - 437