Neural (Tangent Kernel) Collapse

被引:0
|
作者
Seleznova, Mariia [1 ]
Weitzner, Dana [2 ]
Giryes, Raja [2 ]
Kutyniok, Gitta [1 ]
Chou, Hung-Hsu [1 ]
机构
[1] Ludwig Maximilians Univ Munchen, Munich, Germany
[2] Tel Aviv Univ, Tel Aviv, Israel
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work bridges two important concepts: the Neural Tangent Kernel (NTK), which captures the evolution of deep neural networks (DNNs) during training, and the Neural Collapse (NC) phenomenon, which refers to the emergence of symmetry and structure in the last-layer features of well-trained classification DNNs. We adopt the natural assumption that the empirical NTK develops a block structure aligned with the class labels, i.e., samples within the same class have stronger correlations than samples from different classes. Under this assumption, we derive the dynamics of DNNs trained with mean squared (MSE) loss and break them into interpretable phases. Moreover, we identify an invariant that captures the essence of the dynamics, and use it to prove the emergence of NC in DNNs with block-structured NTK. We provide large-scale numerical experiments on three common DNN architectures and three benchmark datasets to support our theory.
引用
收藏
页数:31
相关论文
共 50 条
  • [1] On the Random Conjugate Kernel and Neural Tangent Kernel
    Hu, Zhengmian
    Huang, Heng
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [2] Reverse Engineering the Neural Tangent Kernel
    Simon, James B.
    Anand, Sajant
    DeWeese, Michael R.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [3] A Neural Tangent Kernel Perspective of GANs
    Franceschi, Jean-Yves
    de Bezenac, Emmanuel
    Ayed, Ibrahim
    Chen, Mickael
    Lamprier, Sylvain
    Gallinari, Patrick
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [4] Neural Tangent Kernel: Convergence and Generalization in Neural Networks
    Jacot, Arthur
    Gabriel, Franck
    Hongler, Clement
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [5] Neural Tangent Kernel Empowered Federated Learning
    Yue, Kai
    Jin, Richeng
    Pilgrim, Ryan
    Wong, Chau-Wai
    Baron, Dror
    Dai, Huaiyu
    Proceedings of Machine Learning Research, 2022, 162 : 25783 - 25803
  • [6] Fast Finite Width Neural Tangent Kernel
    Novak, Roman
    Sohl-Dickstein, Jascha
    Schoenholz, Samuel S.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [7] Fast Finite Width Neural Tangent Kernel
    Novak, Roman
    Sohl-Dickstein, Jascha
    Schoenholz, Samuel S.
    arXiv, 2022,
  • [8] Neural Tangent Kernel Empowered Federated Learning
    Yue, Kai
    Jin, Richeng
    Pilgrim, Ryan
    Wong, Chau-Wai
    Baron, Dror
    Dai, Huaiyu
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [9] Neural Tangent Kernel Maximum Mean Discrepancy
    Cheng, Xiuyuan
    Xie, Yao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [10] Spectra of the Conjugate Kernel and Neural Tangent Kernel for Linear-Width Neural Networks
    Fan, Zhou
    Wang, Zhichao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33