Neural (Tangent Kernel) Collapse

被引:0
|
作者
Seleznova, Mariia [1 ]
Weitzner, Dana [2 ]
Giryes, Raja [2 ]
Kutyniok, Gitta [1 ]
Chou, Hung-Hsu [1 ]
机构
[1] Ludwig Maximilians Univ Munchen, Munich, Germany
[2] Tel Aviv Univ, Tel Aviv, Israel
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work bridges two important concepts: the Neural Tangent Kernel (NTK), which captures the evolution of deep neural networks (DNNs) during training, and the Neural Collapse (NC) phenomenon, which refers to the emergence of symmetry and structure in the last-layer features of well-trained classification DNNs. We adopt the natural assumption that the empirical NTK develops a block structure aligned with the class labels, i.e., samples within the same class have stronger correlations than samples from different classes. Under this assumption, we derive the dynamics of DNNs trained with mean squared (MSE) loss and break them into interpretable phases. Moreover, we identify an invariant that captures the essence of the dynamics, and use it to prove the emergence of NC in DNNs with block-structured NTK. We provide large-scale numerical experiments on three common DNN architectures and three benchmark datasets to support our theory.
引用
收藏
页数:31
相关论文
共 50 条
  • [41] What Can the Neural Tangent Kernel Tell Us About Adversarial Robustness?
    Tsilivis, Nikolaos
    Kempe, Julia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [42] On the Generalization Power of Overfitted Two-Layer Neural Tangent Kernel Models
    Ju, Peizhong
    Lin, Xiaojun
    Shroff, Ness B.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [43] DNN-based Topology Optimisation: Spatial Invariance and Neural Tangent Kernel
    Dupuis, Benjamin
    Jacot, Arthur
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [44] Tensor Programs IIb: Architectural Universality of Neural Tangent Kernel Training Dynamics
    Yang, Greg
    Littwin, Etai
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [45] Tight Bounds on the Smallest Eigenvalue of the Neural Tangent Kernel for Deep ReLU Networks
    Nguyen, Quynh
    Mondelli, Marco
    Montufar, Guido
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [46] How does a kernel based on gradients of infinite-width neural networks come to be widely used: a review of the neural tangent kernel
    Tan, Yiqiao
    Liu, Haizhong
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2024, 13 (01)
  • [47] How does a kernel based on gradients of infinite-width neural networks come to be widely used: a review of the neural tangent kernel
    Yiqiao Tan
    Haizhong Liu
    International Journal of Multimedia Information Retrieval, 2024, 13
  • [48] Neural Tangent Kernel Beyond the Infinite-Width Limit: Effects of Depth and Initialization
    Seleznova, Mariia
    Kutyniok, Gitta
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 19522 - 19560
  • [49] Label-Aware Neural Tangent Kernel: Toward Better Generalization and Local Elasticity
    Chen, Shuxiao
    He, Hangfeng
    Su, Weijie J.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [50] Multi-angle Support Vector Survival Analysis with Neural Tangent Kernel Study
    Yue-jing Zhai
    Yu Zhang
    Hai-zhong Liu
    Zhong-rong Zhang
    Arabian Journal for Science and Engineering, 2023, 48 : 10267 - 10284