Tight Bounds on the Smallest Eigenvalue of the Neural Tangent Kernel for Deep ReLU Networks

被引:0
|
作者
Nguyen, Quynh [1 ]
Mondelli, Marco [2 ]
Montufar, Guido [1 ,3 ]
机构
[1] MPI MIS, Leipzig, Germany
[2] IST Austria, Klosterneuburg, Austria
[3] Univ Calif Los Angeles, Los Angeles, CA 90024 USA
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A recent line of work has analyzed the theoretical properties of deep neural networks via the Neural Tangent Kernel (NTK). In particular, the smallest eigenvalue of the NTK has been related to the memorization capacity, the global convergence of gradient descent algorithms and the generalization of deep nets. However, existing results either provide bounds in the two-layer setting or assume that the spectrum of the NTK matrices is bounded away from 0 for multi-layer networks. In this paper, we provide tight bounds on the smallest eigenvalue of NTK matrices for deep ReLU nets, both in the limiting case of infinite widths and for finite widths. In the finite-width setting, the network architectures we consider are fairly general: we require the existence of a wide layer with roughly order of N neurons, N being the number of data samples; and the scaling of the remaining layer widths is arbitrary (up to logarithmic factors). To obtain our results, we analyze various quantities of independent interest: we give lower bounds on the smallest singular value of hidden feature matrices, and upper bounds on the Lipschitz constant of input-output feature maps.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Neural Tangent Kernel Analysis of Deep Narrow Neural Networks
    Lee, Jongmin
    Choi, Joo Young
    Ryu, Ernest K.
    No, Albert
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [2] On the Neural Tangent Kernel of Deep Networks with Orthogonal Initialization
    Huang, Wei
    Du, Weitao
    Da Xu, Richard Yi
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2577 - 2583
  • [3] On the Error Bounds for ReLU Neural Networks
    Katende, Ronald
    Kasumba, Henry
    Kakuba, Godwin
    Mango, John
    IAENG International Journal of Applied Mathematics, 2024, 54 (12) : 2602 - 2611
  • [4] Spectral Analysis of the Neural Tangent Kernel for Deep Residual Networks
    Belfer, Yuval
    Geifman, Amnon
    Galun, Meirav
    Basri, Ronen
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 49
  • [5] Error bounds for approximations with deep ReLU neural networks in Ws,p norms
    Guehring, Ingo
    Kutyniok, Gitta
    Petersen, Philipp
    ANALYSIS AND APPLICATIONS, 2020, 18 (05) : 803 - 859
  • [6] Error bounds for approximations with deep ReLU networks
    Yarotsky, Dmitry
    NEURAL NETWORKS, 2017, 94 : 103 - 114
  • [7] Towards Lower Bounds on the Depth of ReLU Neural Networks
    Hertrich, Christoph
    Basu, Amitabh
    Di Summa, Marco
    Skutella, Martin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [8] TOWARDS LOWER BOUNDS ON THE DEPTH OF RELU NEURAL NETWORKS*
    Hertrich, Christoph
    Basu, Amitabh
    Di Summa, Marco
    Skutella, Martin
    SIAM JOURNAL ON DISCRETE MATHEMATICS, 2023, 37 (02) : 997 - 1029
  • [9] Neural Tangent Kernel: Convergence and Generalization in Neural Networks
    Jacot, Arthur
    Gabriel, Franck
    Hongler, Clement
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [10] "Lossless" Compression of Deep Neural Networks: A High-dimensional Neural Tangent Kernel Approach
    Gu, Lingyu
    Du, Yongqi
    Zhang, Yuan
    Xie, Di
    Pu, Shiliang
    Qiu, Robert C.
    Liao, Zhenyu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,