Provable Guarantees for Neural Networks via Gradient Feature Learning

被引:0
|
作者
Shi, Zhenmei [1 ]
Wei, Junyi [1 ]
Liang, Yingyu [1 ]
机构
[1] Univ Wisconsin, Madison, WI 53706 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks have achieved remarkable empirical performance, while the current theoretical analysis is not adequate for understanding their success, e.g., the Neural Tangent Kernel approach fails to capture their key feature learning ability, while recent analyses on feature learning are typically problem-specific. This work proposes a unified analysis framework for two-layer networks trained by gradient descent. The framework is centered around the principle of feature learning from gradients, and its effectiveness is demonstrated by applications in several prototypical problems such as mixtures of Gaussians and parity functions. The framework also sheds light on interesting network learning phenomena such as feature learning beyond kernels and the lottery ticket hypothesis.
引用
收藏
页数:71
相关论文
共 50 条
  • [31] Recovery Guarantees of Unsupervised Neural Networks for Inverse Problems trained with Gradient Descent
    Buskulic, Nathan
    Queau, Yvain
    Fadili, Jalal
    32ND EUROPEAN SIGNAL PROCESSING CONFERENCE, EUSIPCO 2024, 2024, : 1806 - 1810
  • [32] On the Provable Generalization of Recurrent Neural Networks
    Wang, Lifu
    Shen, Bo
    Hu, Bo
    Cao, Xing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [33] Minimum budget for misinformation detection in online social networks with provable guarantees
    Canh V. Pham
    Dung V. Pham
    Bao Q. Bui
    Anh V. Nguyen
    Optimization Letters, 2022, 16 : 515 - 544
  • [34] Theoretically Provable Spiking Neural Networks
    Zhang, Shao-Qun
    Zhou, Zhi-Hua
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [35] Minimum budget for misinformation detection in online social networks with provable guarantees
    Pham, Canh V.
    Pham, Dung V.
    Bui, Bao Q.
    Nguyen, Anh V.
    OPTIMIZATION LETTERS, 2022, 16 (02) : 515 - 544
  • [36] Fast and Provable Algorithms for Learning Two-Layer Polynomial Neural Networks
    Soltani, Mohammadreza
    Hegde, Chinmay
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (13) : 3361 - 3371
  • [37] Provable convergence of Nesterov?s accelerated gradient method for over-parameterized neural networks
    Liu, Xin
    Pan, Zhisong
    Tao, Wei
    KNOWLEDGE-BASED SYSTEMS, 2022, 251
  • [38] A Reinforcement Learning Approach to Price Cloud Resources With Provable Convergence Guarantees
    Xie, Hong
    Lui, John C. S.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (12) : 7448 - 7460
  • [39] Gradient Starvation: A Learning Proclivity in Neural Networks
    Pezeshki, Mohammad
    Kaba, Sekou-Oumar
    Bengio, Yoshua
    Courville, Aaron
    Precup, Doina
    Lajoie, Guillaume
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [40] Change point detection via feedforward neural networks with theoretical guarantees
    Zhou, Houlin
    Zhu, Hanbing
    Wang, Xuejun
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2024, 193