The Sample Complexity of One-Hidden-Layer Neural Networks

被引:0
|
作者
Vardi, Gal [1 ,2 ,3 ]
Shamir, Ohad [3 ]
Srebro, Nathan [1 ]
机构
[1] TTI Chicago, Chicago, IL 60637 USA
[2] Hebrew Univ Jerusalem, Jerusalem, Israel
[3] Weizmann Inst Sci, Rehovot, Israel
基金
欧洲研究理事会;
关键词
BOUNDS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study norm-based uniform convergence bounds for neural networks, aiming at a tight understanding of how these are affected by the architecture and type of norm constraint, for the simple class of scalar-valued one-hidden-layer networks, and inputs bounded in Euclidean norm. We begin by proving that in general, controlling the spectral norm of the hidden layer weight matrix is insufficient to get uniform convergence guarantees (independent of the network width), while a stronger Frobenius norm control is sufficient, extending and improving on previous work. Motivated by the proof constructions, we identify and analyze two important settings where (perhaps surprisingly) a mere spectral norm control turns out to be sufficient: First, when the network's activation functions are sufficiently smooth (with the result extending to deeper networks); and second, for certain types of convolutional networks. In the latter setting, we study how the sample complexity is additionally affected by parameters such as the amount of overlap between patches and the overall number of patches.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks
    Cao, Yuan
    Gu, Quanquan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Large deviations of one-hidden-layer neural networks
    Hirsch, Christian
    Willhalm, Daniel
    STOCHASTICS AND DYNAMICS, 2024, 24 (08)
  • [3] Recovery Guarantees for One-hidden-layer Neural Networks
    Zhong, Kai
    Song, Zhao
    Jain, Prateek
    Bartlett, Peter L.
    Dhillon, Inderjit S.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [4] Analysis of one-hidden-layer Neural Networks via the Resolvent Method
    Piccolo, Vanessa
    Schroder, Dominik
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] Distributed Parameter Estimation in Randomized One-hidden-layer Neural Networks
    Wang, Yinsong
    Shahrampour, Shahin
    2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 737 - 742
  • [6] Learning One-hidden-layer Neural Networks under General Input Distributions
    Gao, Weihao
    Makkuva, Ashok Vardhan
    Oh, Sewoong
    Viswanath, Pramod
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [7] Guaranteed Recovery of One-Hidden-Layer Neural Networks via Cross Entropy
    Fu, Haoyu
    Chi, Yuejie
    Liang, Yingbin
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 : 3225 - 3235
  • [8] Incremental approximation by one-hidden-layer neural networks: Discrete functions rapprochement
    Beliczynski, B
    ISIE'96 - PROCEEDINGS OF THE IEEE INTERNATIONAL SYMPOSIUM ON INDUSTRIAL ELECTRONICS, VOLS 1 AND 2, 1996, : 392 - 397
  • [9] On the landscape of one-hidden-layer sparse networks and beyond
    Lin, Dachao
    Sun, Ruoyu
    Zhang, Zhihua
    ARTIFICIAL INTELLIGENCE, 2022, 309
  • [10] Learning Narrow One-Hidden-Layer ReLU Networks
    Chen, Sitan
    Dou, Zehao
    Goel, Surbhi
    Klivans, Adam
    Meka, Raghu
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195