The Sample Complexity of One-Hidden-Layer Neural Networks

被引:0
|
作者
Vardi, Gal [1 ,2 ,3 ]
Shamir, Ohad [3 ]
Srebro, Nathan [1 ]
机构
[1] TTI Chicago, Chicago, IL 60637 USA
[2] Hebrew Univ Jerusalem, Jerusalem, Israel
[3] Weizmann Inst Sci, Rehovot, Israel
基金
欧洲研究理事会;
关键词
BOUNDS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study norm-based uniform convergence bounds for neural networks, aiming at a tight understanding of how these are affected by the architecture and type of norm constraint, for the simple class of scalar-valued one-hidden-layer networks, and inputs bounded in Euclidean norm. We begin by proving that in general, controlling the spectral norm of the hidden layer weight matrix is insufficient to get uniform convergence guarantees (independent of the network width), while a stronger Frobenius norm control is sufficient, extending and improving on previous work. Motivated by the proof constructions, we identify and analyze two important settings where (perhaps surprisingly) a mere spectral norm control turns out to be sufficient: First, when the network's activation functions are sufficiently smooth (with the result extending to deeper networks); and second, for certain types of convolutional networks. In the latter setting, we study how the sample complexity is additionally affected by parameters such as the amount of overlap between patches and the overall number of patches.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Feedforward Neural Networks with a Hidden Layer Regularization Method
    Alemu, Habtamu Zegeye
    Wu, Wei
    Zhao, Junhong
    SYMMETRY-BASEL, 2018, 10 (10):
  • [42] Feedforward networks with one hidden layer and their rates of approximation
    Hlavackova, K
    UKACC INTERNATIONAL CONFERENCE ON CONTROL '98, VOLS I&II, 1998, : 727 - 732
  • [43] Modular Expansion of the Hidden Layer in Single Layer Feedforward Neural Networks
    Tissera, Migel D.
    McDonnell, Mark D.
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 2939 - 2945
  • [44] Multilayer neural networks: One or two hidden layers?
    Brightwell, G
    Kenyon, C
    PaugamMoisy, H
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 9: PROCEEDINGS OF THE 1996 CONFERENCE, 1997, 9 : 148 - 154
  • [45] Collapsing multiple hidden layers in feedforward neural networks to a single hidden layer
    Blue, JL
    Hall, LO
    APPLICATIONS AND SCIENCE OF ARTIFICIAL NEURAL NETWORKS II, 1996, 2760 : 44 - 52
  • [46] Interval computing in neural networks:: One layer interval neural networks
    Patiño-Escarcina, RE
    Bedregal, BRC
    Lyra, A
    INTELLIGENT INFORMATION TECHNOLOGY, PROCEEDINGS, 2004, 3356 : 68 - 75
  • [47] Simultaneous Approximation of Polynomial Functions and Its Derivatives by Feedforward Artificial Neural Networks with One Hidden Layer
    Uzentsova, N. S.
    Sidorov, S. P.
    IZVESTIYA SARATOVSKOGO UNIVERSITETA NOVAYA SERIYA-MATEMATIKA MEKHANIKA INFORMATIKA, 2013, 13 (02): : 14 - 14
  • [48] Sample complexity of hidden subgroup problem
    Ye, Zekun
    Li, Lvzhou
    THEORETICAL COMPUTER SCIENCE, 2022, 922 : 108 - 121
  • [49] DEGREE OF APPROXIMATION BY NEURAL AND TRANSLATION NETWORKS WITH A SINGLE HIDDEN LAYER
    MHASKAR, HN
    MICCHELLI, CA
    ADVANCES IN APPLIED MATHEMATICS, 1995, 16 (02) : 151 - 183
  • [50] Simplicity Bias in 1-Hidden Layer Neural Networks
    Morwani, Depen
    Batra, Jatin
    Jain, Prateek
    Netrapalli, Praneeth
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,