Initialization-Dependent Sample Complexity of Linear Predictors and Neural Networks

被引:0
|
作者
Magen, Roey [1 ]
Shamir, Ohad [1 ]
机构
[1] Weizmann Inst Sci, Rehovot, Israel
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We provide several new results on the sample complexity of vector-valued linear predictors (parameterized by a matrix), and more generally neural networks. Focusing on size-independent bounds, where only the Frobenius norm distance of the parameters from some fixed reference matrix W0 is controlled, we show that the sample complexity behavior can be surprisingly different than what we may expect considering the well-studied setting of scalar-valued linear predictors. This also leads to new sample complexity bounds for feed-forward neural networks, tackling some open questions in the literature, and establishing a new convex linear prediction problem that is provably learnable without uniform convergence.
引用
收藏
页数:27
相关论文
共 50 条
  • [21] On the Sample Complexity of the Linear Quadratic Regulator
    Dean, Sarah
    Mania, Horia
    Matni, Nikolai
    Recht, Benjamin
    Tu, Stephen
    FOUNDATIONS OF COMPUTATIONAL MATHEMATICS, 2020, 20 (04) : 633 - 679
  • [22] On the Sample Complexity of the Linear Quadratic Regulator
    Sarah Dean
    Horia Mania
    Nikolai Matni
    Benjamin Recht
    Stephen Tu
    Foundations of Computational Mathematics, 2020, 20 : 633 - 679
  • [23] An overview on weight initialization methods for feedforward neural networks
    de Sousa, Celso A. R.
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 52 - 59
  • [24] Training Linear Neural Networks: Non-Local Convergence and Complexity Results
    Eftekhari, Armin
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [25] Why Lottery Ticket Wins? A Theoretical Perspective of Sample Complexity on Pruned Neural Networks
    Zhang, Shuai
    Wang, Meng
    Liu, Sijia
    Chen, Pin-Yu
    Xiong, Jinjun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [26] Training Linear Neural Networks: Non-Local Convergence and Complexity Results
    Eftekhari, Armin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [27] Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks
    Cao, Yuan
    Gu, Quanquan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [28] Feedforward neural networks initialization based on discriminant learning
    Chumachenko, Kateryna
    Iosifidis, Alexandros
    Gabbouj, Moncef
    NEURAL NETWORKS, 2022, 146 : 220 - 229
  • [29] A New Initialization Method for Artificial Neural Networks: Laplacian
    Catalbas, Burak
    Catalbas, Bahadir
    Morgul, Omer
    2018 26TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2018,
  • [30] Effects of initialization on structure formation and generalization of neural networks
    Shiratsuchi, H
    Gotanda, H
    Inoue, K
    Kumamaru, K
    IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 2644 - 2649