Minimax Lower Bounds for Transfer Learning with Linear and One-hidden Layer Neural Networks

被引:0
|
作者
Kalan, Seyed Mohammadreza Mousavi [1 ]
Fabian, Zalan [1 ]
Avestimehr, Salman [1 ]
Soltanolkotabi, Mahdi [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect Engn, Los Angeles, CA 90089 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transfer learning has emerged as a powerful technique for improving the performance of machine learning models on new domains where labeled training data may be scarce. In this approach a model trained for a source task, where plenty of labeled training data is available, is used as a starting point for training a model on a related target task with only few labeled training data. Despite recent empirical success of transfer learning approaches, the benefits and fundamental limits of transfer learning are poorly understood. In this paper we develop a statistical minimax framework to characterize the fundamental limits of transfer learning in the context of regression with linear and one-hidden layer neural network models. Specifically, we derive a lower-bound for the target generalization error achievable by any algorithm as a function of the number of labeled source and target data as well as appropriate notions of similarity between the source and target tasks. Our lowerbound provides new insights into the benefits and limitations of transfer learning. We further corroborate our theoretical finding with various experiments.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] A novel learning algorithm of single-hidden-layer feedforward neural networks
    Pu, Dong-Mei
    Gao, Da-Qi
    Ruan, Tong
    Yuan, Yu-Bo
    NEURAL COMPUTING & APPLICATIONS, 2017, 28 : S719 - S726
  • [42] Sequential learning artificial fuzzy neural networks (SLAFNN) with single hidden layer
    Rajasekaran, S
    Suresh, D
    Pai, GAV
    NEUROCOMPUTING, 2002, 42 : 287 - 310
  • [43] Learning One-hidden-layer ReLU Networks via Gradient Descent
    Zhang, Xiao
    Yu, Yaodong
    Wang, Lingxiao
    Gu, Quanquan
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [44] Neural Networks with Marginalized Corrupted Hidden Layer
    Li, Yanjun
    Xin, Xin
    Guo, Ping
    NEURAL INFORMATION PROCESSING, PT III, 2015, 9491 : 506 - 514
  • [45] Guaranteed Recovery of One-Hidden-Layer Neural Networks via Cross Entropy
    Fu, Haoyu
    Chi, Yuejie
    Liang, Yingbin
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 : 3225 - 3235
  • [46] Incremental approximation by one-hidden-layer neural networks: Discrete functions rapprochement
    Beliczynski, B
    ISIE'96 - PROCEEDINGS OF THE IEEE INTERNATIONAL SYMPOSIUM ON INDUSTRIAL ELECTRONICS, VOLS 1 AND 2, 1996, : 392 - 397
  • [47] Lower Bounds on the Rate of Learning in Social Networks
    Lobel, Ilan
    Acemoglu, Daron
    Dahleh, Munther
    Ozdaglar, Asuman
    2009 AMERICAN CONTROL CONFERENCE, VOLS 1-9, 2009, : 2825 - +
  • [48] LOWER BOUNDS FOR CONTAMINATION BIAS - GLOBALLY MINIMAX VERSUS LOCALLY LINEAR-ESTIMATION
    HE, XM
    SIMPSON, DG
    ANNALS OF STATISTICS, 1993, 21 (01): : 314 - 337
  • [49] TOWARDS LOWER BOUNDS ON THE DEPTH OF RELU NEURAL NETWORKS*
    Hertrich, Christoph
    Basu, Amitabh
    Di Summa, Marco
    Skutella, Martin
    SIAM JOURNAL ON DISCRETE MATHEMATICS, 2023, 37 (02) : 997 - 1029
  • [50] Towards Lower Bounds on the Depth of ReLU Neural Networks
    Hertrich, Christoph
    Basu, Amitabh
    Di Summa, Marco
    Skutella, Martin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,