Minimax Lower Bounds for Transfer Learning with Linear and One-hidden Layer Neural Networks

被引:0
|
作者
Kalan, Seyed Mohammadreza Mousavi [1 ]
Fabian, Zalan [1 ]
Avestimehr, Salman [1 ]
Soltanolkotabi, Mahdi [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect Engn, Los Angeles, CA 90089 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transfer learning has emerged as a powerful technique for improving the performance of machine learning models on new domains where labeled training data may be scarce. In this approach a model trained for a source task, where plenty of labeled training data is available, is used as a starting point for training a model on a related target task with only few labeled training data. Despite recent empirical success of transfer learning approaches, the benefits and fundamental limits of transfer learning are poorly understood. In this paper we develop a statistical minimax framework to characterize the fundamental limits of transfer learning in the context of regression with linear and one-hidden layer neural network models. Specifically, we derive a lower-bound for the target generalization error achievable by any algorithm as a function of the number of labeled source and target data as well as appropriate notions of similarity between the source and target tasks. Our lowerbound provides new insights into the benefits and limitations of transfer learning. We further corroborate our theoretical finding with various experiments.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Gradient Descent for One-Hidden-Layer Neural Networks: Polynomial Convergence and SQ Lower Bounds
    Vempala, Santosh
    Wilmes, John
    CONFERENCE ON LEARNING THEORY, VOL 99, 2019, 99
  • [2] Superpolynomial Lower Bounds for Learning One-Layer Neural Networks using Gradient Descent
    Goel, Surbhi
    Gollakota, Aravind
    Jin, Zhihan
    Karmalkar, Sushrut
    Klivans, Adam
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [3] Superpolynomial Lower Bounds for Learning One-Layer Neural Networks using Gradient Descent
    Goel, Surbhi
    Gollakota, Aravind
    Jin, Zhihan
    Karmalkar, Sushrut
    Klivans, Adam
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [4] Optimizing One-hidden Layer Neural Network Design Using Evolutionary Programming
    Sulaiman, S. I.
    Rahman, T. K. Abdul
    Musirin, I.
    CSPA: 2009 5TH INTERNATIONAL COLLOQUIUM ON SIGNAL PROCESSING AND ITS APPLICATIONS, PROCEEDINGS, 2009, : 288 - 293
  • [5] Efficiently Learning One Hidden Layer Neural Networks From Queries
    Chen, Sitan
    Klivans, Adam R.
    Meka, Raghu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [6] Strong Minimax Lower Bounds for Learning
    András Antos
    Gábor Lugosi
    Machine Learning, 1998, 30 : 31 - 56
  • [7] Strong minimax lower bounds for learning
    Antos, A
    Lugosi, G
    MACHINE LEARNING, 1998, 30 (01) : 31 - 56
  • [8] Minimax Lower Bounds for Linear Independence Testing
    Ramdas, Aaditya
    Isenberg, David
    Singh, Aarti
    Wasserman, Larry
    2016 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY, 2016, : 965 - 969
  • [9] The One-Hidden Layer Non-parametric Bayesian Kernel Machine
    Chatzis, Sotirios P.
    Korkinof, Dimitrios
    Demiris, Yiannis
    2011 23RD IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2011), 2011, : 825 - 831
  • [10] Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks
    Cao, Yuan
    Gu, Quanquan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32