Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning

被引:0
|
作者
Ji, Kaiyi [1 ]
Yang, Junjie [1 ]
Liang, Yingbin [1 ]
机构
[1] Ohio State Univ, Dept Elect & Comp Engn, Columbus, OH 43210 USA
基金
美国国家科学基金会;
关键词
Computational complexity; convergence rate; finite-sum; meta-learning; multi-step MAML; nonconvex; resampling;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As a popular meta-learning approach, the model-agnostic meta-learning (MAML) algo-rithm has been widely used due to its simplicity and effectiveness. However, the conver-gence of the general multi-step MAML still remains unexplored. In this paper, we develop a new theoretical framework to provide such convergence guarantee for two types of objective functions that are of interest in practice: (a) resampling case (e.g., reinforcement learning), where loss functions take the form in expectation and new data are sampled as the algo-rithm runs; and (b) finite-sum case (e.g., supervised learning), where loss functions take the finite-sum form with given samples. For both cases, we characterize the convergence rate and the computational complexity to attain an epsilon-accurate solution for multi-step MAML in the general nonconvex setting. In particular, our results suggest that an inner-stage stepsize needs to be chosen inversely proportional to the number N of inner-stage steps in order for N-step MAML to have guaranteed convergence. From the technical perspective, we develop novel techniques to deal with the nested structure of the meta gradient for multi-step MAML, which can be of independent interest.
引用
收藏
页数:41
相关论文
共 50 条
  • [1] Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning
    Ji, Kaiyi
    Yang, Junjie
    Liang, Yingbin
    Journal of Machine Learning Research, 2022, 23
  • [2] Bayesian Model-Agnostic Meta-Learning
    Yoon, Jaesik
    Kim, Taesup
    Dia, Ousmane
    Kim, Sungwoong
    Bengio, Yoshua
    Ahn, Sungjin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [3] Probabilistic Model-Agnostic Meta-Learning
    Finn, Chelsea
    Xu, Kelvin
    Levine, Sergey
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [4] Trapezoidal Step Scheduler for Model-Agnostic Meta-Learning in Medical Imaging
    Voon, Wingates
    Hum, Yan Chai
    Tee, Yee Kai
    Yap, Wun-She
    Lai, Khin Wee
    Nisar, Humaira
    Mokayed, Hamam
    PATTERN RECOGNITION, 2025, 161
  • [5] Personalized Federated Learning with Theoretical Guarantees: A Model-Agnostic Meta-Learning Approach
    Fallah, Alireza
    Mokhtari, Aryan
    Ozdaglar, Asuman
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] Knowledge Distillation for Model-Agnostic Meta-Learning
    Zhang, Min
    Wang, Donglin
    Gai, Sibo
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1355 - 1362
  • [7] Meta weight learning via model-agnostic meta-learning
    Xu, Zhixiong
    Chen, Xiliang
    Tang, Wei
    Lai, Jun
    Cao, Lei
    NEUROCOMPUTING, 2021, 432 : 124 - 132
  • [8] On the Convergence Theory of Gradient-Based Model-Agnostic Meta-Learning Algorithms
    Fallah, Alireza
    Mokhtari, Aryan
    Ozdaglar, Asuman
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 1082 - 1091
  • [9] PEER-TO-PEER MODEL-AGNOSTIC META-LEARNING
    Qureshi, Muhammad I.
    Khan, Usman A.
    2024 IEEE 13RD SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING WORKSHOP, SAM 2024, 2024,
  • [10] Task-Robust Model-Agnostic Meta-Learning
    Collins, Liam
    Mokhtari, Aryan
    Shakkottai, Sanjay
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33