Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning

被引:0
|
作者
Ji, Kaiyi [1 ]
Yang, Junjie [1 ]
Liang, Yingbin [1 ]
机构
[1] Ohio State Univ, Dept Elect & Comp Engn, Columbus, OH 43210 USA
基金
美国国家科学基金会;
关键词
Computational complexity; convergence rate; finite-sum; meta-learning; multi-step MAML; nonconvex; resampling;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As a popular meta-learning approach, the model-agnostic meta-learning (MAML) algo-rithm has been widely used due to its simplicity and effectiveness. However, the conver-gence of the general multi-step MAML still remains unexplored. In this paper, we develop a new theoretical framework to provide such convergence guarantee for two types of objective functions that are of interest in practice: (a) resampling case (e.g., reinforcement learning), where loss functions take the form in expectation and new data are sampled as the algo-rithm runs; and (b) finite-sum case (e.g., supervised learning), where loss functions take the finite-sum form with given samples. For both cases, we characterize the convergence rate and the computational complexity to attain an epsilon-accurate solution for multi-step MAML in the general nonconvex setting. In particular, our results suggest that an inner-stage stepsize needs to be chosen inversely proportional to the number N of inner-stage steps in order for N-step MAML to have guaranteed convergence. From the technical perspective, we develop novel techniques to deal with the nested structure of the meta gradient for multi-step MAML, which can be of independent interest.
引用
收藏
页数:41
相关论文
共 50 条
  • [41] Model-agnostic multi-stage loss optimization meta learning
    Yao, Xiao
    Zhu, Jianlong
    Huo, Guanying
    Xu, Ning
    Liu, Xiaofeng
    Zhang, Ce
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (08) : 2349 - 2363
  • [42] Model-agnostic multi-stage loss optimization meta learning
    Xiao Yao
    Jianlong Zhu
    Guanying Huo
    Ning Xu
    Xiaofeng Liu
    Ce Zhang
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 2349 - 2363
  • [43] Model-Agnostic Learning to Meta-Learn
    Devos, Arnout
    Dandi, Yatin
    NEURIPS 2020 WORKSHOP ON PRE-REGISTRATION IN MACHINE LEARNING, VOL 148, 2020, 148 : 155 - 175
  • [44] Convolutional Shrinkage Neural Networks Based Model-Agnostic Meta-Learning for Few-Shot Learning
    Yunpeng He
    Chuanzhi Zang
    Peng Zeng
    Qingwei Dong
    Ding Liu
    Yuqi Liu
    Neural Processing Letters, 2023, 55 : 505 - 518
  • [45] Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning
    Zhang, Ce
    Yao, Xiao
    Shi, Changfeng
    Gu, Min
    MULTIMEDIA SYSTEMS, 2023,
  • [46] Fast Adaptive Active Noise Control Based on Modified Model-Agnostic Meta-Learning Algorithm
    Shi, Dongyuan
    Gan, Woon-Seng
    Lam, Bhan
    Ooi, Kenneth
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 593 - 597
  • [47] Model-Agnostic Meta-Learning for EEG Motor Imagery Decoding in Brain-Computer-Interfacing
    Li, Denghao
    Ortega, Pablo
    Wei, Xiaoxi
    Faisal, Aldo
    2021 10TH INTERNATIONAL IEEE/EMBS CONFERENCE ON NEURAL ENGINEERING (NER), 2021, : 527 - 530
  • [48] Convolutional Shrinkage Neural Networks Based Model-Agnostic Meta-Learning for Few-Shot Learning
    He, Yunpeng
    Zang, Chuanzhi
    Zeng, Peng
    Dong, Qingwei
    Liu, Ding
    Liu, Yuqi
    NEURAL PROCESSING LETTERS, 2023, 55 (01) : 505 - 518
  • [49] Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning
    Zhang, Ce
    Yao, Xiao
    Shi, Changfeng
    Gu, Min
    MULTIMEDIA SYSTEMS, 2023, 29 (06) : 3169 - 3177
  • [50] Learning From Visual Demonstrations via Replayed Task-Contrastive Model-Agnostic Meta-Learning
    Hu, Ziye
    Li, Wei
    Gan, Zhongxue
    Guo, Weikun
    Zhu, Jiwei
    Wen, James Zhiqing
    Zhou, Decheng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (12) : 8756 - 8767