Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning

被引:0
|
作者
Zhang, Ce [1 ]
Yao, Xiao [1 ]
Shi, Changfeng [2 ]
Gu, Min [3 ]
机构
[1] Hohai Univ, Coll IoT Engn, Changzhou, Peoples R China
[2] HoHai Univ, Business Sch, ChangZhou, Peoples R China
[3] First Peoples Hosp Changzhou, Changzhou, Peoples R China
关键词
Machine learning; Few-shot learning; K-FAC; Second-order optimization; Adaptive learning rate;
D O I
10.1007/s00530-023-01159-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Model-agnostic meta-learning (MAML) highlights the ability to quickly adapt to new tasks with only a small amount of labeled training data among many few-shot learning algorithms. However, the computational complexity is high, because the MAML algorithm generates a large number of second-order parameters in the secondary gradient update. In addition, due to the non-convex nature of the neural network, the loss landscape has many flat areas, leading to slow convergence during training, and excessively long training. In this paper, a second-order optimization method called Kronecker-factored Approximate Curvature (K-FAC) is proposed to approximate Natural Gradient Descent. K-FAC reduces the computational complexity by approximating the large matrix of the Fisher information as the Kronecker product of two much smaller matrices, and the second-order parameter information is fully utilized to accelerate the convergence. Moreover, in order to solve the problem that Natural Gradient Descent is sensitive to the learning rate, this paper proposes Kronecker-factored Approximate Curvature with adaptive learning rate for optimizing model-agnostic meta-learning (AK-MAML), which automatically adjusts the learning rate according to the curvature and improves the efficiency of training. Experimental results show that AK-MAML has the ability of faster convergence, lower computation, and higher accuracy on few-shot datasets.
引用
收藏
页码:3169 / 3177
页数:9
相关论文
共 50 条
  • [21] Few-Shot Class-Adaptive Anomaly Detection with Model-Agnostic Meta-Learning
    Feng, Tongtong
    Qi, Qi
    Wang, Jingyu
    Liao, Jianxin
    2021 IFIP NETWORKING CONFERENCE AND WORKSHOPS (IFIP NETWORKING), 2021,
  • [22] Crop Disease Recognition Based on Improved Model-Agnostic Meta-Learning
    Si, Xiuli
    Hong, Biao
    Hu, Yuanhui
    Chu, Lidong
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 75 (03): : 6101 - 6118
  • [23] Generalization of Model-Agnostic Meta-Learning Algorithms: Recurring and Unseen Tasks
    Fallah, Alireza
    Mokhtari, Aryan
    Ozdaglar, Asuman
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [24] Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning
    Ji, Kaiyi
    Yang, Junjie
    Liang, Yingbin
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [25] A Compressed Model-Agnostic Meta-Learning Model Based on Pruning for Disease Diagnosis
    Hu, Xiangjun
    Ding, Xiuxiu
    Bai, Dongpeng
    Zhang, Qingchen
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2023, 32 (02)
  • [26] MODEL-AGNOSTIC META-LEARNING FOR RESILIENCE OPTIMIZATION OF ARTIFICIAL INTELLIGENCE SYSTEM
    Moskalenko, V. V.
    RADIO ELECTRONICS COMPUTER SCIENCE CONTROL, 2023, (02) : 79 - 90
  • [27] Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning
    Ji, Kaiyi
    Yang, Junjie
    Liang, Yingbin
    Journal of Machine Learning Research, 2022, 23
  • [28] Trapezoidal Step Scheduler for Model-Agnostic Meta-Learning in Medical Imaging
    Voon, Wingates
    Hum, Yan Chai
    Tee, Yee Kai
    Yap, Wun-She
    Lai, Khin Wee
    Nisar, Humaira
    Mokayed, Hamam
    PATTERN RECOGNITION, 2025, 161
  • [29] Cross Domain Adaptation of Crowd Counting with Model-Agnostic Meta-Learning
    Hou, Xiaoyu
    Xu, Jihui
    Wu, Jinming
    Xu, Huaiyu
    APPLIED SCIENCES-BASEL, 2021, 11 (24):
  • [30] Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and Personalized Federated Learning
    Wang, Bokun
    Yuan, Zhuoning
    Ying, Yiming
    Yang, Tianbao
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24