Meta-Learning with Neural Bandit Scheduler

被引:0
|
作者
Qi, Yunzhe [1 ]
Ban, Yikun [1 ]
Wei, Tianxin [1 ]
Zou, Jiaru [1 ]
Yao, Huaxiu [2 ]
He, Jingrui [1 ]
机构
[1] Univ Illinois, Champaign, IL 61820 USA
[2] Univ North Carolina Chapel Hill, Chapel Hill, NC USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023) | 2023年
基金
美国食品与农业研究所; 美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning has been proven an effective learning paradigm for training machine learning models with good generalization ability. Apart from the common practice of uniformly sampling the meta-training tasks, existing methods working on task scheduling strategies are mainly based on pre-defined sampling protocols or the assumed task-model correlations, and greedily make scheduling decisions, which can lead to sub-optimal performance bottlenecks of the meta-model. In this paper, we propose a novel task scheduling framework under Contextual Bandits settings, named BASS, which directly optimizes the task scheduling strategy based on the status of the meta-model. By balancing the exploitation and exploration in meta-learning task scheduling, BASS can help tackle the challenge of limited knowledge about the task distribution during the early stage of meta-training, while simultaneously exploring potential benefits for forthcoming meta-training iterations through an adaptive exploration strategy. Theoretical analysis and extensive experiments are presented to show the effectiveness of our proposed framework.
引用
收藏
页数:35
相关论文
共 50 条
  • [41] Meta-Modelling Meta-Learning
    Hartmann, Thomas
    Moawad, Assaad
    Schockaert, Cedric
    Fouquet, Francois
    Le Traon, Yves
    2019 ACM/IEEE 22ND INTERNATIONAL CONFERENCE ON MODEL DRIVEN ENGINEERING LANGUAGES AND SYSTEMS (MODELS 2019), 2019, : 300 - 305
  • [42] Stateless neural meta-learning using second-order gradients
    Huisman, Mike
    Plaat, Aske
    van Rijn, Jan N.
    MACHINE LEARNING, 2022, 111 (09) : 3227 - 3244
  • [43] Meta-Learning Stationary Stochastic Process Prediction with Convolutional Neural Processes
    Foong, Andrew Y. K.
    Bruinsma, Wessel P.
    Gordon, Jonathan
    Dubois, Yann
    Requeima, James
    Turner, Richard E.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [44] Learning Tensor Representations for Meta-Learning
    Deng, Samuel
    Guo, Yilin
    Hsu, Daniel
    Mandal, Debmalya
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [45] Meta-learning for fast incremental learning
    Oohira, T
    Yamauchi, K
    Omori, T
    ARTIFICAIL NEURAL NETWORKS AND NEURAL INFORMATION PROCESSING - ICAN/ICONIP 2003, 2003, 2714 : 157 - 164
  • [46] Learning to Propagate for Graph Meta-Learning
    Liu, Lu
    Zhou, Tianyi
    Long, Guodong
    Jiang, Jing
    Zhang, Chengqi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [47] Subspace Learning for Effective Meta-Learning
    Jiang, Weisen
    Kwok, James T.
    Zhang, Yu
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 10177 - 10194
  • [48] Stateless neural meta-learning using second-order gradients
    Mike Huisman
    Aske Plaat
    Jan N. van Rijn
    Machine Learning, 2022, 111 : 3227 - 3244
  • [49] Meta-learning pseudo-differential operators with deep neural networks
    Feliu-Faba, Jordi
    Fan, Yuwei
    Ying, Lexing
    JOURNAL OF COMPUTATIONAL PHYSICS, 2020, 408
  • [50] Hierarchical Meta-learning Models with Deep Neural Networks for Spectrum Assignment
    Rutagemwa, Humphrey
    Baddour, Kareem E.
    Rong, Bo
    2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2019,