Learning Multi-Objective Curricula for Robotic Policy Learning

被引:0
|
作者
Kang, Jikun [1 ,2 ]
Liu, Miao [3 ]
Gupta, Abhinav [2 ,4 ]
Pal, Christopher [2 ,5 ]
Liu, Xue [1 ,2 ]
Fu, Jie [6 ]
机构
[1] McGill Univ, Montreal, PQ, Canada
[2] Mila Quebec AI Inst, Montreal, PQ, Canada
[3] IBM Res, Yorktown Hts, NY USA
[4] Univ Montreal, Montreal, PQ, Canada
[5] Polytech Montreal, Montreal, PQ, Canada
[6] Beijing Acad AI, Beijing, Peoples R China
来源
关键词
ACL; Hyper-net; Multi-objective Curricula;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Various automatic curriculum learning (ACL) methods have been proposed to improve the sample efficiency and final performance of robots' policies learning. They are designed to control how a robotic agent collects data, which is inspired by how humans gradually adapt their learning processes to their capabilities. In this paper, we propose a unified automatic curriculum learning framework to create multi-objective but coherent curricula that are generated by a set of parametric curriculum modules. Each curriculum module is instantiated as a neural network and is responsible for generating a particular curriculum. In order to coordinate those potentially conflicting modules in a unified parameter space, we propose a multi-task hyper-net learning framework that uses a single hyper-net to parameterize all those curriculum modules. We evaluate our method on a series of robotic manipulation tasks and demonstrate its superiority over other state-of-the-art ACL methods in terms of sample efficiency and final performance. Our code is available at https://github.com/luciferkonn/MOC_CoRL22.
引用
收藏
页码:847 / 858
页数:12
相关论文
共 50 条
  • [1] Multi-objective learning control for robotic manipulator
    Win, KKK
    Cheah, CC
    JOURNAL OF ROBOTIC SYSTEMS, 2004, 21 (10): : 539 - 557
  • [2] Multi-objective learning control for robotic manipulator
    Win, KKK
    Cheah, CC
    IEEE ICIT' 02: 2002 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY, VOLS I AND II, PROCEEDINGS, 2002, : 1084 - 1089
  • [3] Incremental reinforcement learning for multi-objective robotic tasks
    Garcia, Javier
    Iglesias, Roberto
    Rodriguez, Miguel A.
    Regueiro, Carlos V.
    KNOWLEDGE AND INFORMATION SYSTEMS, 2017, 51 (03) : 911 - 940
  • [4] Incremental reinforcement learning for multi-objective robotic tasks
    Javier García
    Roberto Iglesias
    Miguel A. Rodríguez
    Carlos V. Regueiro
    Knowledge and Information Systems, 2017, 51 : 911 - 940
  • [5] Nondominated Policy-Guided Learning in Multi-Objective Reinforcement Learning
    Kim, Man-Je
    Park, Hyunsoo
    Ahn, Chang Wook
    ELECTRONICS, 2022, 11 (07)
  • [6] Multi-Objective Recommendation via Multivariate Policy Learning
    Jeunen, Olivier
    Mandav, Jatin
    Potapov, Ivan
    Agarwal, Nakul
    Vaid, Sourabh
    Shi, Wenzhe
    Ustimenko, Aleksei
    PROCEEDINGS OF THE EIGHTEENTH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2024, 2024, : 712 - 721
  • [7] Multi-objective Reinforcement Learning with Path Integral Policy Improvement
    Ariizumi, Ryo
    Sago, Hayato
    Asai, Toru
    Azuma, Shun-ichi
    2023 62ND ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS, SICE, 2023, : 1418 - 1423
  • [8] Neuroevolutionary diversity policy search for multi-objective reinforcement learning
    Zhou, Dan
    Du, Jiqing
    Arai, Sachiyo
    INFORMATION SCIENCES, 2024, 657
  • [9] A Generalized Algorithm for Multi-Objective Reinforcement Learning and Policy Adaptation
    Yang, Runzhe
    Sun, Xingyuan
    Narasimhan, Karthik
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [10] Personalized robotic control via constrained multi-objective reinforcement learning
    He, Xiangkun
    Hu, Zhongxu
    Yang, Haohan
    Lv, Chen
    NEUROCOMPUTING, 2024, 565