An imitation learning framework for generating multi-modal trajectories from unstructured demonstrations

被引:3
|
作者
Peng, Jian-Wei [1 ]
Hu, Min-Chun [2 ]
Chu, Wei-Ta [1 ]
机构
[1] Natl Cheng Kung Univ, Dept Comp Sci & Informat Engn, Tainan, Taiwan
[2] Natl Tsing Hua Univ, Dept Comp Sci, Hsinchu, Taiwan
关键词
Trajectory generation; Motion synthesis; Imitation learning; Reinforcement learning; Generative adversarial networks; HUMAN MOTION PREDICTION;
D O I
10.1016/j.neucom.2022.05.076
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The main challenge of the trajectory generation problem is to generate long-term as well as diverse tra-jectories. Generative Adversarial Imitation Learning (GAIL) is a well-known model-free imitation learning algorithm that can be utilized to generate trajectory data, while vanilla GAIL would fail to capture multi -modal demonstrations. Recent methods propose latent variable models to solve this problem; however, previous works may have a mode missing problem. In this work, we propose a novel method to generate long-term trajectories that are controllable by a continuous latent variable based on GAIL and a condi-tional Variational Autoencoder (cVAE). We further assume that subsequences of the same trajectory should be encoded to similar locations in the latent space. Therefore, we introduce a contrastive loss in the training of the encoder. In our motion synthesis task, we propose to first construct a low-dimensional motion manifold by using a VAE to reduce the burden of our imitation learning model. Our experimental results show that the proposed model outperforms the state-of-the-art methods and can be applied to motion synthesis.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:712 / 723
页数:12
相关论文
共 50 条
  • [1] Multi-Modal Imitation Learning from Unstructured Demonstrations using Generative Adversarial Nets
    Hausman, Karol
    Chebotar, Yevgen
    Schaal, Stefan
    Sukhatme, Gaurav
    Lim, Joseph J.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [2] BAGAIL: Multi-modal imitation learning from imbalanced demonstrations
    Gu, Sijia
    Zhu, Fei
    NEURAL NETWORKS, 2024, 174
  • [3] Burn-In Demonstrations for Multi-Modal Imitation Learning
    Kuefler, Alex
    Kochenderfer, Mykel J.
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 1071 - 1078
  • [4] Modelling flight trajectories with multi-modal generative adversarial imitation learning
    Spatharis, Christos
    Blekas, Konstantinos
    Vouros, George A.
    APPLIED INTELLIGENCE, 2024, : 7118 - 7134
  • [5] Multi-Modal Inverse Constrained Reinforcement Learning from a Mixture of Demonstrations
    Qiao, Guanren
    Liu, Guiliang
    Poupart, Pascal
    Xu, Zhiqiang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Learning Forceful Manipulation Skills from Multi-modal Human Demonstrations
    Le, An T.
    Guo, Meng
    van Duijkeren, Niels
    Rozo, Leonel
    Krug, Robert
    Kupcsik, Andras G.
    Buerger, Mathias
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 7770 - 7777
  • [7] A Multi-modal Framework for Robots to Learn Manipulation Tasks from Human Demonstrations
    Congcong Yin
    Qiuju Zhang
    Journal of Intelligent & Robotic Systems, 2023, 107
  • [8] A Multi-modal Framework for Robots to Learn Manipulation Tasks from Human Demonstrations
    Yin, Congcong
    Zhang, Qiuju
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2023, 107 (04)
  • [9] Multi-Modal Imitation Learning Method with Cosine Similarity
    Hao S.
    Liu Q.
    Xu P.
    Zhang L.
    Huang Z.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2023, 60 (06): : 1358 - 1372
  • [10] Triple-GAIL: A Multi-Modal Imitation Learning Framework with Generative Adversarial Nets
    Fei, Cong
    Wang, Bin
    Zhuang, Yuzheng
    Zhang, Zongzhang
    Hao, Jianye
    Zhang, Hongbo
    Ji, Xuewu
    Liu, Wulong
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 2929 - 2935