An imitation learning framework for generating multi-modal trajectories from unstructured demonstrations

被引:3
|
作者
Peng, Jian-Wei [1 ]
Hu, Min-Chun [2 ]
Chu, Wei-Ta [1 ]
机构
[1] Natl Cheng Kung Univ, Dept Comp Sci & Informat Engn, Tainan, Taiwan
[2] Natl Tsing Hua Univ, Dept Comp Sci, Hsinchu, Taiwan
关键词
Trajectory generation; Motion synthesis; Imitation learning; Reinforcement learning; Generative adversarial networks; HUMAN MOTION PREDICTION;
D O I
10.1016/j.neucom.2022.05.076
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The main challenge of the trajectory generation problem is to generate long-term as well as diverse tra-jectories. Generative Adversarial Imitation Learning (GAIL) is a well-known model-free imitation learning algorithm that can be utilized to generate trajectory data, while vanilla GAIL would fail to capture multi -modal demonstrations. Recent methods propose latent variable models to solve this problem; however, previous works may have a mode missing problem. In this work, we propose a novel method to generate long-term trajectories that are controllable by a continuous latent variable based on GAIL and a condi-tional Variational Autoencoder (cVAE). We further assume that subsequences of the same trajectory should be encoded to similar locations in the latent space. Therefore, we introduce a contrastive loss in the training of the encoder. In our motion synthesis task, we propose to first construct a low-dimensional motion manifold by using a VAE to reduce the burden of our imitation learning model. Our experimental results show that the proposed model outperforms the state-of-the-art methods and can be applied to motion synthesis.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:712 / 723
页数:12
相关论文
共 50 条
  • [21] Accessing learning objects within the multi-modal description framework
    Chen, JS
    Heinrich, E
    Kemp, E
    INTERNATIONAL CONFERENCE ON COMPUTERS IN EDUCATION, VOLS I AND II, PROCEEDINGS, 2002, : 117 - 118
  • [22] A Multi-Modal Hashing Learning Framework for Automatic Image Annotation
    Wang, Jiale
    Li, Guohui
    2017 IEEE SECOND INTERNATIONAL CONFERENCE ON DATA SCIENCE IN CYBERSPACE (DSC), 2017, : 14 - 21
  • [23] Master of Puppets: Multi-modal Robot Activity Segmentation from Teleoperated Demonstrations
    Coppola, Claudio
    Jamone, Lorenzo
    2022 IEEE INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING, ICDL, 2022, : 88 - 94
  • [24] Multi-Modal Knowledge Graph Transformer Framework for Multi-Modal Entity Alignment
    Li, Qian
    Ji, Cheng
    Guo, Shu
    Liang, Zhaoji
    Wang, Lihong
    Li, Jianxin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 987 - 999
  • [25] Multi-Modal Anomaly Detection for Unstructured and Uncertain Environments
    Ji, Tianchen
    Vuppala, Sri Theja
    Chowdhary, Girish
    Driggs-Campbell, Katherine
    CONFERENCE ON ROBOT LEARNING, VOL 155, 2020, 155 : 1443 - 1455
  • [26] Unsupervised Trajectory Segmentation and Promoting of Multi-Modal Surgical Demonstrations
    Shao, Zhenzhou
    Zhao, Hongfa
    Xie, Jiexin
    Qu, Ying
    Guan, Yong
    Tan, Jindong
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 777 - 782
  • [27] A Unified Deep Learning Framework for Multi-Modal Multi-Dimensional Data
    Xi, Pengcheng
    Goubran, Rafik
    Shu, Chang
    2019 IEEE INTERNATIONAL SYMPOSIUM ON MEDICAL MEASUREMENTS AND APPLICATIONS (MEMEA), 2019,
  • [28] Unsupervised Multi-modal Learning
    Iqbal, Mohammed Shameer
    ADVANCES IN ARTIFICIAL INTELLIGENCE (AI 2015), 2015, 9091 : 343 - 346
  • [29] Learning Multi-modal Similarity
    McFee, Brian
    Lanckriet, Gert
    JOURNAL OF MACHINE LEARNING RESEARCH, 2011, 12 : 491 - 523
  • [30] TSC-DL: Unsupervised Trajectory Segmentation of Multi-Modal Surgical Demonstrations with Deep Learning
    Murali, Adithyavairavan
    Garg, Animesh
    Krishnan, Sanjay
    Pokorny, Florian T.
    Abbeel, Pieter
    Darrell, Trevor
    Goldberg, Ken
    2016 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2016, : 4150 - 4157