LAGOON: Language-Guided Motion Control

被引:0
|
作者
Xu, Shusheng [1 ,2 ]
Wang, Huaijie [1 ,2 ]
Ouyang, Yutao [2 ,3 ]
Gao, Jiaxuan [1 ,2 ]
Meng, Zhiyu [1 ,2 ]
Yu, Chao [1 ]
Wu, Yi [1 ,2 ]
机构
[1] Tsinghua Univ, Beijing, Peoples R China
[2] Shanghai Qi Zhi Inst, Shanghai, Peoples R China
[3] Xiamen Univ, Xiamen, Peoples R China
来源
2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024) | 2024年
关键词
D O I
10.1109/ICRA57147.2024.10610467
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We aim to control a robot to physically behave in the real world following any high-level language command like "cartwheel" or "kick". Although human motion datasets exist, this task remains particularly challenging since generative models can produce physically unrealistic motions, which will be more severe for robots due to different body structures and physical properties. Deploying such a motion to a physical robot can cause even greater difficulties due to the sim2real gap. We develop LAnguage-Guided mOtion cONtrol (LAGOON), a multi-phase reinforcement learning (RL) method to generate physically realistic robot motions under language commands. LAGOON first leverages a pretrained model to generate a human motion from a language command. Then an RL phase trains a control policy in simulation to mimic the generated human motion. Finally, with domain randomization, our learned policy can be deployed to a quadrupedal robot, leading to a quadrupedal robot that can take diverse behaviors in the real world under natural language commands.
引用
收藏
页码:9743 / 9750
页数:8
相关论文
共 50 条
  • [1] Language-guided Human Motion Synthesis with Atomic Actions
    Zhai, Yuanhao
    Huang, Mingzhen
    Luan, Tianyu
    Dong, Lu
    Nwogu, Ifeoma
    Lyu, Siwei
    Doermann, David
    Yuan, Junsong
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5262 - 5271
  • [2] Language-guided Image Reflection Separation
    Zhong, Haofeng
    Hong, Yuchen
    Weng, Shuchen
    Liang, Jinxiu
    Shi, Boxin
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 24913 - 24922
  • [3] DocEdit: Language-Guided Document Editing
    Mathur, Puneet
    Jain, Rajiv
    Gu, Jiuxiang
    Dernoncourt, Franck
    Manocha, Dinesh
    Morariu, Vlad I.
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1914 - 1922
  • [4] CLIP-It! Language-Guided Video Summarization
    Narasimhan, Medhini
    Rohrbach, Anna
    Darrell, Trevor
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] LIDNeRF: Language-Guided NeRF Editing With InstructDiffusion
    Kulkarni, Vaishali
    Sharma, Khushal Hemant
    Shah, Manan
    Vinay, Aniruddh
    INTERNATIONAL JOURNAL OF INTELLIGENT INFORMATION TECHNOLOGIES, 2025, 21 (01)
  • [6] Language-Guided Controller Synthesis for Linear Systems
    Gol, Ebru Aydin
    Lazar, Mircea
    Belta, Calin
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2014, 59 (05) : 1163 - 1176
  • [7] mmFilter: Language-Guided Video Analytics at the Edge
    Hu, Zhiming
    Ye, Ning
    Phillips, Caleb
    Capes, Tim
    Mohomed, Iqbal
    PROCEEDINGS OF THE 2020 21ST INTERNATIONAL MIDDLEWARE CONFERENCE INDUSTRIAL TRACK (MIDDLEWARE INDUSTRY '20), 2020, : 1 - 7
  • [8] A Hardware Accelerator for Language-Guided Reinforcement Learning
    Shiri, Aidin
    Mazumder, Arnab Neelim
    Prakash, Bharat
    Homayoun, Houman
    Waytowich, Nicholas R.
    Mohsenin, Tinoosh
    IEEE DESIGN & TEST, 2022, 39 (03) : 37 - 44
  • [9] Preference-Conditioned Language-Guided Abstraction
    Peng, Andi
    Bobu, Andreea
    Li, Belinda Z.
    Sumers, Theodore R.
    Sucholutsky, Ilia
    Kumar, Nishanth
    Grifths, Thomas L.
    Shah, Julie A.
    PROCEEDINGS OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024, 2024, : 572 - 581
  • [10] LGDN: Language-Guided Denoising Network for Video-Language Modeling
    Lu, Haoyu
    Ding, Mingyu
    Fei, Nanyi
    Huo, Yuqi
    Lu, Zhiwu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,