Efficient Bimanual Handover and Rearrangement via Symmetry-Aware Actor-Critic Learning

被引:3
|
作者
Li, Yunfei [1 ]
Pan, Chaoyi [2 ]
Xu, Huazhe [1 ,4 ,5 ]
Wang, Xiaolong [3 ]
Wu, Yi [1 ,5 ]
机构
[1] Tsinghua Univ, Inst Interdisciplinary Informat Sci, Beijing, Peoples R China
[2] Tsinghua Univ, Dept Elect Engn, Beijing, Peoples R China
[3] Univ Calif San Diego, Dept Elect & Comp Engn, San Diego, CA USA
[4] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
[5] Shanghai Qi Zhi Inst, Shanghai, Peoples R China
关键词
D O I
10.1109/ICRA48891.2023.10160739
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Bimanual manipulation is important for building intelligent robots that unlock richer skills than single arms. We consider a multi-object bimanual rearrangement task, where a reinforcement learning (RL) agent aims to jointly control two arms to rearrange these objects as fast as possible. Solving this task efficiently is challenging for an RL agent due to the requirement of discovering precise intra-arm coordination in an exponentially large control space. We develop a symmetry-aware actor-critic framework that leverages the interchangeable roles of the two manipulators in the bimanual control setting to reduce the policy search space. To handle the compositionality over multiple objects, we augment training data with an object-centric relabeling technique. The overall approach produces an RL policy that can rearrange up to 8 objects with a success rate of over 70% in simulation. We deploy the policy to two Franka Panda arms and further show a successful demo on human-robot collaboration. Videos can be found at https: //sites.google.com/view/bimanual.
引用
收藏
页码:3867 / 3874
页数:8
相关论文
共 50 条
  • [31] Actor-Critic Learning Based on Adaptive Importance Sampling
    Cheng Yuhu
    Feng Huanting
    Wang Xuesong
    CHINESE JOURNAL OF ELECTRONICS, 2010, 19 (04): : 583 - 588
  • [32] DAC: The Double Actor-Critic Architecture for Learning Options
    Zhang, Shangtong
    Whiteson, Shimon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [33] Merging with Extraction Method for Transfer Learning in Actor-Critic
    Takano, Toshiaki
    Takase, Haruhiko
    Kawanaka, Hiroharu
    Tsuruoka, Shinji
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2011, 15 (07) : 814 - 821
  • [34] Actor-Critic reinforcement learning based on prior knowledge
    Yang, Zhenyu, 1600, Transport and Telecommunication Institute, Lomonosova street 1, Riga, LV-1019, Latvia (18):
  • [35] Actor-Critic Reinforcement Learning for Tracking Control in Robotics
    Pane, Yudha P.
    Nageshrao, Subramanya P.
    Babuska, Robert
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5819 - 5826
  • [36] Visual Navigation with Actor-Critic Deep Reinforcement Learning
    Shao, Kun
    Zhao, Dongbin
    Zhu, Yuanheng
    Zhang, Qichao
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [37] Reinforcement learning with actor-critic for knowledge graph reasoning
    Zhang, Linli
    Li, Dewei
    Xi, Yugeng
    Jia, Shuai
    SCIENCE CHINA-INFORMATION SCIENCES, 2020, 63 (06)
  • [38] Actor-critic reinforcement learning for bidding in bilateral negotiation
    Arslan, Furkan
    Aydogan, Reyhan
    TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2022, 30 (05) : 1695 - 1714
  • [39] Reinforcement learning with actor-critic for knowledge graph reasoning
    Linli Zhang
    Dewei Li
    Yugeng Xi
    Shuai Jia
    Science China Information Sciences, 2020, 63
  • [40] A Sandpile Model for Reliable Actor-Critic Reinforcement Learning
    Peng, Yiming
    Chen, Gang
    Zhang, Mengjie
    Pang, Shaoning
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 4014 - 4021