Efficient Bimanual Handover and Rearrangement via Symmetry-Aware Actor-Critic Learning

被引:3
|
作者
Li, Yunfei [1 ]
Pan, Chaoyi [2 ]
Xu, Huazhe [1 ,4 ,5 ]
Wang, Xiaolong [3 ]
Wu, Yi [1 ,5 ]
机构
[1] Tsinghua Univ, Inst Interdisciplinary Informat Sci, Beijing, Peoples R China
[2] Tsinghua Univ, Dept Elect Engn, Beijing, Peoples R China
[3] Univ Calif San Diego, Dept Elect & Comp Engn, San Diego, CA USA
[4] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
[5] Shanghai Qi Zhi Inst, Shanghai, Peoples R China
关键词
D O I
10.1109/ICRA48891.2023.10160739
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Bimanual manipulation is important for building intelligent robots that unlock richer skills than single arms. We consider a multi-object bimanual rearrangement task, where a reinforcement learning (RL) agent aims to jointly control two arms to rearrange these objects as fast as possible. Solving this task efficiently is challenging for an RL agent due to the requirement of discovering precise intra-arm coordination in an exponentially large control space. We develop a symmetry-aware actor-critic framework that leverages the interchangeable roles of the two manipulators in the bimanual control setting to reduce the policy search space. To handle the compositionality over multiple objects, we augment training data with an object-centric relabeling technique. The overall approach produces an RL policy that can rearrange up to 8 objects with a success rate of over 70% in simulation. We deploy the policy to two Franka Panda arms and further show a successful demo on human-robot collaboration. Videos can be found at https: //sites.google.com/view/bimanual.
引用
收藏
页码:3867 / 3874
页数:8
相关论文
共 50 条
  • [41] Reinforcement learning with actor-critic for knowledge graph reasoning
    Linli ZHANG
    Dewei LI
    Yugeng XI
    Shuai JIA
    Science China(Information Sciences), 2020, 63 (06) : 223 - 225
  • [42] Distributed Actor-Critic Learning Using Emphatic Weightings
    Stankovic, Milos S.
    Beko, Marko
    Stankovic, Srdjan S.
    2022 8TH INTERNATIONAL CONFERENCE ON CONTROL, DECISION AND INFORMATION TECHNOLOGIES (CODIT'22), 2022, : 1167 - 1172
  • [43] Actor-Critic Reinforcement Learning for Control With Stability Guarantee
    Han, Minghao
    Zhang, Lixian
    Wang, Jun
    Pan, Wei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6217 - 6224
  • [44] Uncertainty Weighted Actor-Critic for Offline Reinforcement Learning
    Wu, Yue
    Zhai, Shuangfei
    Srivastava, Nitish
    Susskind, Joshua
    Zhang, Jian
    Salakhutdinov, Ruslan
    Goh, Hanlin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [45] Deep Actor-Critic Reinforcement Learning for Anomaly Detection
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [46] Fast Learning in an Actor-Critic Architecture with Reward and Punishment
    Balkenius, Christian
    Winberg, Stefan
    TENTH SCANDINAVIAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2008, 173 : 20 - 27
  • [47] MARS: Malleable Actor-Critic Reinforcement Learning Scheduler
    Baheri, Betis
    Tronge, Jacob
    Fang, Bo
    Li, Ang
    Chaudhary, Vipin
    Guan, Qiang
    2022 IEEE INTERNATIONAL PERFORMANCE, COMPUTING, AND COMMUNICATIONS CONFERENCE, IPCCC, 2022,
  • [48] Regularized Soft Actor-Critic for Behavior Transfer Learning
    Tan, Mingxi
    Tian, Andong
    Denoyer, Ludovic
    2022 IEEE CONFERENCE ON GAMES, COG, 2022, : 516 - 519
  • [49] PAC-Bayesian Soft Actor-Critic Learning
    Tasdighi, Bahareh
    Akgul, Abdullah
    Haussmann, Manuel
    Brink, Kenny Kazimirzak
    Kandemir, Melih
    SYMPOSIUM ON ADVANCES IN APPROXIMATE BAYESIAN INFERENCE, 2024, 253 : 127 - 145
  • [50] Averaged Soft Actor-Critic for Deep Reinforcement Learning
    Ding, Feng
    Ma, Guanfeng
    Chen, Zhikui
    Gao, Jing
    Li, Peng
    COMPLEXITY, 2021, 2021