Accelerated Sim-to-Real Deep Reinforcement Learning: Learning Collision Avoidance from Human Player

被引:21
|
作者
Niu, Hanlin [1 ]
Ji, Ze [2 ]
Arvin, Farshad [1 ]
Lennox, Barry [1 ]
Yin, Hujun [1 ]
Carrasco, Joaquin [1 ]
机构
[1] Univ Manchester, Dept Elect & Elect Engn, Manchester, Lancs, England
[2] Cardiff Univ, Sch Engn, Cardiff, Wales
基金
英国工程与自然科学研究理事会;
关键词
D O I
10.1109/IEEECONF49454.2021.9382693
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This paper presents a sensor-level mapless collision avoidance algorithm for use in mobile robots that map raw sensor data to linear and angular velocities and navigate in an unknown environment without a map. An efficient training strategy is proposed to allow a robot to learn from both human experience data and self-exploratory data. A game format simulation framework is designed to allow the human player to tele-operate the mobile robot to a goal and human action is also scored using the reward function. Both human player data and self-playing data are sampled using prioritized experience replay algorithm. The proposed algorithm and training strategy have been evaluated in two different experimental configurations: Environment 1, a simulated cluttered environment, and Environment 2, a simulated corridor environment, to investigate the performance. It was demonstrated that the proposed method achieved the same level of reward using only 16% of the training steps required by the standard Deep Deterministic Policy Gradient (DDPG) method in Environment 1 and 20% of that in Environment 2. In the evaluation of 20 random missions, the proposed method achieved no collision in less than 2 h and 2.5 h of training time in the two Gazebo environments respectively. The method also generated smoother trajectories than DDPG. The proposed method has also been implemented on a real robot in the real-world environment for performance evaluation. We can confirm that the trained model with the simulation software can be directly applied into the real-world scenario without further fine-tuning, further demonstrating its higher robustness than DDPG. The video and code are available: https://youtu.be/BmwxevgsdGc https://github.com/hanlinniu/turtlebot3_ddpg_collision_avoidance
引用
收藏
页码:144 / 149
页数:6
相关论文
共 50 条
  • [1] Sim-to-Real in Reinforcement Learning for Everyone
    Vacaro, Juliano
    Marques, Guilherme
    Oliveira, Bruna
    Paz, Gabriel
    Paula, Thomas
    Staehler, Wagston
    Murphy, David
    2019 LATIN AMERICAN ROBOTICS SYMPOSIUM, 2019 BRAZILIAN SYMPOSIUM ON ROBOTICS (SBR) AND 2019 WORKSHOP ON ROBOTICS IN EDUCATION (LARS-SBR-WRE 2019), 2019, : 305 - 310
  • [2] Sim-to-Real Transfer in Deep Reinforcement Learning for Robotics: a Survey
    Zhao, Wenshuai
    Queralta, Jorge Pena
    Westerlund, Tomi
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 737 - 744
  • [3] Sim-to-Real Control of Trifinger Robot by Deep Reinforcement Learning
    Wan, Qiang
    Wu, Tianyang
    Ye, Jiawei
    Wan, Lipeng
    Lau, Xuguang
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2024, PT VI, 2025, 15206 : 300 - 314
  • [4] Sim-to-Real: Mapless Navigation for USVs Using Deep Reinforcement Learning
    Wang, Ning
    Wang, Yabiao
    Zhao, Yuming
    Wang, Yong
    Li, Zhigang
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2022, 10 (07)
  • [5] Sim-to-Real Deep Reinforcement Learning with Manipulators for Pick-and-Place
    Liu, Wenxing
    Niu, Hanlin
    Skilton, Robert
    Carrasco, Joaquin
    TOWARDS AUTONOMOUS ROBOTIC SYSTEMS, TAROS 2023, 2023, 14136 : 240 - 252
  • [6] Grounded action transformation for sim-to-real reinforcement learning
    Josiah P. Hanna
    Siddharth Desai
    Haresh Karnan
    Garrett Warnell
    Peter Stone
    Machine Learning, 2021, 110 : 2469 - 2499
  • [7] Grounded action transformation for sim-to-real reinforcement learning
    Hanna, Josiah P.
    Desai, Siddharth
    Karnan, Haresh
    Warnell, Garrett
    Stone, Peter
    MACHINE LEARNING, 2021, 110 (09) : 2469 - 2499
  • [8] Cloud-Edge Training Architecture for Sim-to-Real Deep Reinforcement Learning
    Cao, Hongpeng
    Theile, Mirco
    Wyrwal, Federico G.
    Caccamo, Marco
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 9363 - 9370
  • [9] Meta Reinforcement Learning for Sim-to-real Domain Adaptation
    Arndt, Karol
    Hazara, Murtaza
    Ghadirzadeh, Ali
    Kyrki, Ville
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 2725 - 2731
  • [10] Sim-to-real transfer of active suspension control using deep reinforcement learning
    Wiberg, Viktor
    Wallin, Erik
    Falldin, Arvid
    Semberg, Tobias
    Rossander, Morgan
    Wadbro, Eddie
    Servin, Martin
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2024, 179