Deep Reinforcement Learning For Visual Navigation of Wheeled Mobile Robots

被引:2
|
作者
Nwaonumah, Ezebuugo [1 ]
Samanta, Biswanath [1 ]
机构
[1] Georgia Southern Univ, Dept Mech Engn, Statesboro, GA 30460 USA
来源
关键词
asynchronous advantage actor-critic (A3C); convolutional neural network (CNN); deep neural network (DNN); deep reinforcement learning (DRL); machine learning (ML); mapless navigation; reinforcement learning (RL); ResNet50; robotics; robot operating system (ROS);
D O I
10.1109/southeastcon44009.2020.9249654
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
A study is presented on applying deep reinforcement learning (DRL) for visual navigation of wheeled mobile robots (WMR) in dynamic and unknown environments. Two DRL algorithms, namely, value-learning deep Q-network (DQN) and policy gradient based asynchronous advantage actor critic (A3C), have been considered. RGB (red, green and blue) and depth images have been used as inputs in implementation of both DRL algorithms to generate control commands for autonomous navigation of WMR in simulation environments. The initial DRL networks were generated and trained progressively in OpenAI Gym Gazebo based simulation environments within robot operating system (ROS) framework for a popular target WMR, Kobuki TurtleBot2. A pre-trained deep neural network ResNet50 was used after further training with regrouped objects commonly found in laboratory setting for target-driven mapless visual navigation of Turlebot2 through DRL. The performance of A3C with multiple computation threads (4, 6, and 8) was simulated on a desktop. The navigation performance of DQN and A3C networks, in terms of reward statistics and completion time, was compared in three simulation environments. As expected, A3C with multiple threads (4, 6, and 8) performed better than DQN and the performance of A3C improved with number of threads. Details of the methodology, simulation results are presented and recommendations for future work towards real-time implementation through transfer learning of the DRL models are outlined.
引用
收藏
页数:8
相关论文
共 50 条
  • [41] Dynamic Path Planning for Mobile Robots with Deep Reinforcement Learning
    Yang, Laiyi
    Bi, Jing
    Yuan, Haitao
    IFAC PAPERSONLINE, 2022, 55 (11): : 19 - 24
  • [42] Deep Reinforcement Learning Based Mobile Robot Navigation: A Review
    Zhu, Kai
    Zhang, Tao
    TSINGHUA SCIENCE AND TECHNOLOGY, 2021, 26 (05) : 674 - 691
  • [43] Deep Reinforcement Learning Based Mobile Robot Navigation:A Review
    Kai Zhu
    Tao Zhang
    TsinghuaScienceandTechnology, 2021, 26 (05) : 674 - 691
  • [44] Visual Navigation for Mobile Robots: A Survey
    Francisco Bonin-Font
    Alberto Ortiz
    Gabriel Oliver
    Journal of Intelligent and Robotic Systems, 2008, 53 : 263 - 296
  • [45] Visual navigation for mobile robots: A survey
    Bonin-Font, Francisco
    Ortiz, Alberto
    Oliver, Gabriel
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2008, 53 (03) : 263 - 296
  • [46] Continuous Control with Deep Reinforcement Learning for Mobile Robot Navigation
    Xiang, Jiaqi
    Li, Qingdong
    Dong, Xiwang
    Ren, Zhang
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 1501 - 1506
  • [47] A Brief Survey: Deep Reinforcement Learning in Mobile Robot Navigation
    Jiang, Haoge
    Wang, Han
    Yau, Wei-Yun
    Wan, Kong-Wah
    PROCEEDINGS OF THE 15TH IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA 2020), 2020, : 592 - 597
  • [48] Challenges for visual navigation of mobile robots
    Kurashiki K.
    Seimitsu Kogaku Kaishi/Journal of the Japan Society for Precision Engineering, 2019, 85 (01): : 41 - 44
  • [49] Distributed Deep Reinforcement Learning based Indoor Visual Navigation
    Hsu, Shih-Hsi
    Chan, Shoo-Hung
    Wu, Ping-Tsang
    Xiao, Kun
    Fu, Li-Chen
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 2532 - 2537
  • [50] Collision Anticipation via Deep Reinforcement Learning for Visual Navigation
    Gutierrez-Maestro, Eduardo
    Lopez-Sastre, Roberto J.
    Maldonado-Bascon, Saturnino
    PATTERN RECOGNITION AND IMAGE ANALYSIS, PT I, 2020, 11867 : 386 - 397