Obstacle avoidance path planning of mobile robot based on improved DQN

被引:0
|
作者
Tian X. [1 ]
Dong X. [1 ,2 ]
机构
[1] School of Electrical Engineering and Electronic Information, Xihua University, Chengdu
[2] Sichuan University, Jinjiang College, Meishan
关键词
deep reinforcement learning; DQN algorithm; mobile robot; obstacle avoidance; path planning;
D O I
10.13695/j.cnki.12-1222/o3.2024.04.012
中图分类号
学科分类号
摘要
Aiming at the problems such as long learning time, poor exploration ability and sparse reward in obstacle avoidance path planning for robots under general reinforcement learning methods, an obstacle avoidance path planning for mobile robots based on improved Deep Q network (DQN) was proposed. Firstly, based on the traditional DQN algorithm, the obstacle learning rules are designed to remember and avoid obstacles, avoid repeated learning of the same obstacle, and improve the learning efficiency and success rate. Secondly, a reward optimization method is proposed, which uses the difference of access times between states to give rewards, balances the access times of state points, and avoids excessive access. At the same time, by calculating the Euclidean distance from the target point, it is inclined to choose the path close to the target, and cancel the penalty of staying away from the target, and realize the adaptive optimization of the reward mechanism. Finally, the dynamic exploration factor function is designed, and the reinforcement learning strategy is used to select action and learning in the later training to improve the performance and learning efficiency of the algorithm. The experimental simulation results show that compared with the traditional DQN algorithm, the improved algorithm can shorten the training time by 40.25%, the obstacle avoidance success rate by 79.8% and the path length by 2.25%, all of which show better performance. © 2024 Editorial Department of Journal of Chinese Inertial Technology. All rights reserved.
引用
收藏
页码:406 / 416
页数:10
相关论文
共 25 条
  • [1] Liu N, Lu T, Cai Y, Et al., A review of robot manipulation skills learning methods, Acta Automatica Sinica, 45, 3, pp. 458-470, (2019)
  • [2] Wang P, Li X, Song C, Et al., Research on dynamic path planning of wheeled robot based on deep reinforcement learning on the slope ground, Journal of Robotics, pp. 1-20, (2020)
  • [3] Wang L, Wang Y, Sun L., A global path planning algorithm based on obstacle landing node detection, Journal of Chinese Inertial Technology, 30, pp. 388-394, (2022)
  • [4] Lavalle S M., Rapidly-exploring random trees: a new tool for path planning, Research Report, (1999)
  • [5] Zheng K, Yin D, Yin S, Et al., Multi-stage mission plan method of multi-UAVs deployed in different bases based on improved A<sup>*</sup>algorithm, Journal of Chinese Inertial Technology, 30, pp. 248-256, (2022)
  • [6] Luo Q, Wang H, Zheng Y, Et al., Research on path planning of mobile robot based on improved ant colony algorithm, Neural Computing and Applications, 32, 6, pp. 1555-1566, (2020)
  • [7] Hari S, Rathinam S, Darbha S, Et al., Optimal UAV route planning for persistent monitoring missions, IEEE Transactions on Robotics, 37, 2, pp. 550-566, (2021)
  • [8] Xu J, Shao K, Wang J, Et al., Mobile robot dynamic obstacle avoidance method based on improved reinforce- ment learning, Journal of Chinese Inertial Technology, 31, pp. 92-99, (2023)
  • [9] Liu Q, Zhai J, Zhang Z, Et al., A survey on deep reinforcement learning, Chinese Journal of Computers, 41, 1, pp. 1-27, (2018)
  • [10] Zhang L, Tang L, Zhang S, Et al., A self-adaptive reinforcement-exploration Q-learning algorithm, Symmetry, 13, 6, (2021)