Robot Dynamic Path Planning Based on Prioritized Experience Replay and LSTM Network

被引:0
|
作者
Li, Hongqi [1 ]
Zhong, Peisi [1 ]
Liu, Li [2 ]
Wang, Xiao [1 ]
Liu, Mei [3 ]
Yuan, Jie [1 ]
机构
[1] Shandong Univ Sci & Technol, Coll Mech & Elect Engn, Qingdao 266590, Peoples R China
[2] Shandong Univ Sci & Technol, Coll Comp Sci & Engn, Qingdao 266590, Peoples R China
[3] Shandong Univ Sci & Technol, Coll Energy Storage Technol, Qingdao 266590, Peoples R China
来源
IEEE ACCESS | 2025年 / 13卷
基金
中国国家自然科学基金;
关键词
Heuristic algorithms; Long short term memory; Path planning; Convergence; Robots; Training; Planning; Adaptation models; Accuracy; Deep reinforcement learning; DDQN; LSTM network; mobile robot; path planning; prioritized experience replay; LEARNING ALGORITHM;
D O I
10.1109/ACCESS.2025.3532449
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To address the issues of slow convergence speed, poor dynamic adaptability, and path redundancy in the Double Deep Q Network (DDQN) within complex obstacle environments, this paper proposes an enhanced algorithm within the deep reinforcement learning framework. This algorithm, termed LPDDQN, integrates Prioritized Experience Replay (PER) and the Long Short Term Memory (LSTM) network to improve upon the DDQN algorithm. First, Prioritized Experience Replay (PER) is utilized to prioritize experience data and optimize storage and sampling operations through the SumTree structure, rather than the conventional experience queue. Second, the LSTM network is introduced to enhance the dynamic adaptability of the DDQN algorithm. Owing to the introduction of the LSTM model, the experience samples must be sliced and populated. The performance of the proposed LPDDQN algorithm is compared with five other path planning algorithms in both static and dynamic environments. Simulation analysis shows that in a static environment, LPDDQN demonstrates significant improvements over traditional DDQN in terms of convergence, number of moving steps, success rate, and number of turns, with respective improvements of 24.07%, 17.49%, 37.73%, and 61.54%. In dynamic and complex environments, the success rates of all algorithms, except TLD3 and the LPDDQN, decreased significantly. Further analysis reveals that the LPDDQN outperforms the TLD3 by 18.87%, 2.41%, and 39.02% in terms of moving steps, success rate, and number of turns, respectively.
引用
收藏
页码:22283 / 22299
页数:17
相关论文
共 50 条
  • [41] Robot Dynamic Path Planning Based on Improved A* and DWA Algorithms
    Guan, Chenxi
    Wang, Shuying
    2022 4TH INTERNATIONAL CONFERENCE ON CONTROL AND ROBOTICS, ICCR, 2022, : 53 - 58
  • [42] Mobile Robot Dynamic Path Planning Based on Genetic Algorithm
    Wu, Bing
    Wang, Yanping
    PROCEEDINGS OF ANNUAL CONFERENCE OF CHINA INSTITUTE OF COMMUNICATIONS, 2010, : 96 - +
  • [43] Mobile Robot Path Planning Based on Dynamic Movement Primitives
    Jiang, Minghao
    Chen, Yang
    Zheng, Wenlei
    Wu, Huaiyu
    Cheng, Lei
    2016 IEEE INTERNATIONAL CONFERENCE ON INFORMATION AND AUTOMATION (ICIA), 2016, : 980 - 985
  • [44] A NEW ROBOT PATH PLANNING METHOD BASED ON LSTM NEURAL NETWORK AND RAPIDLY-EXPLORING RANDOM TREE ALGORITHM
    Hu, Weifei
    Tang, Feng
    Liu, Zhenyu
    Tan, Jianrong
    PROCEEDINGS OF ASME 2021 INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, IDETC-CIE2021, VOL 8B, 2021,
  • [45] Path planning based on a recurrent neural network for an evolutionary robot
    School of Control Science and Engineering, Shandong University, Jinan 250061, China
    不详
    不详
    Harbin Gongcheng Daxue Xuebao, 2009, 8 (898-902):
  • [46] Path planning method of a humanoid robot based on a home network
    Yu, Jiayuan
    Zhang, Lei
    Bian, Gui-Bin
    Liu, Chenxi
    2019 IEEE 4TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2019), 2019, : 875 - 880
  • [47] Graph neural network based method for robot path planning
    Diao, Xingrong
    Chi, Wenzheng
    Wang, Jiankun
    BIOMIMETIC INTELLIGENCE AND ROBOTICS, 2024, 4 (01):
  • [48] A cooperative EV charging scheduling strategy based on double deep Q-network and Prioritized experience replay
    Zhang, Yanyu
    Rao, Xinpeng
    Liu, Chunyang
    Zhang, Xibeng
    Zhou, Yi
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 118
  • [49] A modified dueling DQN algorithm for robot path planning incorporating priority experience replay and artificial potential fieldsA modified dueling dqn algorithm for robot path planning incorporating priority experience replay and artificial potential fieldsX. Yue et al.
    Chang Li
    Xiaofeng Yue
    Zeyuan Liu
    Guoyuan Ma
    Hongbo Zhang
    Yuan Zhou
    Juan Zhu
    Applied Intelligence, 2025, 55 (6)
  • [50] Two Techniques That Enhance the Performance of Multi-robot Prioritized Path Planning
    Andreychuk, Anton
    Yakovlev, Konstantin
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 2177 - 2179