Robot Path Planning for Persistent Monitoring Based on Improved Deep Q Networks

被引:0
|
作者
Wang X. [1 ,2 ]
Chen Y. [1 ,2 ]
Hu M. [1 ,2 ]
Li X. [1 ,2 ]
机构
[1] Engineering Research Center for Metallurgical Automation and Measurement Technology, Ministry of Education, Wuhan University of Science and Technology, Hubei, Wuhan
[2] Institute of Robotics and Intelligent Systems, Wuhan University of Science and Technology, Hubei, Wuhan
来源
Binggong Xuebao/Acta Armamentarii | 2024年 / 45卷 / 06期
关键词
mobile robot; path planning; persistent monitoring; reinforcement learning;
D O I
10.12382/bgxb.2023.0227
中图分类号
学科分类号
摘要
Persistent monitoring refers to the long-term monitoring of road network environment by planning the patrol route of mobile robots in the road network, so as to achieve the purpose of ensuring environmental safety. The sites to be monitored in the environment are usually limited by the maximum allowable monitoring period (revisit period). A fixed monitoring period should not be set for an optimal monitoring path, otherwise, the monitoring process is easy to be destroyed by malicious intruders. To solve the above problems, a robot monitoring path planning algorithm based on improved Deep Q Networks (DQN) is proposed, the decision-making method of DQN is improved, and a monitoring path with high monitoring frequency, good security (ability to prevent intelligent intrusion) and non-fixed period is planned for robot. Simulated and experimental results show that the proposed algorithm can efficiently cover all nodes to be monitored. Compared with the traditional DQN algorithm, the proposed algorithm does not make the monitoring fall into the cyclic path, and enhances the anti-intrusion ability of the persistent monitoring system. © 2024 China Ordnance Industry Corporation. All rights reserved.
引用
收藏
页码:1813 / 1823
页数:10
相关论文
共 21 条
  • [1] PINTO S C, ANDERSSON S B, HENDRICKX J M, Et al., Optimal periodic multi-agent persistent monitoring of a finite set of targets with uncertain states[C], Proceedings of 2020 American Control Conference, pp. 5207-5212, (2020)
  • [2] HARI S, RATHINAM S, DARBHA S, Et al., Optimal UAV route planning for persistent monitoring missions[J], IEEE Transactions on Robotics, 37, 2, pp. 550-566, (2021)
  • [3] MAINI P, YU K, SUJIT P B, Et al., Persistent monitoring with refueling on a terrain using a team of aerial and ground robots, Proceedings of 2018 IEEE / RSJ International Conference on Intelligent Robots and Systems, pp. 8493-8498, (2018)
  • [4] MAINI P, TOKEKAR P, SUJIT P B., Visibility-based persistent monitoring of piecewise linear features on a terrain using multiple aerial and ground robots [J], IEEE Transactions on Automation Science and Engineering, 18, 4, pp. 1692-1704, (2021)
  • [5] LI Q, GAMA F, RIBEIRO A, Et al., Graph neural networks for decentralized path planning, Proceedings of 2020 IEEE/ RSJ International Conference on Intelligent Robots and Systems, pp. 1901-1903, (2020)
  • [6] TOLSTAYA E, GAMA F, PAULOS J, Et al., Learning decentralized controllers for robot swarms with graph neural networks [C], Proceedings of Conference on Robot Learning, pp. 521-531, (2019)
  • [7] LEW T, SHARMA A, HARRISON J, Et al., Safe active dynamics learning and control: a sequential exploration-exploitation framework, IEEE Transactions on Robotics, 38, 5, pp. 2888-2907, (2022)
  • [8] ZHANG Z, WANG X H, ZHANG Q R, Et al., Multi-robot cooperative pursuit via potential field-enhanced reinforcement learning, Proceedings of 2022 International Conference on Robotics and Automation, pp. 8808-8814, (2022)
  • [9] ASARKAYA A S, AKSARAY D, Temporal-logic-constrained hybrid reinforcement learning to perform optimal aerial monitoring with delivery drones, Proceedings of 2021 International Conference on Unmanned Aircraft Systems, pp. 285-294, (2021)
  • [10] MD KABA, UZUNBAS M G, LIM S N., A reinforcement learning approach to the view planning problem, Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition, pp. 5094-5102, (2017)