Deep Reinforcement Learning Visual Target Navigation Method Based on Attention Mechanism and Reward Shaping

被引:0
|
作者
Meng, Yiyue [1 ]
Guo, Chi [2 ]
Liu, Jingnan [1 ]
机构
[1] GNSS Research Center, Wuhan University, Wuhan,430079, China
[2] Hubei Luojia Laboratory, Wuhan,430079, China
关键词
'current - Attention mechanisms - Attention mecha‑ nism - Deep reinforcement learning - Navigation methods - Reinforcement learnings - Reward shaping - Visual Navigation - Visual target navigation - Visual targets;
D O I
10.13203/j.whugis20230193
中图分类号
学科分类号
摘要
Objectives: As one of the important tasks of visual navigation, visual target navigation requires the agent to explore and navigate to the target and issue the done action only relying on visual image infor‑ mation and target information. Presently, the existing methods usually adopt deep reinforcement learning framework to solve visual target navigation problems. However, there are still some shortcomings: (1) The existing methods ignore the relationship between the state of the current and previous time step, resulting in poor navigation performance. (2) The reward settings of the existing methods are fixed and sparse. The agents cannot obtain better navigation strategies under sparse reward. To solve these problems, we propose a deep reinforcement learning visual target navigation method based on attention mechanism and reward shaping. This method can further improve the performance of visual target navigation tasks. Methods: First, the method obtains the area of path focused by the agent at the previous time step based on scaled dot production attention between previous visual image and action. Then, the method obtains the area of path focused by the agent at current time step based on scaled dot production attention between current visual image and previous focused area of path to introduce the state relationship. Besides, to obtain the current focused area of target, we also utilize scaled dot production attention mechanism. We concatenate the current focused area of path and target to build a better state of the agent. Additionally, we propose a reward reshaping rule to solve the problem of sparse reward and apply the cosine similarity between the visual image and target to automatically build a reward space with target preference. Finally, the attention method and reward reshap‑ ing method are combined together to form the deep reinforcement learning visual target navigation method based on attention mechanism and reward shaping. Results: We conduct experiments on AI2-THOR dataset and use success rate (SR) and success weighted by path length (SPL) to evaluate the performance of visual target navigation methods. The results indicate that our method shows 7% improvement in SR and 20% in SPL, which means that the agent can learn a better navigation strategy. In addition, the ablation study shows that the introduction of state relationship and reward shaping can both improve the navigation perfor‑ mance. Conclusions: To draw a conclusion, the proposed deep reinforcement learning visual target naviga‑ tion method based on attention mechanism and reward shaping can further improve the navigation success rate and efficiency by building better states and reward space. © 2024 Editorial Department of Geomatics and Information Science of Wuhan University. All rights reserved.
引用
收藏
页码:1100 / 1108
相关论文
共 50 条
  • [41] Offline reward shaping with scaling human preference feedback for deep reinforcement learning
    Li, Jinfeng
    Luo, Biao
    Xu, Xiaodong
    Huang, Tingwen
    NEURAL NETWORKS, 2025, 181
  • [42] Reward shaping to improve the performance of deep reinforcement learning in perishable inventory management
    De Moor, Bram J.
    Gijsbrechts, Joren
    Boute, Robert N.
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2022, 301 (02) : 535 - 545
  • [43] Deep Reinforcement Learning For Visual Navigation of Wheeled Mobile Robots
    Nwaonumah, Ezebuugo
    Samanta, Biswanath
    IEEE SOUTHEASTCON 2020, 2020,
  • [44] Collision Anticipation via Deep Reinforcement Learning for Visual Navigation
    Gutierrez-Maestro, Eduardo
    Lopez-Sastre, Roberto J.
    Maldonado-Bascon, Saturnino
    PATTERN RECOGNITION AND IMAGE ANALYSIS, PT I, 2020, 11867 : 386 - 397
  • [45] Autonomous Visual Navigation using Deep Reinforcement Learning: An Overview
    Ejaz, Muhammad Mudassir
    Tang, Tong Boon
    Lu, Cheng-Kai
    2019 17TH IEEE STUDENT CONFERENCE ON RESEARCH AND DEVELOPMENT (SCORED), 2019, : 294 - 299
  • [46] A Survey on Visual Navigation for Artificial Agents With Deep Reinforcement Learning
    Zeng, Fanyu
    Wang, Chen
    Ge, Shuzhi Sam
    IEEE ACCESS, 2020, 8 : 135426 - 135442
  • [47] Visual Navigation with Actor-Critic Deep Reinforcement Learning
    Shao, Kun
    Zhao, Dongbin
    Zhu, Yuanheng
    Zhang, Qichao
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [48] Residual Network for Deep Reinforcement Learning with Attention Mechanism
    Zhu, Hanhua
    Kaneko, Tomoyuki
    JOURNAL OF INFORMATION SCIENCE AND ENGINEERING, 2021, 37 (03) : 517 - 533
  • [49] A ship target recognition method based on biological visual attention mechanism
    Ma Xiao
    Chen Zhongwei
    Suo Jun
    Zhuansun Xiaobo
    Ni Jiazheng
    Zhang Shuai
    Liu Mo
    AOPC 2021: NOVEL TECHNOLOGIES AND INSTRUMENTS FOR ASTRONOMICAL MULTI-BAND OBSERVATIONS, 2021, 12069
  • [50] Reward Mechanism Design for Deep Reinforcement Learning-Based Microgrid Energy Management
    Hu, Mingjie
    Han, Baohui
    Lv, Shilin
    Bao, Zhejing
    Lu, Lingxia
    Yu, Miao
    2023 6TH INTERNATIONAL CONFERENCE ON RENEWABLE ENERGY AND POWER ENGINEERING, REPE 2023, 2023, : 201 - 205