Deep Reinforcement Learning Visual Target Navigation Method Based on Attention Mechanism and Reward Shaping

被引:0
|
作者
Meng, Yiyue [1 ]
Guo, Chi [2 ]
Liu, Jingnan [1 ]
机构
[1] GNSS Research Center, Wuhan University, Wuhan,430079, China
[2] Hubei Luojia Laboratory, Wuhan,430079, China
关键词
'current - Attention mechanisms - Attention mecha‑ nism - Deep reinforcement learning - Navigation methods - Reinforcement learnings - Reward shaping - Visual Navigation - Visual target navigation - Visual targets;
D O I
10.13203/j.whugis20230193
中图分类号
学科分类号
摘要
Objectives: As one of the important tasks of visual navigation, visual target navigation requires the agent to explore and navigate to the target and issue the done action only relying on visual image infor‑ mation and target information. Presently, the existing methods usually adopt deep reinforcement learning framework to solve visual target navigation problems. However, there are still some shortcomings: (1) The existing methods ignore the relationship between the state of the current and previous time step, resulting in poor navigation performance. (2) The reward settings of the existing methods are fixed and sparse. The agents cannot obtain better navigation strategies under sparse reward. To solve these problems, we propose a deep reinforcement learning visual target navigation method based on attention mechanism and reward shaping. This method can further improve the performance of visual target navigation tasks. Methods: First, the method obtains the area of path focused by the agent at the previous time step based on scaled dot production attention between previous visual image and action. Then, the method obtains the area of path focused by the agent at current time step based on scaled dot production attention between current visual image and previous focused area of path to introduce the state relationship. Besides, to obtain the current focused area of target, we also utilize scaled dot production attention mechanism. We concatenate the current focused area of path and target to build a better state of the agent. Additionally, we propose a reward reshaping rule to solve the problem of sparse reward and apply the cosine similarity between the visual image and target to automatically build a reward space with target preference. Finally, the attention method and reward reshap‑ ing method are combined together to form the deep reinforcement learning visual target navigation method based on attention mechanism and reward shaping. Results: We conduct experiments on AI2-THOR dataset and use success rate (SR) and success weighted by path length (SPL) to evaluate the performance of visual target navigation methods. The results indicate that our method shows 7% improvement in SR and 20% in SPL, which means that the agent can learn a better navigation strategy. In addition, the ablation study shows that the introduction of state relationship and reward shaping can both improve the navigation perfor‑ mance. Conclusions: To draw a conclusion, the proposed deep reinforcement learning visual target naviga‑ tion method based on attention mechanism and reward shaping can further improve the navigation success rate and efficiency by building better states and reward space. © 2024 Editorial Department of Geomatics and Information Science of Wuhan University. All rights reserved.
引用
收藏
页码:1100 / 1108
相关论文
共 50 条
  • [21] Deep Reinforcement Learning Visual Navigation Model Integrating Memory- prediction Mechanism
    Xiao, Qian
    Yi, Pengfei
    Liu, Rui
    Dong, Jing
    Zhou, Dongsheng
    Zhang, Qiang
    PROCEEDINGS OF THE 2021 IEEE 24TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN (CSCWD), 2021, : 109 - 114
  • [22] A Hierarchical Reinforcement Learning Algorithm Based on Attention Mechanism for UAV Autonomous Navigation
    Liu, Zun
    Cao, Yuanqiang
    Chen, Jianyong
    Li, Jianqiang
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (11) : 13309 - 13320
  • [23] Belief Reward Shaping in Reinforcement Learning
    Marom, Ofir
    Rosman, Benjamin
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3762 - 3769
  • [24] Reward Shaping in Episodic Reinforcement Learning
    Grzes, Marek
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 565 - 573
  • [25] Multigrid Reinforcement Learning with Reward Shaping
    Grzes, Marek
    Kudenko, Daniel
    ARTIFICIAL NEURAL NETWORKS - ICANN 2008, PT I, 2008, 5163 : 357 - 366
  • [26] Modular deep reinforcement learning from reward and punishment for robot navigation
    Wang, Jiexin
    Elfwing, Stefan
    Uchibe, Eiji
    NEURAL NETWORKS, 2021, 135 : 115 - 126
  • [27] Visual Target-Driven Robot Crowd Navigation with Limited FOV Using Self-Attention Enhanced Deep Reinforcement Learning
    Li, Yinbei
    Lyu, Qingyang
    Yang, Jiaqiang
    Salam, Yasir
    Wang, Baixiang
    SENSORS, 2025, 25 (03)
  • [28] An image caption model based on attention mechanism and deep reinforcement learning
    Bai, Tong
    Zhou, Sen
    Pang, Yu
    Luo, Jiasai
    Wang, Huiqian
    Du, Ya
    FRONTIERS IN NEUROSCIENCE, 2023, 17
  • [29] Deep Reinforcement Learning Recommendation System based on GRU and Attention Mechanism
    Hou, Yan-e
    Gu, Wenbo
    Yang, Kang
    Dang, Lanxue
    ENGINEERING LETTERS, 2023, 31 (02) : 695 - 701
  • [30] FUSION SPARSE AND SHAPING REWARD FUNCTION IN SOFT ACTOR-CRITIC DEEP REINFORCEMENT LEARNING FOR MOBILE ROBOT NAVIGATION
    Abu Bakar, Mohamad Hafiz
    Shamsudin, Abu Ubaidah
    Soomro, Zubair Adil
    Tadokoro, Satoshi
    Salaan, C. J.
    JOURNAL OF THE KOREAN SOCIETY OF RADIOLOGY, 2024, 86 (02): : 37 - 49