Deep Reinforcement Learning Visual Target Navigation Method Based on Attention Mechanism and Reward Shaping

被引:0
|
作者
Meng, Yiyue [1 ]
Guo, Chi [2 ]
Liu, Jingnan [1 ]
机构
[1] GNSS Research Center, Wuhan University, Wuhan,430079, China
[2] Hubei Luojia Laboratory, Wuhan,430079, China
关键词
'current - Attention mechanisms - Attention mecha‑ nism - Deep reinforcement learning - Navigation methods - Reinforcement learnings - Reward shaping - Visual Navigation - Visual target navigation - Visual targets;
D O I
10.13203/j.whugis20230193
中图分类号
学科分类号
摘要
Objectives: As one of the important tasks of visual navigation, visual target navigation requires the agent to explore and navigate to the target and issue the done action only relying on visual image infor‑ mation and target information. Presently, the existing methods usually adopt deep reinforcement learning framework to solve visual target navigation problems. However, there are still some shortcomings: (1) The existing methods ignore the relationship between the state of the current and previous time step, resulting in poor navigation performance. (2) The reward settings of the existing methods are fixed and sparse. The agents cannot obtain better navigation strategies under sparse reward. To solve these problems, we propose a deep reinforcement learning visual target navigation method based on attention mechanism and reward shaping. This method can further improve the performance of visual target navigation tasks. Methods: First, the method obtains the area of path focused by the agent at the previous time step based on scaled dot production attention between previous visual image and action. Then, the method obtains the area of path focused by the agent at current time step based on scaled dot production attention between current visual image and previous focused area of path to introduce the state relationship. Besides, to obtain the current focused area of target, we also utilize scaled dot production attention mechanism. We concatenate the current focused area of path and target to build a better state of the agent. Additionally, we propose a reward reshaping rule to solve the problem of sparse reward and apply the cosine similarity between the visual image and target to automatically build a reward space with target preference. Finally, the attention method and reward reshap‑ ing method are combined together to form the deep reinforcement learning visual target navigation method based on attention mechanism and reward shaping. Results: We conduct experiments on AI2-THOR dataset and use success rate (SR) and success weighted by path length (SPL) to evaluate the performance of visual target navigation methods. The results indicate that our method shows 7% improvement in SR and 20% in SPL, which means that the agent can learn a better navigation strategy. In addition, the ablation study shows that the introduction of state relationship and reward shaping can both improve the navigation perfor‑ mance. Conclusions: To draw a conclusion, the proposed deep reinforcement learning visual target naviga‑ tion method based on attention mechanism and reward shaping can further improve the navigation success rate and efficiency by building better states and reward space. © 2024 Editorial Department of Geomatics and Information Science of Wuhan University. All rights reserved.
引用
收藏
页码:1100 / 1108
相关论文
共 50 条
  • [1] Generalization in Deep Reinforcement Learning for Robotic Navigation by Reward Shaping
    Miranda, Victor R. F.
    Neto, Armando A.
    Freitas, Gustavo M.
    Mozelli, Leonardo A.
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (06) : 6013 - 6020
  • [2] An Improvement on Mapless Navigation with Deep Reinforcement Learning: A Reward Shaping Approach
    Alipanah, Arezoo
    Moosavian, S. Ali A.
    2022 10TH RSI INTERNATIONAL CONFERENCE ON ROBOTICS AND MECHATRONICS (ICROM), 2022, : 261 - 266
  • [3] Hindsight Reward Shaping in Deep Reinforcement Learning
    de Villiers, Byron
    Sabatta, Deon
    2020 INTERNATIONAL SAUPEC/ROBMECH/PRASA CONFERENCE, 2020, : 653 - 659
  • [4] Obstacle Avoidance and Navigation Utilizing Reinforcement Learning with Reward Shaping
    Zhang, Daniel
    Bailey, Colleen P.
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS II, 2020, 11413
  • [5] Drone Navigation and Target Interception Using Deep Reinforcement Learning: A Cascade Reward Approach
    Darwish, Ali A.
    Nakhmani, Arie
    IEEE Journal on Indoor and Seamless Positioning and Navigation, 2023, 1 : 130 - 140
  • [6] Distributed Deep Reinforcement Learning based Indoor Visual Navigation
    Hsu, Shih-Hsi
    Chan, Shoo-Hung
    Wu, Ping-Tsang
    Xiao, Kun
    Fu, Li-Chen
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 2532 - 2537
  • [7] Visual Navigation With Multiple Goals Based on Deep Reinforcement Learning
    Rao, Zhenhuan
    Wu, Yuechen
    Yang, Zifei
    Zhang, Wei
    Lu, Shijian
    Lu, Weizhi
    Zha, ZhengJun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (12) : 5445 - 5455
  • [8] Reward Shaping Based Federated Reinforcement Learning
    Hu, Yiqiu
    Hua, Yun
    Liu, Wenyan
    Zhu, Jun
    IEEE ACCESS, 2021, 9 : 67259 - 67267
  • [9] Autonomous navigation of UAV in complex environment : a deep reinforcement learning method based on temporal attention
    Liu, Shuyuan
    Zou, Shufan
    Chang, Xinghua
    Liu, Huayong
    Zhang, Laiping
    Deng, Xiaogang
    APPLIED INTELLIGENCE, 2025, 55 (04)
  • [10] Skill-Based Hierarchical Reinforcement Learning for Target Visual Navigation
    Wang, Shuo
    Wu, Zhihao
    Hu, Xiaobo
    Lin, Youfang
    Lv, Kai
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8920 - 8932