Reinforcement learning framework for UAV-based target localization applications

被引:14
|
作者
Shurrab, Mohammed [1 ]
Mizouni, Rabeb [1 ]
Singh, Shakti [1 ]
Otrok, Hadi [1 ]
机构
[1] Khalifa Univ, Elect Engn & Comp Sci Dept, Abu Dhabi 127788, U Arab Emirates
关键词
Target localization; Unmanned aerial vehicle (UAV); Reinforcement learning (RL); Deep Q-network (DQN); Data-driven; Deep reinforcement learning (DRL); Smart environmental monitoring (SEM); INTERNET; SYSTEM; THINGS; IOT;
D O I
10.1016/j.iot.2023.100867
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Smart environmental monitoring has gained prominence, where target localization is of utmost importance. Employing UAVs for localization tasks is appealing owing to their low-cost, light-weight, and high maneuverability. However, UAVs lack the autonomy of decision-making if met with uncertain situations. Therefore, reinforcement learning (RL) can introduce intelligence to UAVs, where they learn to act based on the presented situation. Existing works focus on UAV trajectory optimization, navigation, and target tracking. These methods are application-specific and cannot be adapted to localization tasks since they require prior knowledge of the target. Moreover, the current RL-based autonomous target localization systems are lacking since-1) they must keep track of all visited locations and their corresponding readings, 2) they require retraining when encountering new environments, and 3) they are not scalable since the agent's movement is limited to slow speeds and for specific environments. Therefore, this work proposes a data-driven UAV target localization system based on Q-learning, which employs tabular methods to learn the optimal policy. Deep Q-network (DQN) is introduced to enhance the RL model and alleviate the curse of dimensionality. The proposed models enable smart decision-making, where the sensory information gathered by the UAV is exploited to produce the best action. Moreover, the UAV movement is modeled based on motion physics, where the actions correspond to linear velocities and heading angles. The proposed approach is compared with different benchmarks, where the results indicate that a more efficient, scalable, and adaptable localization is achieved, irrespective of the environment or source characteristics, without retraining.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] An Informative Path Planning Framework for Active Learning in UAV-Based Semantic Mapping
    Rueckin, Julius
    Magistri, Federico
    Stachniss, Cyrill
    Popovic, Marija
    IEEE TRANSACTIONS ON ROBOTICS, 2023, 39 (06) : 4279 - 4296
  • [22] UAV-based Smart Agriculture: a Review of UAV Sensing and Applications
    Moradi, Salaheddin
    Bokani, Ayub
    Hassan, Jahan
    2022 32ND INTERNATIONAL TELECOMMUNICATION NETWORKS AND APPLICATIONS CONFERENCE (ITNAC), 2022, : 181 - 184
  • [23] A Review on UAV-Based Applications for Precision Agriculture
    Tsouros, Dimosthenis C.
    Bibi, Stamatia
    Sarigiannidis, Panagiotis G.
    INFORMATION, 2019, 10 (11)
  • [24] Experimental Comparison of UAV-Based RSSI and AoA Localization
    Scazzoli, Davide
    Moro, Stefano
    Teeda, Vineeth
    Upadhyay, Prabhat Kumar
    Magarini, Maurizio
    IEEE SENSORS LETTERS, 2024, 8 (01) : 1 - 4
  • [25] The Impact of Dataset on Offline Reinforcement Learning Performance in UAV-Based Emergency Network Recovery Tasks
    Eo, Jeyeon
    Lee, Dongsu
    Kwon, Minhae
    IEEE COMMUNICATIONS LETTERS, 2024, 28 (05) : 1058 - 1061
  • [26] Development of UAV-Based Target Tracking and Recognition Systems
    Wang, Shuaijun
    Jiang, Fan
    Zhang, Bin
    Ma, Rui
    Hao, Qi
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2020, 21 (08) : 3409 - 3422
  • [27] Trajectory Design for UAV-Based Internet of Things Data Collection: A Deep Reinforcement Learning Approach
    Wang, Yang
    Gao, Zhen
    Zhang, Jun
    Cao, Xianbin
    Zheng, Dezhi
    Gao, Yue
    Ng, Derrick Wing Kwan
    Renzo, Marco Di
    IEEE Internet of Things Journal, 2022, 9 (05): : 3899 - 3912
  • [28] Adaptive Informative Path Planning Using Deep Reinforcement Learning for UAV-based Active Sensing
    Rueckin, Julius
    Jin, Liren
    Popovic, Marija
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 4473 - 4479
  • [29] Trajectory Design for UAV-Based Internet of Things Data Collection: A Deep Reinforcement Learning Approach
    Wang, Yang
    Gao, Zhen
    Zhang, Jun
    Cao, Xianbin
    Zheng, Dezhi
    Gao, Yue
    Ng, Derrick Wing Kwan
    Di Renzo, Marco
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (05): : 3899 - 3912
  • [30] Future UAV-Based ITS: A Comprehensive Scheduling Framework
    Ghazzai, Hakim
    Menouar, Hamid
    Kadri, Abdullah
    Massoud, Yehia
    IEEE ACCESS, 2019, 7 : 75678 - 75695