FUSION SPARSE AND SHAPING REWARD FUNCTION IN SOFT ACTOR-CRITIC DEEP REINFORCEMENT LEARNING FOR MOBILE ROBOT NAVIGATION

被引:0
|
作者
Abu Bakar, Mohamad Hafiz [1 ]
Shamsudin, Abu Ubaidah [1 ]
Soomro, Zubair Adil [1 ]
Tadokoro, Satoshi [2 ]
Salaan, C. J. [3 ]
机构
[1] Univ Tun Hussein Onn Malaysia, Fac Elect & Elect Engn, Batu Pahat 86400, Johor, Malaysia
[2] Tohoku Univ, 2 Chome 1-1 Katahira,Aoba Ward, Sendai, Miyagi 9808577, Japan
[3] MSU Iligan Inst Technol, Dept Elect Engn & Technol, Andres Bonifacio Ave, Lanao Del Norte 9200, Philippines
来源
关键词
Soft Actor Critic Deep Reinforcement Learning (SAC DRL); Deep Reinforcement Learning; Mobile robot navigation; Reward function; Sparse reward; Shaping reward;
D O I
暂无
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Nowadays, the advancement in autonomous robots is the latest influenced by the development of a world surrounded by new technologies. Deep Reinforcement Learning (DRL) allows systems to operate automatically, so the robot will learn the next movement based on the interaction with the environment. Moreover, since robots require continuous action, Soft Actor Critic Deep Reinforcement Learning (SAC DRL) is considered the latest DRL approach solution. SAC is used because its ability to control continuous action to produce more accurate movements. SAC fundamental is robust against unpredictability, but some weaknesses have been identified, particularly in the exploration process for accuracy learning with faster maturity. To address this issue, the study identified a solution using a reward function appropriate for the system to guide in the learning process. This research proposes several types of reward functions based on sparse and shaping reward in SAC method to investigate the effectiveness of mobile robot learning. Finally, the experiment shows that using fusion sparse and shaping rewards in the SAC DRL successfully navigates to the target position and can also increase accuracy based on the average error result of 4.99%.
引用
收藏
页码:37 / 49
页数:13
相关论文
共 50 条
  • [41] Enhancing HVAC Control Systems Using a Steady Soft Actor-Critic Deep Reinforcement Learning Approach
    Sun, Hongtao
    Hu, Yushuang
    Luo, Jinlu
    Guo, Qiongyu
    Zhao, Jianzhe
    BUILDINGS, 2025, 15 (04)
  • [42] Optimal Dispatch of Integrated Electricity-gas System With Soft Actor-critic Deep Reinforcement Learning
    Qiao, Ji
    Wang, Xinying
    Zhang, Qing
    Zhang, Dongxia
    Pu, Tianjiao
    Zhongguo Dianji Gongcheng Xuebao/Proceedings of the Chinese Society of Electrical Engineering, 2021, 41 (03): : 819 - 832
  • [43] On the sample complexity of actor-critic method for reinforcement learning with function approximation
    Harshat Kumar
    Alec Koppel
    Alejandro Ribeiro
    Machine Learning, 2023, 112 : 2433 - 2467
  • [44] SOFT ACTOR-CRITIC REINFORCEMENT LEARNING FOR ROBOTIC MANIPULATOR WITH HINDSIGHT EXPERIENCE REPLAY
    Yan, Tao
    Zhang, Wenan
    Yang, Simon X.
    Yu, Li
    INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2019, 34 (05): : 536 - 543
  • [45] Optimising maize threshing process with temporal proximity soft actor-critic deep reinforcement learning algorithm
    Zhang, Qiang
    Fang, Xuwen
    Gao, Xiaodi
    Zhang, Jinsong
    Zhao, Xuelin
    Yu, Lulu
    Yu, Chunsheng
    Zhou, Deyi
    Zhou, Haigen
    Zhang, Li
    Wu, Xinling
    BIOSYSTEMS ENGINEERING, 2024, 248 : 229 - 239
  • [46] SAC-FACT: Soft Actor-Critic Reinforcement Learning for Counterfactual Explanations
    Ezzeddine, Fatima
    Ayoub, Omran
    Andreoletti, Davide
    Giordano, Silvia
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, XAI 2023, PT I, 2023, 1901 : 195 - 216
  • [47] CONTROLLED SENSING AND ANOMALY DETECTION VIA SOFT ACTOR-CRITIC REINFORCEMENT LEARNING
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4198 - 4202
  • [48] A Developmental Actor-Critic Reinforcement Learning Approach for Task-Nonspecific Robot
    Li, Xiaoan
    Yang, Yuan
    Sun, Yunming
    Zhang, Lu
    2016 IEEE CHINESE GUIDANCE, NAVIGATION AND CONTROL CONFERENCE (CGNCC), 2016, : 2231 - 2237
  • [49] A soft actor-critic deep reinforcement learning method for multi-timescale coordinated operation of microgrids
    Chunchao Hu
    Zexiang Cai
    Yanxu Zhang
    Rudai Yan
    Yu Cai
    Bowei Cen
    Protection and Control of Modern Power Systems, 2022, 7
  • [50] Reinforcement learning for automatic quadrilateral mesh generation: A soft actor-critic approach
    Pan, Jie
    Huang, Jingwei
    Cheng, Gengdong
    Zeng, Yong
    NEURAL NETWORKS, 2023, 157 : 288 - 304