FUSION SPARSE AND SHAPING REWARD FUNCTION IN SOFT ACTOR-CRITIC DEEP REINFORCEMENT LEARNING FOR MOBILE ROBOT NAVIGATION

被引:0
|
作者
Abu Bakar, Mohamad Hafiz [1 ]
Shamsudin, Abu Ubaidah [1 ]
Soomro, Zubair Adil [1 ]
Tadokoro, Satoshi [2 ]
Salaan, C. J. [3 ]
机构
[1] Univ Tun Hussein Onn Malaysia, Fac Elect & Elect Engn, Batu Pahat 86400, Johor, Malaysia
[2] Tohoku Univ, 2 Chome 1-1 Katahira,Aoba Ward, Sendai, Miyagi 9808577, Japan
[3] MSU Iligan Inst Technol, Dept Elect Engn & Technol, Andres Bonifacio Ave, Lanao Del Norte 9200, Philippines
来源
关键词
Soft Actor Critic Deep Reinforcement Learning (SAC DRL); Deep Reinforcement Learning; Mobile robot navigation; Reward function; Sparse reward; Shaping reward;
D O I
暂无
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Nowadays, the advancement in autonomous robots is the latest influenced by the development of a world surrounded by new technologies. Deep Reinforcement Learning (DRL) allows systems to operate automatically, so the robot will learn the next movement based on the interaction with the environment. Moreover, since robots require continuous action, Soft Actor Critic Deep Reinforcement Learning (SAC DRL) is considered the latest DRL approach solution. SAC is used because its ability to control continuous action to produce more accurate movements. SAC fundamental is robust against unpredictability, but some weaknesses have been identified, particularly in the exploration process for accuracy learning with faster maturity. To address this issue, the study identified a solution using a reward function appropriate for the system to guide in the learning process. This research proposes several types of reward functions based on sparse and shaping reward in SAC method to investigate the effectiveness of mobile robot learning. Finally, the experiment shows that using fusion sparse and shaping rewards in the SAC DRL successfully navigates to the target position and can also increase accuracy based on the average error result of 4.99%.
引用
收藏
页码:37 / 49
页数:13
相关论文
共 50 条
  • [31] Deep Reinforcement Learning in VizDoom via DQN and Actor-Critic Agents
    Bakhanova, Maria
    Makarov, Ilya
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, IWANN 2021, PT I, 2021, 12861 : 138 - 150
  • [32] A Deep Actor-Critic Reinforcement Learning Framework for Dynamic Multichannel Access
    Zhong, Chen
    Lu, Ziyang
    Gursoy, M. Cenk
    Velipasalar, Senem
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2019, 5 (04) : 1125 - 1139
  • [33] Applying Online Expert Supervision in Deep Actor-Critic Reinforcement Learning
    Zhang, Jin
    Chen, Jiansheng
    Huang, Yiqing
    Wan, Weitao
    Li, Tianpeng
    PATTERN RECOGNITION AND COMPUTER VISION, PT II, 2018, 11257 : 469 - 478
  • [34] Multi-agent dual actor-critic framework for reinforcement learning navigation
    Xiong, Fengguang
    Zhang, Yaodan
    Kuang, Xinhe
    He, Ligang
    Han, Xie
    APPLIED INTELLIGENCE, 2025, 55 (02)
  • [35] Sparse Actor-Critic: Sparse Tsallis Entropy Regularized Reinforcement Learning in a Continuous Action Space
    Choy, Jaegoo
    Lee, Kyungjae
    Oh, Songhwai
    2020 17TH INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS (UR), 2020, : 68 - 73
  • [36] Coverage Path Planning Using Actor-Critic Deep Reinforcement Learning
    Garrido-Castaneda, Sergio Isahi
    Vasquez, Juan Irving
    Antonio-Cruz, Mayra
    SENSORS, 2025, 25 (05)
  • [37] Fully distributed actor-critic architecture for multitask deep reinforcement learning
    Valcarcel Macua, Sergio
    Davies, Ian
    Tukiainen, Aleksi
    De Cote, Enrique Munoz
    KNOWLEDGE ENGINEERING REVIEW, 2021, 36
  • [38] A novel reward-shaping-based soft actor-critic for random trajectory tracking of AUVs
    Zhang, Yue
    Zhang, Tianze
    Li, Yibin
    Zhuang, Yinghao
    Wang, Daichao
    OCEAN ENGINEERING, 2025, 322
  • [39] A soft actor-critic deep reinforcement learning method for multi-timescale coordinated operation of microgrids
    Hu, Chunchao
    Cai, Zexiang
    Zhang, Yanxu
    Yan, Rudai
    Cai, Yu
    Cen, Bowei
    PROTECTION AND CONTROL OF MODERN POWER SYSTEMS, 2022, 7 (01)
  • [40] On the sample complexity of actor-critic method for reinforcement learning with function approximation
    Kumar, Harshat
    Koppel, Alec
    Ribeiro, Alejandro
    MACHINE LEARNING, 2023, 112 (07) : 2433 - 2467