A novel approach for self-driving car in partially observable environment using life long reinforcement learning

被引:1
|
作者
Quadir, Md Abdul [1 ]
Jaiswal, Dibyanshu [1 ]
Mohan, Senthilkumar [2 ]
Innab, Nisreen [3 ]
Sulaiman, Riza [4 ]
Alaoui, Mohammed Kbiri [5 ]
Ahmadian, Ali [6 ,7 ]
机构
[1] Vellore Inst Technol, Sch Comp Sci & Engn, Chennai 600127, India
[2] Vellore Inst Technol, Sch Comp Sci Engn & Informat Syst, Vellore 632014, Tamilnadu, India
[3] AlMaarefa Univ, Coll Appl Sci, Dept Comp Sci & Informat Syst, Riyadh, Saudi Arabia
[4] Univ Kebangsaan Malaysia, Inst Visual Informat, Bangi 43600, Malaysia
[5] King Khalid Univ, Coll Sci, Dept Math, Abha 61413, POB 9004, Saudi Arabia
[6] Mediterranea Univ Reggio Calabria, Decis Lab, Reggio Di Calabria, Italy
[7] Istanbul Okan Univ, Fac Engn & Nat Sci, Istanbul, Turkiye
来源
关键词
Reinforcement Learning; Lifelong Learning; Self-driving car; Lifelong reinforcement learning; Partially observable Environment; POLICY; GAMES;
D O I
10.1016/j.segan.2024.101356
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
Despite ground-breaking advancements in robotics, gaming, and other challenging domains, reinforcement learning still faces significant challenges in solving dynamic, open-world problems. Since reinforcement learning algorithms usually perform poorly when exposed to new tasks outside of their data distribution, continuous learning algorithms have drawn significant attention. In parallel with work on lifelong learning algorithms, there is a need for challenging environments, properly planned trials, and metrics to measure research success. In this context, a Deep Asynchronous Autonomous Learning System (DAALS) is proposed in this paper for training a selfdriving car in a partially observable environment for real-world scenarios in a continuous state-action space. To cater to three different use cases, three different algorithms were used. To train their agents for learning and upgrading discrete state policies, DAALS used the Asynchronous Advantage Stager Reviewer (AASR) algorithm. To train its agent for continuous state spaces, DAALS also uses an Extensive Deterministic Policy Gradient (EDPG) algorithm. To train the agent in a lifelong form of learning for partially observable environments, DAALS uses a Deep Deterministic Policy Gradient Novel Lifelong Learning Algorithm (DDPGNLLA). The system offers flexibility to the user to train the agents for both discrete and continuous state-action spaces. Compared to previous models in continuous state-action spaces, Deep deterministic policy gradient lifelong learning algorithm outperforms previous models by 46.09%. Furthermore, the Deep Asynchronous Autonomous System tends to outperform all previous reinforcement learning algorithms, making our proposed approach a real-world solution. As DAALS has tested on number of different environments it provides the insights on how modern Artificial Intelligence (AI) solutions can be generalized making it one of the better solutions for AI general domain problems.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] SELF-DRIVING CAR USING LIDAR SENSING AND IMAGE PROCESSING
    Memon, Sidra
    Ahmed, Muhammad
    Narejo, Sanam
    Baig, Umer Ahmed
    Chowdry, Bhawani Shankar
    Anjum, M. Rizwan
    INTERNATIONAL JOURNAL OF GRID AND DISTRIBUTED COMPUTING, 2020, 13 (02): : 77 - 88
  • [22] Parallel, Angular and Perpendicular Parking for Self-Driving Cars using Deep Reinforcement Learning
    Sousa, Bruno
    Ribeiro, Tiago
    Coelho, Joana
    Lopes, Gil
    Ribeiro, A. Fernando
    2022 IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS (ICARSC), 2022, : 40 - 46
  • [23] Advanced Self-driving Car Using CNN: Udacity Simulator
    Hemalatha, T.
    Sivakumar, T. K.
    ARTIFICIAL INTELLIGENCE AND KNOWLEDGE PROCESSING, AIKP 2023, 2024, 2127 : 381 - 396
  • [24] Comparison of Deep Learning Models in Pothole Avoidance for Self-Driving Car
    Rosyid, Harits Ar
    Burhan, Oemar Syarif
    Zaeni, Ilham Ari Elbaith
    Pee, Ahmad Naim Che
    2021 7TH INTERNATIONAL CONFERENCE ON ELECTRICAL, ELECTRONICS AND INFORMATION ENGINEERING (ICEEIE 2021), 2021, : 466 - 471
  • [25] Partially observable environment estimation with uplift inference for reinforcement learning based recommendation
    Shang, Wenjie
    Li, Qingyang
    Qin, Zhiwei
    Yu, Yang
    Meng, Yiping
    Ye, Jieping
    MACHINE LEARNING, 2021, 110 (09) : 2603 - 2640
  • [26] Partially observable environment estimation with uplift inference for reinforcement learning based recommendation
    Wenjie Shang
    Qingyang Li
    Zhiwei Qin
    Yang Yu
    Yiping Meng
    Jieping Ye
    Machine Learning, 2021, 110 : 2603 - 2640
  • [27] Imitative Reinforcement Learning Fusing Vision and Pure Pursuit for Self-driving
    Peng, Mingxing
    Gong, Zhihao
    Sun, Chen
    Chen, Long
    Cao, Dongpu
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 3298 - 3304
  • [28] Toward Self-Driving Bicycles Using State-of-the-Art Deep Reinforcement Learning Algorithms
    Choi, SeungYoon
    Le, Tuyen P.
    Nguyen, Quang D.
    Abu Layek, Md
    Lee, SeungGwan
    Chung, TaeChoong
    SYMMETRY-BASEL, 2019, 11 (02):
  • [29] Continuous improvement of self-driving cars using dynamic confidence-aware reinforcement learning
    Zhong Cao
    Kun Jiang
    Weitao Zhou
    Shaobing Xu
    Huei Peng
    Diange Yang
    Nature Machine Intelligence, 2023, 5 : 145 - 158
  • [30] Continuous improvement of self-driving cars using dynamic confidence-aware reinforcement learning
    Cao, Zhong
    Jiang, Kun
    Zhou, Weitao
    Xu, Shaobing
    Peng, Huei
    Yang, Diange
    NATURE MACHINE INTELLIGENCE, 2023, 5 (02) : 145 - 158