Spatiotemporal Costmap Inference for MPC Via Deep Inverse Reinforcement Learning

被引:15
|
作者
Lee, Keuntaek [1 ]
Isele, David [2 ]
Theodorou, Evangelos A. [3 ]
Bae, Sangjae [2 ]
机构
[1] Georgia Inst Technol, Dept Elect & Comp Engn, Atlanta, GA 30318 USA
[2] Honda Res Inst USA Inc, Div Res, San Jose, CA 95110 USA
[3] Georgia Inst Technol, Sch Aerosp Engn, Atlanta, GA 30318 USA
关键词
Learning from demonstration; reinforcement learning; optimization and optimal control; motion and path planning; autonomous vehicle navigation;
D O I
10.1109/LRA.2022.3146635
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
It can he difficult to autonomously produce driver behavior so that it appears natural to other traffic participants. Through Inverse Reinforcement Learning (IRL), we can automate this process by learning the underlying reward function from human demonstrations. We propose a new IRL algorithm that learns a goal-conditioned spatio-temporal reward function. The resulting costmap is used by Model Predictive Controllers (MPCs) to perform a task without any hand-designing or hand-tuning of the cost function. We evaluate our proposed Goal-conditioned SpatioTemporal Zeroing Maximum Entropy Deep IRL (GSTZ)-MEDIRL framework together with MPC in the CARLA simulator for autonomous driving, lane keeping, and lane changing tasks in a challenging dense traffic highway scenario. Our proposed methods show higher success rates compared to other baseline methods including behavior cloning, state-of-the-art RL policies, and MPC with a learning-based behavior prediction model.
引用
收藏
页码:3194 / 3201
页数:8
相关论文
共 50 条
  • [11] Symbolic Task Inference in Deep Reinforcement Learning
    Hasanbeig, Hosein
    Jeppu, Natasha Yogananda
    Abate, Alessandro
    Melham, Tom
    Kroening, Daniel
    Journal of Artificial Intelligence Research, 2024, 80 : 1099 - 1137
  • [12] Deep reinforcement learning with significant multiplications inference
    Ivanov, Dmitry A.
    Larionov, Denis A.
    Kiselev, Mikhail V.
    Dylov, Dmitry V.
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [13] Deep reinforcement learning with significant multiplications inference
    Dmitry A. Ivanov
    Denis A. Larionov
    Mikhail V. Kiselev
    Dmitry V. Dylov
    Scientific Reports, 13
  • [14] Symbolic Task Inference in Deep Reinforcement Learning
    Hasanbeig, Hosein
    Jeppu, Natasha Yogananda
    Abate, Alessandro
    Melham, Tom
    Kroening, Daniel
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2024, 80 : 1099 - 1137
  • [15] Spatiotemporal path tracking via deep reinforcement learning of robot for manufacturing internal logistics
    Fan, Fei
    Xu, Guanglin
    Feng, Na
    Li, Lin
    Jiang, Wei
    Yu, Lianqin
    Xiong, Xiaoshuang
    JOURNAL OF MANUFACTURING SYSTEMS, 2023, 69 : 150 - 169
  • [16] Deep Inverse Reinforcement Learning for Sepsis Treatment
    Yu, Chao
    Ren, Guoqi
    Liu, Jiming
    2019 IEEE INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI), 2019, : 373 - 375
  • [17] Deep Inverse Reinforcement Learning by Logistic Regression
    Uchibe, Eiji
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT I, 2016, 9947 : 23 - 31
  • [18] An Online Training Method for Augmenting MPC with Deep Reinforcement Learning
    Bellegarda, Guillaume
    Byl, Katie
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 5453 - 5459
  • [19] Machining sequence learning via inverse reinforcement learning
    Sugisawa, Yasutomo
    Takasugi, Keigo
    Asakawa, Naoki
    PRECISION ENGINEERING-JOURNAL OF THE INTERNATIONAL SOCIETIES FOR PRECISION ENGINEERING AND NANOTECHNOLOGY, 2022, 73 : 477 - 487
  • [20] Automatic inverse treatment planning of Gamma Knife radiosurgery via deep reinforcement learning
    Liu, Yingzi
    Shen, Chenyang
    Wang, Tonghe
    Zhang, Jiahan
    Yang, Xiaofeng
    Liu, Tian
    Kahn, Shannon
    Shu, Hui-Kuo
    Tian, Zhen
    MEDICAL PHYSICS, 2022, 49 (05) : 2877 - 2889