Efficient Sampling-Based Maximum Entropy Inverse Reinforcement Learning With Application to Autonomous Driving

被引:72
|
作者
Wu, Zheng [1 ]
Sun, Liting [1 ]
Zhan, Wei [1 ]
Yang, Chenyu [2 ]
Tomizuka, Masayoshi [1 ]
机构
[1] Univ Calif Berkeley, Dept Mech Engn, Berkeley, CA 94709 USA
[2] Shanghai Jiao Tong Univ, Dept Comp Sci, Shanghai 200240, Peoples R China
关键词
Learning from demonstration; intelligent transportation systems; inverse reinforcement learning; autonomous driving; social human-robot interaction; ALGORITHMS;
D O I
10.1109/LRA.2020.3005126
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In the past decades, we have witnessed significant progress in the domain of autonomous driving. Advanced techniques based on optimization and reinforcement learning become increasingly powerful when solving the forward problem: given designed reward/cost functions, how we should optimize them and obtain driving policies that interact with the environment safely and efficiently. Such progress has raised another equally important question: what should we optimize? Instead of manually specifying the reward functions, it is desired that we can extract what human drivers try to optimize from real traffic data and assign that to autonomous vehicles to enable more naturalistic and transparent interaction between humans and intelligent agents. To address this issue, we present an efficient sampling-based maximum-entropy inverse reinforcement learning (IRL) algorithm in this letter. Different from existing IRL algorithms, by introducing an efficient continuous-domain trajectory sampler, the proposed algorithm can directly learn the reward functions in the continuous domain while considering the uncertainties in demonstrated trajectories from human drivers. We evaluate the proposed algorithm via real-world driving data, including both non-interactive and interactive scenarios. The experimental results show that the proposed algorithm achieves more accurate prediction performance with faster convergence speed and better generalization compared to other baseline IRL algorithms.
引用
收藏
页码:5355 / 5362
页数:8
相关论文
共 50 条
  • [21] Sampling-based Safe Reinforcement Learning for Nonlinear Dynamical Systems
    Suttle, Wesley A.
    Sharma, Vipul K.
    Kosaraju, Krishna C.
    Sivaranjani, S.
    Liu, Ji
    Gupta, Vijay
    Sadler, Brian M.
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [22] Optimistic Thompson Sampling-based Algorithms for Episodic Reinforcement Learning
    Hu, Bingshan
    Zhang, Tianyue H.
    Hegde, Nidhi
    Schmidt, Mark
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 890 - 899
  • [23] Estimation and control using sampling-based Bayesian reinforcement learning
    Slade, Patrick
    Sunberg, Zachary N.
    Kochenderfer, Mykel J.
    IET CYBER-PHYSICAL SYSTEMS: THEORY & APPLICATIONS, 2020, 5 (01) : 127 - 135
  • [24] Revisiting Maximum Entropy Inverse Reinforcement Learning: New Perspectives and Algorithm
    Snoswell, Aaron J.
    Singh, Surya P. N.
    Ye, Nan
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 241 - 249
  • [25] Infinite Time Horizon Maximum Causal Entropy Inverse Reinforcement Learning
    Bloem, Michael
    Bambos, Nicholas
    2014 IEEE 53RD ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2014, : 4911 - 4916
  • [26] Infinite Time Horizon Maximum Causal Entropy Inverse Reinforcement Learning
    Zhou, Zhengyuan
    Bloem, Michael
    Bambos, Nicholas
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2018, 63 (09) : 2787 - 2802
  • [27] AdaBoost maximum entropy deep inverse reinforcement learning with truncated gradient
    Song, Li
    Li, Dazi
    Wang, Xiao
    Xu, Xin
    INFORMATION SCIENCES, 2022, 602 : 328 - 350
  • [28] A Practical Sampling-based Motion Planning Method for Autonomous Driving in Unstructured Environments
    Jin, Xianjian
    Yan, Zeyuan
    Yang, Hang
    Wang, Qikang
    IFAC PAPERSONLINE, 2021, 54 (10): : 449 - 453
  • [29] A Sampling-Based Local Trajectory Planner for Autonomous Driving along a Reference Path
    Li, Xiaohui
    Sun, Zhenping
    Kurt, Arda
    Zhu, Qi
    2014 IEEE INTELLIGENT VEHICLES SYMPOSIUM PROCEEDINGS, 2014, : 376 - 381
  • [30] Research on Trajectory Prediction of Vehicle Lane Change for Autonomous Driving Based on Inverse Reinforcement Learning
    Zhan, Ming
    Fan, Jingjing
    Jin, Linhao
    SEVENTH INTERNATIONAL CONFERENCE ON TRAFFIC ENGINEERING AND TRANSPORTATION SYSTEM, ICTETS 2023, 2024, 13064