Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning

被引:0
|
作者
Wang, Yongjie [1 ]
Niu, Yuchen [1 ]
Xiao, Mei [1 ]
Zhu, Wenying [1 ]
You, Xinshang [2 ]
机构
[1] Changan Univ, Sch Transportat Engn, Xian 710064, Peoples R China
[2] Hebei Univ Sci & Technol, Sch Econ & Management, Shijiazhuang 050018, Peoples R China
基金
中国国家自然科学基金;
关键词
Human driving behavior; Autonomous vehicle; Occluded pedestrian; Inverse reinforcement learning; Reinforcement learning; PERFORMANCE;
D O I
10.1016/j.physa.2024.130052
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Driving behavior modeling is extremely crucial for designing safe, intelligent, and personalized autonomous driving systems. In this paper, a modeling framework based on Markov Decision Processes (MDPs) is introduced that emulates drivers' decision-making processes. The framework combines the Deep Maximum Entropy Inverse Reinforcement Learning (Deep MEIRL) and a reinforcement learning algorithm-proximal strategy optimization (PPO). A neural network structure is customized for Deep MEIRL, which uses the velocity of the ego vehicle, the pedestrian position, the velocity of surrounding vehicles, the lateral distance, the surrounding vehicles' type, and the distance to the crosswalk to recover the nonlinear reward function. The dataset of drone-based video footage is collected in Xi'an (China) to train and validate the framework. The outcomes demonstrate that Deep MEIRL-PPO outperforms traditional modeling frameworks (Maximum Entropy Inverse Reinforcement Learning (MEIRL)- PPO) in modeling and predicting human driving behavior. Specifically, in predicting human driving behavior, Deep MEIRL-PPO outperforms MEIRL-PPO by 50.71% and 43.90% on the basis of the MAE and HD, respectively. Furthermore, it is discovered that Deep MEIRL-PPO accurately learns the behavior of human drivers avoiding potential conflicts when lines of sight are occluded. This research can contribute to aiding self-driving vehicles in learning human driving behavior and avoiding unforeseen risks.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] An open framework for human-like autonomous driving using Inverse Reinforcement Learning
    Vasquez, Dizan
    Yu, Yufeng
    Kumar, Suryansh
    Laugier, Christian
    2014 IEEE VEHICLE POWER AND PROPULSION CONFERENCE (VPPC), 2014,
  • [22] Analyzing the Suitability of Cost Functions for Explaining and Imitating Human Driving Behavior based on Inverse Reinforcement Learning
    Naumann, Maximilian
    Sun, Liting
    Zhan, Wei
    Tomizuka, Masayoshi
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 5481 - 5487
  • [23] Revisiting Maximum Entropy Inverse Reinforcement Learning: New Perspectives and Algorithm
    Snoswell, Aaron J.
    Singh, Surya P. N.
    Ye, Nan
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 241 - 249
  • [24] Infinite Time Horizon Maximum Causal Entropy Inverse Reinforcement Learning
    Bloem, Michael
    Bambos, Nicholas
    2014 IEEE 53RD ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2014, : 4911 - 4916
  • [25] Infinite Time Horizon Maximum Causal Entropy Inverse Reinforcement Learning
    Zhou, Zhengyuan
    Bloem, Michael
    Bambos, Nicholas
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2018, 63 (09) : 2787 - 2802
  • [26] Simulation of Vehicle Interaction Behavior in Merging Scenarios: A Deep Maximum Entropy-Inverse Reinforcement Learning Method Combined With Game Theory
    Li, Wenli
    Qiu, Fanke
    Li, Lingxi
    Zhang, Yinan
    Wang, Kan
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (01): : 1079 - 1093
  • [27] Deep Inverse Reinforcement Learning for Behavior Prediction in Autonomous Driving: Accurate Forecasts of Vehicle Motion
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    IEEE SIGNAL PROCESSING MAGAZINE, 2021, 38 (01) : 87 - 96
  • [28] A Predictive Reinforcement Learning Framework for Modeling Human Decision Making Behavior
    Kianifar, Rezvan
    Towhidkhah, Farzad
    2009 14TH INTERNATIONAL COMPUTER CONFERENCE, 2009, : 482 - 487
  • [29] Deep Reinforcement Learning in Maximum Entropy Framework with Automatic Adjustment of Mixed Temperature Parameters for Path Planning
    Chen, Yingying
    Ying, Fengkang
    Li, Xiangjian
    Liu, Huashan
    2023 7TH INTERNATIONAL CONFERENCE ON ROBOTICS, CONTROL AND AUTOMATION, ICRCA, 2023, : 78 - 82
  • [30] Estimation of personal driving style via deep inverse reinforcement learning
    Kishikawa, Daiko
    Arai, Sachiyo
    ARTIFICIAL LIFE AND ROBOTICS, 2021, 26 (03) : 338 - 346