Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning

被引:0
|
作者
Wang, Yongjie [1 ]
Niu, Yuchen [1 ]
Xiao, Mei [1 ]
Zhu, Wenying [1 ]
You, Xinshang [2 ]
机构
[1] Changan Univ, Sch Transportat Engn, Xian 710064, Peoples R China
[2] Hebei Univ Sci & Technol, Sch Econ & Management, Shijiazhuang 050018, Peoples R China
基金
中国国家自然科学基金;
关键词
Human driving behavior; Autonomous vehicle; Occluded pedestrian; Inverse reinforcement learning; Reinforcement learning; PERFORMANCE;
D O I
10.1016/j.physa.2024.130052
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Driving behavior modeling is extremely crucial for designing safe, intelligent, and personalized autonomous driving systems. In this paper, a modeling framework based on Markov Decision Processes (MDPs) is introduced that emulates drivers' decision-making processes. The framework combines the Deep Maximum Entropy Inverse Reinforcement Learning (Deep MEIRL) and a reinforcement learning algorithm-proximal strategy optimization (PPO). A neural network structure is customized for Deep MEIRL, which uses the velocity of the ego vehicle, the pedestrian position, the velocity of surrounding vehicles, the lateral distance, the surrounding vehicles' type, and the distance to the crosswalk to recover the nonlinear reward function. The dataset of drone-based video footage is collected in Xi'an (China) to train and validate the framework. The outcomes demonstrate that Deep MEIRL-PPO outperforms traditional modeling frameworks (Maximum Entropy Inverse Reinforcement Learning (MEIRL)- PPO) in modeling and predicting human driving behavior. Specifically, in predicting human driving behavior, Deep MEIRL-PPO outperforms MEIRL-PPO by 50.71% and 43.90% on the basis of the MAE and HD, respectively. Furthermore, it is discovered that Deep MEIRL-PPO accurately learns the behavior of human drivers avoiding potential conflicts when lines of sight are occluded. This research can contribute to aiding self-driving vehicles in learning human driving behavior and avoiding unforeseen risks.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Estimation of personal driving style via deep inverse reinforcement learning
    Daiko Kishikawa
    Sachiyo Arai
    Artificial Life and Robotics, 2021, 26 : 338 - 346
  • [32] A Modified Maximum Entropy Inverse Reinforcement Learning Approach for Microgrid hnergy Scheduling
    Lin, Yanbin
    Das, Avijit
    Ni, Zhen
    2023 IEEE POWER & ENERGY SOCIETY GENERAL MEETING, PESGM, 2023,
  • [33] Maximum Entropy Inverse Reinforcement Learning in Continuous State Spaces with Path Integrals
    Aghasadeghi, Navid
    Bretl, Timothy
    2011 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, 2011, : 1561 - 1566
  • [34] Uncertainty-aware human-like driving policy learning with deep Bayesian inverse reinforcement learning
    Zeng, Di
    Zheng, Ling
    Yang, Xiantong
    Li, Yinong
    TRANSPORTMETRICA A-TRANSPORT SCIENCE, 2024,
  • [35] Soft Policy Gradient Method for Maximum Entropy Deep Reinforcement Learning
    Shi, Wenjie
    Song, Shiji
    Wu, Cheng
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3425 - 3431
  • [36] Conditional Predictive Behavior Planning With Inverse Reinforcement Learning for Human-Like Autonomous Driving
    Huang, Zhiyu
    Liu, Haochen
    Wu, Jingda
    Lv, Chen
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (07) : 7244 - 7258
  • [37] Parameterized MDPs and Reinforcement Learning Problems--A Maximum Entropy Principle-Based Framework
    Srivastava, Amber
    Salapaka, Srinivasa M.
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (09) : 9339 - 9351
  • [38] Allocation method of communication interference resource based on deep reinforcement learning of maximum policy entropy
    Rao, Ning
    Xu, Hua
    Qi, Zisen
    Song, Bailin
    Shi, Yunhao
    Xibei Gongye Daxue Xuebao/Journal of Northwestern Polytechnical University, 2021, 39 (05): : 1077 - 1086
  • [39] Driving Decision and Control for Automated Lane Change Behavior based on Deep Reinforcement Learning
    Shi, Tianyu
    Wang, Pin
    Cheng, Xuxin
    Chan, Ching-Yao
    Huang, Ding
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 2895 - 2900
  • [40] Human-like Highway Trajectory Modeling based on Inverse Reinforcement Learning
    Sun, Ruoyu
    Hu, Shaochi
    Zhao, Huijing
    Moze, Mathieu
    Aioun, Francois
    Guillemard, Franck
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 1482 - 1489