Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning

被引:0
|
作者
Wang, Yongjie [1 ]
Niu, Yuchen [1 ]
Xiao, Mei [1 ]
Zhu, Wenying [1 ]
You, Xinshang [2 ]
机构
[1] Changan Univ, Sch Transportat Engn, Xian 710064, Peoples R China
[2] Hebei Univ Sci & Technol, Sch Econ & Management, Shijiazhuang 050018, Peoples R China
基金
中国国家自然科学基金;
关键词
Human driving behavior; Autonomous vehicle; Occluded pedestrian; Inverse reinforcement learning; Reinforcement learning; PERFORMANCE;
D O I
10.1016/j.physa.2024.130052
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Driving behavior modeling is extremely crucial for designing safe, intelligent, and personalized autonomous driving systems. In this paper, a modeling framework based on Markov Decision Processes (MDPs) is introduced that emulates drivers' decision-making processes. The framework combines the Deep Maximum Entropy Inverse Reinforcement Learning (Deep MEIRL) and a reinforcement learning algorithm-proximal strategy optimization (PPO). A neural network structure is customized for Deep MEIRL, which uses the velocity of the ego vehicle, the pedestrian position, the velocity of surrounding vehicles, the lateral distance, the surrounding vehicles' type, and the distance to the crosswalk to recover the nonlinear reward function. The dataset of drone-based video footage is collected in Xi'an (China) to train and validate the framework. The outcomes demonstrate that Deep MEIRL-PPO outperforms traditional modeling frameworks (Maximum Entropy Inverse Reinforcement Learning (MEIRL)- PPO) in modeling and predicting human driving behavior. Specifically, in predicting human driving behavior, Deep MEIRL-PPO outperforms MEIRL-PPO by 50.71% and 43.90% on the basis of the MAE and HD, respectively. Furthermore, it is discovered that Deep MEIRL-PPO accurately learns the behavior of human drivers avoiding potential conflicts when lines of sight are occluded. This research can contribute to aiding self-driving vehicles in learning human driving behavior and avoiding unforeseen risks.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Car-Following Behavior Modeling With Maximum Entropy Deep Inverse Reinforcement Learning
    Nan, Jiangfeng
    Deng, Weiwen
    Zhang, Ruzheng
    Zhao, Rui
    Wang, Ying
    Ding, Juan
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (02): : 3998 - 4010
  • [2] RRT-based maximum entropy inverse reinforcement learning for robust and efficient driving behavior prediction
    Hosoma, Shinpei
    Sugasaki, Masato
    Arie, Hiroaki
    Shimosaka, Masamichi
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1353 - 1359
  • [3] Comparison and Deduction of Maximum Entropy Deep Inverse Reinforcement Learning
    Chen, Guannan
    Fu, Yanfang
    Liu, Yu
    Dang, Xiangbin
    Hao, Jiajun
    Liu, Xinchen
    2023 IEEE 2ND INDUSTRIAL ELECTRONICS SOCIETY ANNUAL ON-LINE CONFERENCE, ONCON, 2023,
  • [4] A Study of Continuous Maximum Entropy Deep Inverse Reinforcement Learning
    Chen, Xi-liang
    Cao, Lei
    Xu, Zhi-xiong
    Lai, Jun
    Li, Chen-xi
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2019, 2019
  • [5] Driving Behavior Modeling Using Naturalistic Human Driving Data With Inverse Reinforcement Learning
    Huang, Zhiyu
    Wu, Jingda
    Lv, Chen
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (08) : 10239 - 10251
  • [6] Maximum Entropy Inverse Reinforcement Learning Based on Frenet Frame Sampling for Human-like Autonomous Driving
    Zhang, Tangyike
    Sun, Shuo
    Shi, Jiamin
    Chen, Shitao
    Ang, Marcelo H.
    Xin, Jingmin
    Zheng, Nanning
    2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 1820 - 1827
  • [7] Learning the Car-following Behavior of Drivers Using Maximum Entropy Deep Inverse Reinforcement Learning
    Zhou, Yang
    Fu, Rui
    Wang, Chang
    JOURNAL OF ADVANCED TRANSPORTATION, 2020, 2020
  • [8] Efficient Sampling-Based Maximum Entropy Inverse Reinforcement Learning With Application to Autonomous Driving
    Wu, Zheng
    Sun, Liting
    Zhan, Wei
    Yang, Chenyu
    Tomizuka, Masayoshi
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 5355 - 5362
  • [9] AdaBoost maximum entropy deep inverse reinforcement learning with truncated gradient
    Song, Li
    Li, Dazi
    Wang, Xiao
    Xu, Xin
    INFORMATION SCIENCES, 2022, 602 : 328 - 350
  • [10] Continuous Deep Maximum Entropy Inverse Reinforcement Learning using online POMDP
    Silva, Junior A. R.
    Grassi Jr, Valdir
    Wolf, Denis Fernando
    2019 19TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS (ICAR), 2019, : 382 - 387