Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning

被引:0
|
作者
Wang, Yongjie [1 ]
Niu, Yuchen [1 ]
Xiao, Mei [1 ]
Zhu, Wenying [1 ]
You, Xinshang [2 ]
机构
[1] Changan Univ, Sch Transportat Engn, Xian 710064, Peoples R China
[2] Hebei Univ Sci & Technol, Sch Econ & Management, Shijiazhuang 050018, Peoples R China
基金
中国国家自然科学基金;
关键词
Human driving behavior; Autonomous vehicle; Occluded pedestrian; Inverse reinforcement learning; Reinforcement learning; PERFORMANCE;
D O I
10.1016/j.physa.2024.130052
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Driving behavior modeling is extremely crucial for designing safe, intelligent, and personalized autonomous driving systems. In this paper, a modeling framework based on Markov Decision Processes (MDPs) is introduced that emulates drivers' decision-making processes. The framework combines the Deep Maximum Entropy Inverse Reinforcement Learning (Deep MEIRL) and a reinforcement learning algorithm-proximal strategy optimization (PPO). A neural network structure is customized for Deep MEIRL, which uses the velocity of the ego vehicle, the pedestrian position, the velocity of surrounding vehicles, the lateral distance, the surrounding vehicles' type, and the distance to the crosswalk to recover the nonlinear reward function. The dataset of drone-based video footage is collected in Xi'an (China) to train and validate the framework. The outcomes demonstrate that Deep MEIRL-PPO outperforms traditional modeling frameworks (Maximum Entropy Inverse Reinforcement Learning (MEIRL)- PPO) in modeling and predicting human driving behavior. Specifically, in predicting human driving behavior, Deep MEIRL-PPO outperforms MEIRL-PPO by 50.71% and 43.90% on the basis of the MAE and HD, respectively. Furthermore, it is discovered that Deep MEIRL-PPO accurately learns the behavior of human drivers avoiding potential conflicts when lines of sight are occluded. This research can contribute to aiding self-driving vehicles in learning human driving behavior and avoiding unforeseen risks.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Deep Reinforcement Learning Based Train Driving Optimization
    Huang, Jin
    Zhang, Ende
    Zhang, Jiarui
    Huang, Siguang
    Zhong, Zhihua
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 2375 - 2381
  • [42] WISEMOVE: A Framework to Investigate Safe Deep Reinforcement Learning for Autonomous Driving
    Lee, Jaeyoung
    Balakrishnan, Aravind
    Gaurav, Ashish
    Czarnecki, Krzysztof
    Sedwards, Sean
    QUANTITATIVE EVALUATION OF SYSTEMS (QEST 2019), 2019, 11785 : 350 - 354
  • [43] Modeling Driver Behavior using Adversarial Inverse Reinforcement Learning
    Sackmann, Moritz
    Bey, Henrik
    Hofmann, Ulrich
    Thielecke, Joern
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1683 - 1690
  • [44] Off-policy asymptotic and adaptive maximum entropy deep reinforcement learning
    Zhang, Huihui
    Han, Xu
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, : 2417 - 2429
  • [45] A latent space method with maximum entropy deep reinforcement learning for data assimilation
    Zhang, Jinding
    Zhang, Kai
    Wang, Zhongzheng
    Zhou, Wensheng
    Liu, Chen
    Zhang, Liming
    Ma, Xiaopeng
    Liu, Piyang
    Bian, Ziwei
    Kang, Jinzheng
    Yang, Yongfei
    Yao, Jun
    GEOENERGY SCIENCE AND ENGINEERING, 2024, 243
  • [46] CADRE: A Cascade Deep Reinforcement Learning Framework for Vision-Based Autonomous Urban Driving
    Zhao, Yinuo
    Wu, Kun
    Xu, Zhiyuan
    Che, Zhengping
    Lu, Qi
    Tang, Jian
    Liu, Chi Harold
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3481 - 3489
  • [47] Active Perception in Adversarial Scenarios using Maximum Entropy Deep Reinforcement Learning
    Shen, Macheng
    How, Jonathan P.
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 3384 - 3390
  • [48] Inverse Reinforcement Learning Based Stochastic Driver Behavior Learning
    Ozkan, Mehmet F.
    Rocque, Abishek J.
    Ma, Yao
    IFAC PAPERSONLINE, 2021, 54 (20): : 882 - 888
  • [49] Driver Behavior Modeling via Inverse Reinforcement Learning Based on Particle Swarm Optimization
    Liu, Zeng-Jie
    Wu, Huai-Ning
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 7232 - 7237
  • [50] Bayesian Inverse Reinforcement Learning-based Reward Learning for Automated Driving
    Zeng, Di
    Zheng, Ling
    Li, Yinong
    Yang, Xiantong
    Jixie Gongcheng Xuebao/Journal of Mechanical Engineering, 2024, 60 (10): : 245 - 260