A Nash Q-Learning Based Motion Decision Algorithm With Considering Interaction to Traffic Participants

被引:25
|
作者
Xu, Can [1 ]
Zhao, Wanzhong [1 ]
Li, Lin [1 ]
Chen, Qingyun [1 ]
Kuang, Dengming [1 ]
Zhou, Jianhao [1 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Dept Vehicle Engn, Nanjing 210016, Peoples R China
关键词
Trajectory; Prediction algorithms; Acceleration; Predictive models; Autonomous vehicles; Kinematics; Roads; Nash Q-learning; motion decision; interaction; trajectory prediction; highly autonomous driving; DRIVER;
D O I
10.1109/TVT.2020.3027352
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In order to improve the efficiency and comfort of autonomous vehicles while ensuring safety, the decision algorithm needs to interact with human drivers, infer the most probable behavior and then makes advantageous decision. This paper proposes a Nash-Q learning based motion decision algorithm to consider the interaction. First, the local trajectory of surrounding vehicle is predicted by kinematic constraints, which can reflect the short-term motion trend. Then, the future action space is built based the predicted local trajectory that consists of five basis actions. With that, the Nash-Q learning process can be implemented by the game between these basis actions. By elimination of strictly dominated actions and the Lemke-Howson method, the autonomous vehicle can decide the optimal action and infer the behavior of surrounding vehicle. Finally, the lane merging scenario is built to test the performance contrast to the existing methods. The driver in loop experiment is further designed to verify the interaction performance in multi-vehicle traffic. The results show that the Nash-Q learning based algorithm can improve the efficiency and comfort by 15.75% and 20.71% to the Stackelberg game and the no-interaction method respectively while the safety is ensured. It can also make real-time interaction with human drivers in multi-vehicle traffic.
引用
收藏
页码:12621 / 12634
页数:14
相关论文
共 50 条
  • [1] Lane-changing decision method based Nash Q-learning with considering the interaction of surrounding vehicles
    Zhou, Xiaochuan
    Kuang, Dengming
    Zhao, Wanzhong
    Xu, Can
    Feng, Jian
    Wang, Chunyan
    IET INTELLIGENT TRANSPORT SYSTEMS, 2020, 14 (14) : 2064 - 2072
  • [2] Enhancing Nash Q-learning and Team Q-learning mechanisms by using bottlenecks
    Ghazanfari, Behzad
    Mozayani, Nasser
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2014, 26 (06) : 2771 - 2783
  • [3] Motion Planning for Lunar Rover Based on Behavior Decision Field Q-Learning
    Pan, Haining
    Yuan, Ye
    Ju, Hehua
    Cui, Pingyuan
    2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 1834 - 1839
  • [4] Adaptive Traffic Control Algorithm Based on Back-Pressure and Q-Learning
    Maipradit, Arnan
    Gao, Juntao
    Kawakami, Tomoya
    Ito, Minoru
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 1995 - 1999
  • [5] Backward Q-learning: The combination of Sarsa algorithm and Q-learning
    Wang, Yin-Hao
    Li, Tzuu-Hseng S.
    Lin, Chih-Jui
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2013, 26 (09) : 2184 - 2193
  • [6] An ARM-based Q-learning algorithm
    Hsu, Yuan-Pao
    Hwang, Kao-Shing
    Lin, Hsin-Yi
    ADVANCED INTELLIGENT COMPUTING THEORIES AND APPLICATIONS: WITH ASPECTS OF CONTEMPORARY INTELLIGENT COMPUTING TECHNIQUES, 2007, 2 : 11 - +
  • [7] A Nash-Stackelberg Fuzzy Q-Learning Decision Approach in Heterogeneous Cognitive Networks
    Haddad, Majed
    Altman, Zwi
    Elayoubi, Salah Eddine
    Altman, Eitan
    2010 IEEE GLOBAL TELECOMMUNICATIONS CONFERENCE GLOBECOM 2010, 2010,
  • [8] Optimizing traffic flow with Q-learning and genetic algorithm for congestion control
    Deepika, Gitanjali
    Pandove, Gitanjali
    EVOLUTIONARY INTELLIGENCE, 2024, 17 (5-6) : 4179 - 4197
  • [9] Q-learning based Wi-Fi Direct Grouping Algorithm considering Optical Backhaul
    Lim, W.
    2017 XXXIIND GENERAL ASSEMBLY AND SCIENTIFIC SYMPOSIUM OF THE INTERNATIONAL UNION OF RADIO SCIENCE (URSI GASS), 2017,
  • [10] A Q-learning algorithm for Markov decision processes with continuous state spaces
    Hu, Jiaqiao
    Yang, Xiangyu
    Hu, Jian-Qiang
    Peng, Yijie
    SYSTEMS & CONTROL LETTERS, 2024, 187