Hierarchical Interpretable Imitation Learning for End-to-End Autonomous Driving

被引:67
|
作者
Teng, Siyu [1 ,2 ]
Chen, Long [3 ,4 ]
Ai, Yunfeng [5 ]
Zhou, Yuanye [6 ]
Xuanyuan, Zhe [1 ]
Hu, Xuemin [7 ]
机构
[1] HKBU United Int Coll, BNU, Zhuhai 999077, Peoples R China
[2] Hong Kong Baptist Univ, Kowloon, Hong Kong 999077, Peoples R China
[3] Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
[4] Waytous Inc Qingdao, Qingdao 266109, Peoples R China
[5] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[6] Malardalen Univ, S-72214 Vasteras, Sweden
[7] Hubei Univ, Sch Comp Sci & Informat Engn, Wuhan 430062, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Semantics; Data models; Autonomous vehicles; Cameras; Reinforcement learning; Predictive models; Robustness; Autonomous driving; imitation learning; motion planning; end-to-End driving; interpretability;
D O I
10.1109/TIV.2022.3225340
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
End-to-end autonomous driving provides a simple and efficient framework for autonomous driving systems, which can directly obtain control commands from raw perception data. However, it fails to address stability and interpretability problems in complex urban scenarios. In this paper, we construct a two-stage end-to-end autonomous driving model for complex urban scenarios, named HIIL (Hierarchical Interpretable Imitation Learning), which integrates interpretable BEV mask and steering angle to solve the problems shown above. In Stage One, we propose a pretrained Bird's Eye View (BEV) model which leverages a BEV mask to present an interpretation of the surrounding environment. In Stage Two, we construct an Interpretable Imitation Learning (IIL) model that fuses BEV latent feature from Stage One with an additional steering angle from Pure-Pursuit (PP) algorithm. In the HIIL model, visual information is converted to semantic images by the semantic segmentation network, and the semantic images are encoded to extract the BEV latent feature, which are decoded to predict BEV masks and fed to the IIL as perception data. In this way, the BEV latent feature bridges the BEV and IIL models. Visual information can be supplemented by the calculated steering angle for PP algorithm, speed vector, and location information, thus it could have better performance in complex and terrible scenarios. Our HIIL model meets an urgent requirement for interpretability and robustness of autonomous driving. We validate the proposed model in the CARLA simulator with extensive experiments which show remarkable interpretability, generalization, and robustness capability in unknown scenarios for navigation tasks.
引用
收藏
页码:673 / 683
页数:11
相关论文
共 50 条
  • [41] Enhancing scene understanding based on deep learning for end-to-end autonomous driving
    Hu, Jie
    Kong, Huifang
    Zhang, Qian
    Liu, Runwu
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 116
  • [42] Towards End-to-End Escape in Urban Autonomous Driving Using Reinforcement Learning
    Sakhai, Mustafa
    Wielgosz, Maciej
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 2, INTELLISYS 2023, 2024, 823 : 21 - 40
  • [43] End-to-end Learning Approach for Autonomous Driving: A Convolutional Neural Network Model
    Wang, Yaqin
    Liu, Dongfang
    Jeon, Hyewon
    Chu, Zhiwei
    Matson, Eric T.
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 833 - 839
  • [44] End-to-End Urban Autonomous Driving With Safety Constraints
    Hou, Changmeng
    Zhang, Wei
    IEEE ACCESS, 2024, 12 : 132198 - 132209
  • [45] A Review of End-to-End Autonomous Driving in Urban Environments
    Coelho, Daniel
    Oliveira, Miguel
    IEEE ACCESS, 2022, 10 : 75296 - 75311
  • [46] SEECAD: Semantic End-to-End Communication for Autonomous Driving
    Ribouh, Soheyb
    Hadid, Abdenour
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 1808 - 1813
  • [47] End-to-end Spatiotemporal Attention Model for Autonomous Driving
    Zhao, Ruijie
    Zhang, Yanxin
    Huang, Zhiqing
    Yin, Chenkun
    PROCEEDINGS OF 2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2020), 2020, : 2649 - 2653
  • [48] Bayesian Imitation Learning for End-to-End Mobile Manipulation
    Du, Yuqing
    Ho, Daniel
    Alemi, Alexander A.
    Jang, Eric
    Khansari, Mohi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [49] XBG: End-to-End Imitation Learning for Autonomous Behaviour in Human-Robot Interaction and Collaboration
    Cardenas-Perez, Carlos
    Romualdi, Giulio
    Elobaid, Mohamed
    Dafarra, Stefano
    L'Erario, Giuseppe
    Traversaro, Silvio
    Morerio, Pietro
    Del Bue, Alessio
    Pucci, Daniele
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (12): : 11617 - 11624
  • [50] SuperDriverAI: Towards Design and Implementation for End-to-End Learning-Based Autonomous Driving
    Aoki, Shunsuke
    Yamamoto, Issei
    Shiotsuka, Daiki
    Inoue, Yuichi
    Tokuhiro, Kento
    Miwa, Keita
    IEEE Vehicular Networking Conference, VNC, 2023, 2023-April : 195 - 198