Hierarchical Interpretable Imitation Learning for End-to-End Autonomous Driving

被引:67
|
作者
Teng, Siyu [1 ,2 ]
Chen, Long [3 ,4 ]
Ai, Yunfeng [5 ]
Zhou, Yuanye [6 ]
Xuanyuan, Zhe [1 ]
Hu, Xuemin [7 ]
机构
[1] HKBU United Int Coll, BNU, Zhuhai 999077, Peoples R China
[2] Hong Kong Baptist Univ, Kowloon, Hong Kong 999077, Peoples R China
[3] Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
[4] Waytous Inc Qingdao, Qingdao 266109, Peoples R China
[5] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[6] Malardalen Univ, S-72214 Vasteras, Sweden
[7] Hubei Univ, Sch Comp Sci & Informat Engn, Wuhan 430062, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
Semantics; Data models; Autonomous vehicles; Cameras; Reinforcement learning; Predictive models; Robustness; Autonomous driving; imitation learning; motion planning; end-to-End driving; interpretability;
D O I
10.1109/TIV.2022.3225340
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
End-to-end autonomous driving provides a simple and efficient framework for autonomous driving systems, which can directly obtain control commands from raw perception data. However, it fails to address stability and interpretability problems in complex urban scenarios. In this paper, we construct a two-stage end-to-end autonomous driving model for complex urban scenarios, named HIIL (Hierarchical Interpretable Imitation Learning), which integrates interpretable BEV mask and steering angle to solve the problems shown above. In Stage One, we propose a pretrained Bird's Eye View (BEV) model which leverages a BEV mask to present an interpretation of the surrounding environment. In Stage Two, we construct an Interpretable Imitation Learning (IIL) model that fuses BEV latent feature from Stage One with an additional steering angle from Pure-Pursuit (PP) algorithm. In the HIIL model, visual information is converted to semantic images by the semantic segmentation network, and the semantic images are encoded to extract the BEV latent feature, which are decoded to predict BEV masks and fed to the IIL as perception data. In this way, the BEV latent feature bridges the BEV and IIL models. Visual information can be supplemented by the calculated steering angle for PP algorithm, speed vector, and location information, thus it could have better performance in complex and terrible scenarios. Our HIIL model meets an urgent requirement for interpretability and robustness of autonomous driving. We validate the proposed model in the CARLA simulator with extensive experiments which show remarkable interpretability, generalization, and robustness capability in unknown scenarios for navigation tasks.
引用
收藏
页码:673 / 683
页数:11
相关论文
共 50 条
  • [21] End-to-End Autonomous Driving Decision Based on Deep Reinforcement Learning
    Huang, Zhiqing
    Zhang, Ji
    Tian, Rui
    Zhang, Yanxin
    CONFERENCE PROCEEDINGS OF 2019 5TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS (ICCAR), 2019, : 658 - 662
  • [22] Evaluation of End-To-End Learning for Autonomous Driving: The Good, the Bad and the Ugly
    Varisteas, Georgios
    Frank, Raphael
    Alamdari, Seyed Amin Sajadi
    Voos, Holger
    State, Radu
    2019 2ND INTERNATIONAL CONFERENCE ON INTELLIGENT AUTONOMOUS SYSTEMS (ICOIAS 2019), 2019, : 110 - 117
  • [23] End-to-End Learning of Behavioural Inputs for Autonomous Driving in Dense Traffic
    Shrestha, Jatan
    Idoko, Simon
    Sharma, Basant
    Singh, Arun Kumar
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2023, : 10020 - 10027
  • [24] Multi-task Learning with Attention for End-to-end Autonomous Driving
    Ishihara, Keishi
    Kanervisto, Anssi
    Miura, Jun
    Hautamaki, Ville
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 2896 - 2905
  • [25] End-to-end Autonomous Driving Perception with Sequential Latent Representation Learning
    Chen, Jianyu
    Xu, Zhuo
    Tomizuka, Masayoshi
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 1999 - 2006
  • [26] End-to-End Autonomous Driving Decision Based on Deep Reinforcement Learning
    Huang Z.-Q.
    Qu Z.-W.
    Zhang J.
    Zhang Y.-X.
    Tian R.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2020, 48 (09): : 1711 - 1719
  • [27] Learning End-to-end Autonomous Driving using Guided Auxiliary Supervision
    Mehta, Ashish
    Subramanian, Adithya
    Subramanian, Anbumani
    ELEVENTH INDIAN CONFERENCE ON COMPUTER VISION, GRAPHICS AND IMAGE PROCESSING (ICVGIP 2018), 2018,
  • [28] End-to-end deep learning for reverse driving trajectory of autonomous bulldozer
    You, Ke
    Ding, Lieyun
    Jiang, Yutian
    Wu, Zhangang
    Zhou, Cheng
    KNOWLEDGE-BASED SYSTEMS, 2022, 252
  • [29] DriveGPT4: Interpretable End-to-End Autonomous Driving Via Large Language Model
    Xu, Zhenhua
    Zhang, Yujia
    Xie, Enze
    Zhao, Zhen
    Guo, Yong
    Wong, Kwan-Yee K.
    Li, Zhenguo
    Zhao, Hengshuang
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (10): : 8186 - 8193
  • [30] GenAD: Generative End-to-End Autonomous Driving
    Zheng, Wenzhao
    Song, Ruiqi
    Guo, Xianda
    Zhan, Chenming
    Chen, Long
    COMPUTER VISION - ECCV 2024, PT LXV, 2025, 15123 : 87 - 104