MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement Learning

被引:46
|
作者
Li, Quanyi [1 ]
Peng, Zhenghao [2 ]
Feng, Lan [4 ]
Zhang, Qihang [3 ]
Xue, Zhenghai [3 ]
Zhou, Bolei [5 ]
机构
[1] Chinese Univ Hong Kong, Ctr Perceptual & Interact Intelligence, Hong Kong, Peoples R China
[2] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[3] Chinese Univ Hong Kong, Dept Informat Engn, Hong Kong, Peoples R China
[4] Swiss Fed Inst Technol, CH-8092 Zurich, Switzerland
[5] Univ Calif Los Angeles, Los Angeles, CA 90095 USA
关键词
Task analysis; Roads; Reinforcement learning; Benchmark testing; Training; Safety; Autonomous vehicles; autonomous driving; simulation;
D O I
10.1109/TPAMI.2022.3190471
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Driving safely requires multiple capabilities from human and intelligent agents, such as the generalizability to unseen environments, the safety awareness of the surrounding traffic, and the decision-making in complex multi-agent settings. Despite the great success of Reinforcement Learning (RL), most of the RL research works investigate each capability separately due to the lack of integrated environments. In this work, we develop a new driving simulation platform called MetaDrive to support the research of generalizable reinforcement learning algorithms for machine autonomy. MetaDrive is highly compositional, which can generate an infinite number of diverse driving scenarios from both the procedural generation and the real data importing. Based on MetaDrive, we construct a variety of RL tasks and baselines in both single-agent and multi-agent settings, including benchmarking generalizability across unseen scenes, safe exploration, and learning multi-agent traffic. The generalization experiments conducted on both procedurally generated scenarios and real-world scenarios show that increasing the diversity and the size of the training set leads to the improvement of the RL agent's generalizability. We further evaluate various safe reinforcement learning and multi-agent reinforcement learning algorithms in MetaDrive environments and provide the benchmarks. Source code, documentation, and demo video are available at https://metadriverse.github.io/metadrive.
引用
收藏
页码:3461 / 3475
页数:15
相关论文
共 50 条
  • [41] A Deep Reinforcement Learning Approach for Composing Moving IoT Services
    Neiat, Azadeh Ghari
    Bouguettaya, Athman
    Bahutair, Mohammed
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2022, 15 (05) : 2538 - 2550
  • [42] GALOIS: Boosting Deep Reinforcement Learning via Generalizable Logic Synthesis
    Cao, Yushi
    Li, Zhiming
    Yang, Tianpei
    Zhang, Hao
    Zheng, Yan
    Li, Yi
    Hao, Jianye
    Liu, Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [43] Generalizable Task Representation Learning for Offline Meta-Reinforcement Learning with Data Limitations
    Zhou, Renzhe
    Gao, Chen-Xiao
    Zhang, Zongzhang
    Yu, Yang
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 15, 2024, : 17132 - 17140
  • [44] Decompose a Task into Generalizable Subtasks in Multi-Agent Reinforcement Learning
    Tian, Zikang
    Chen, Ruizhi
    Hu, Xing
    Li, Ling
    Zhang, Rui
    Wu, Fan
    Peng, Shaohui
    Guo, Jiaming
    Du, Zidong
    Guo, Qi
    Chen, Yunji
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [45] Generalizable Resource Allocation in Stream Processing via Deep Reinforcement Learning
    Ni, Xiang
    Li, Jing
    Yu, Mo
    Zhou, Wang
    Wu, Kun-Lung
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 857 - 864
  • [46] Pre-Trained Image Encoder for Generalizable Visual Reinforcement Learning
    Yuan, Zhecheng
    Xue, Zhengrong
    Yuan, Bo
    Wang, Xueqian
    Wu, Yi
    Gao, Yang
    Xu, Huazhe
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [47] Learning Human Dynamics in Autonomous Driving Scenarios
    Wang, Jingbo
    Yuan, Ye
    Luo, Zhengyi
    Xie, Kevin
    Lin, Dahua
    Iqbal, Umar
    Fidler, Sanja
    Khamis, Sameh
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 20739 - 20749
  • [48] Research on Anthropomorphic Obstacle Avoidance Trajectory Planning for Adaptive Driving Scenarios Based on Inverse Reinforcement Learning Theory
    Wu, Jian
    Yan, Yang
    Liu, Yulong
    Liu, Yahui
    ENGINEERING, 2024, 33 : 133 - 145
  • [49] Real-Time Fatigue Monitoring System in Diverse Driving Scenarios
    Shajahan, Thasnimol Valuthottiyil
    Srinivasan, Babji
    Srinivasan, Rajagopalan
    2024 IEEE SPACE, AEROSPACE AND DEFENCE CONFERENCE, SPACE 2024, 2024, : 124 - 127
  • [50] Reinforcement and Imitation Learning for Diverse Visuomotor Skills
    Zhu, Yuke
    Wang, Ziyu
    Merel, Josh
    Rusu, Andrei
    Erez, Tom
    Cabi, Serkan
    Tunyasuvunakool, Saran
    Kramar, Janos
    Hadsell, Raia
    de Freitas, Nando
    Heess, Nicolas
    ROBOTICS: SCIENCE AND SYSTEMS XIV, 2018,