Trajectory Planning With Deep Reinforcement Learning in High-Level Action Spaces

被引:7
|
作者
Williams, Kyle R. [1 ]
Schlossman, Rachel [1 ]
Whitten, Daniel [1 ]
Ingram, Joe
Musuvathy, Srideep [1 ]
Pagan, James [1 ]
Williams, Kyle A. [1 ]
Green, Sam [2 ]
Patel, Anirudh [2 ]
Mazumdar, Anirban [3 ]
Parish, Julie [1 ]
机构
[1] Sandia Natl Labs, Albuquerque, CA 94551 USA
[2] Semiot Labs, Los Altos, CA 94022 USA
[3] Georgia Inst Technol, Atlanta, GA 30332 USA
关键词
Trajectory; Planning; Trajectory planning; Training; Reinforcement learning; Optimization; Aerodynamics; OPTIMIZATION;
D O I
10.1109/TAES.2022.3218496
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
This article presents a technique for trajectory planning based on parameterized high-level actions. These high-level actions are subtrajectories that have variable shape and duration. The use of high-level actions can improve the performance of guidance algorithms. Specifically, we show how the use of high-level actions improves the performance of guidance policies that are generated via reinforcement learning (RL). RL has shown great promise for solving complex control, guidance, and coordination problems but can still suffer from long training times and poor performance. This work shows how the use of high-level actions reduces the required number of training steps and increases the path performance of an RL-trained guidance policy. We demonstrate the method on a space-shuttle guidance example. We show the proposed method increases the path performance (latitude range) by 18% compared with a baseline RL implementation. Similarly, we show the proposed method achieves steady state during training with approximately 75% fewer training steps. We also show how the guidance policy enables effective performance in an obstacle field. Finally, this article develops a loss function term for policy-gradient-based deep RL, which is analogous to an antiwindup mechanism in feedback control. We demonstrate that the inclusion of this term in the underlying optimization increases the average policy return in our numerical example.
引用
收藏
页码:2513 / 2529
页数:17
相关论文
共 50 条
  • [1] Using Deep Reinforcement Learning to Learn High-Level Policies on the ATRIAS Biped
    Li, Tianyu
    Geyer, Hartmut
    Atkeson, Christopher G.
    Rai, Akshara
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 263 - 269
  • [2] Deep Distributional Reinforcement Learning Based High-Level Driving Policy Determination
    Min, Kyushik
    Kim, Hayoung
    Huh, Kunsoo
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2019, 4 (03): : 416 - 424
  • [3] DEEP REINFORCEMENT LEARNING IN LINEAR DISCRETE ACTION SPACES
    van Heeswijk, Wouter
    La Poutre, Han
    2020 WINTER SIMULATION CONFERENCE (WSC), 2020, : 1063 - 1074
  • [4] Global Multi-Phase Path Planning Through High-Level Reinforcement Learning
    Salamat, Babak
    Olzem, Sebastian-Sven
    Elsbacher, Gerhard
    Tonello, Andrea M.
    IEEE OPEN JOURNAL OF CONTROL SYSTEMS, 2024, 3 : 405 - 415
  • [5] Deep Reinforcement Learning for Autonomous Driving using High-Level Heterogeneous Graph Representations
    Schier, Maximilian
    Reinders, Christoph
    Rosenhahn, Bodo
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 7147 - 7153
  • [6] A Deep Reinforcement Learning Approach for Federated Learning Optimization with UAV Trajectory Planning
    Zhang, Chunyu
    Liu, Yiming
    Zhang, Zhi
    2023 IEEE 34TH ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS, PIMRC, 2023,
  • [7] Learning Portable Representations for High-Level Planning
    James, Steven
    Rosman, Benjamin
    Konidaris, George
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [8] Hierarchical Skill Learning for High-level Planning
    MacGlashan, James
    PROCEEDINGS OF THE TWENTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-10), 2010, : 1988 - 1989
  • [9] Reinforcement learning for instance segmentation with high-level priors
    Hilt, Paul
    Zarvandi, Maedeh
    Kaziakhmedov, Edgar
    Bhide, Sourabh
    Laptin, Maria
    Pape, Constantin
    Kreshuk, Anna
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 3915 - 3924
  • [10] Learning to explore by reinforcement over high-level options
    Juncheng Liu
    Brendan McCane
    Steven Mills
    Machine Vision and Applications, 2024, 35