Learning Transferable Policies for Monocular Reactive MAV Control

被引:13
|
作者
Daftry, Shreyansh [1 ]
Bagnell, J. Andrew [1 ]
Hebert, Martial [1 ]
机构
[1] Carnegie Mellon Univ, Inst Robot, Pittsburgh, PA 15213 USA
关键词
Transfer learning; Domain adaptation; Reactive control; Autonomous monocular navigation; Micro aerial vehicles;
D O I
10.1007/978-3-319-50115-4_1
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
The ability to transfer knowledge gained in previous tasks into new contexts is one of the most important mechanisms of human learning. Despite this, adapting autonomous behavior to be reused in partially similar settings is still an open problem in current robotics research. In this paper, we take a small step in this direction and propose a generic framework for learning transferable motion policies. Our goal is to solve a learning problem in a target domain by utilizing the training data in a different but related source domain. We present this in the context of an autonomous MAV flight using monocular reactive control, and demonstrate the efficacy of our proposed approach through extensive real-world flight experiments in outdoor cluttered environments.
引用
收藏
页码:3 / 11
页数:9
相关论文
共 50 条
  • [31] Reinforcement learning for reactive power control
    Vlachogiannis, JG
    Hatziargyriou, ND
    IEEE TRANSACTIONS ON POWER SYSTEMS, 2004, 19 (03) : 1317 - 1325
  • [32] Development of a Transferable Reactive Force Field for Cobalt
    LaBrosse, Matthew R.
    Johnson, J. Karl
    van Duin, Adri C. T.
    JOURNAL OF PHYSICAL CHEMISTRY A, 2010, 114 (18): : 5855 - 5861
  • [33] DEEP REINFORCEMENT LEARNING FOR TRANSFER OF CONTROL POLICIES
    Cunningham, James D.
    Miller, Simon W.
    Yukish, Michael A.
    Simpson, Timothy W.
    Tucker, Conrad S.
    PROCEEDINGS OF THE ASME INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, 2019, VOL 2A, 2020,
  • [34] Data Efficient Learning of Robust Control Policies
    Jha, Susmit
    Lincoln, Patrick
    2018 56TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2018, : 856 - 861
  • [35] Learning Force Control Policies for Compliant Manipulation
    Kalakrishnan, Mrinal
    Righetti, Ludovic
    Pastor, Peter
    Schaal, Stefan
    2011 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, 2011, : 4639 - 4644
  • [36] EpiCURB: Learning to Derive Epidemic Control Policies
    Rusu A.C.
    Farrahi K.
    Niranjan M.
    IEEE Pervasive Computing, 2024, 23 (01) : 57 - 62
  • [37] Reactive Avoidance Using Embedded Stereo Vision for MAV Flight
    Oleynikova, Helen
    Honegger, Dominik
    Pollefeys, Marc
    2015 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2015, : 50 - 56
  • [38] Learning and generalizing force control policies for sculpting
    Koropouli, Vasiliki
    Hirche, Sandra
    Lee, Dongheui
    2012 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2012, : 1493 - 1498
  • [39] Learning Continuous-Action Control Policies
    Pazis, Jason
    Lagoudakis, Michail G.
    ADPRL: 2009 IEEE SYMPOSIUM ON ADAPTIVE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2009, : 169 - 176
  • [40] Learning Control Policies from Optimal Trajectories
    Zelch, Christoph
    Peters, Jan
    von Stryk, Oskar
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 2529 - 2535