Supervised autonomy for online learning in human-robot interaction

被引:30
|
作者
Senft, Emmanuel [1 ]
Baxter, Paul [2 ]
Kennedy, James [1 ]
Lemaignan, Severin [1 ]
Belpaeme, Tony [1 ,3 ]
机构
[1] Plymouth Univ, Plymouth PL4 8AA, Devon, England
[2] Univ Lincoln, Lincoln Ctr Autonomous Syst, Lincoln LN6 7TS, England
[3] Univ Ghent, Dept Elect & Informat Syst, Imec IDLab, Ghent, Belgium
关键词
Human-Robot interaction; Reinforcement learning; Interactive machine learning; Robotics; Progressive Autonomy; Supervised autonomy;
D O I
10.1016/j.patrec.2017.03.015
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When a robot is learning it needs to explore its environment and how its environment responds on its actions. When the environment is large and there are a large number of possible actions the robot can take, this exploration phase can take prohibitively long. However, exploration can often be optimised by letting a human expert guide the robot during its learning. Interactive machine learning, in which a human user interactively guides the robot as it learns, has been shown to be an effective way to teach a robot. It requires an intuitive control mechanism to allow the human expert to provide feedback on the robot's progress. This paper presents a novel method which combines Reinforcement Learning and Supervised Progressively Autonomous Robot Competencies (SPARC). By allowing the user to fully control the robot and by treating rewards as implicit, SPARC aims to learn an action policy while maintaining human supervisory oversight of the robot's behaviour. This method is evaluated and compared to Interactive Reinforcement Learning in a robot teaching task. Qualitative and quantitative results indicate that SPARC allows for safer and faster learning by the robot, whilst not placing a high workload on the human teacher. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:77 / 86
页数:10
相关论文
共 50 条
  • [11] Online Robot Teaching With Natural Human-Robot Interaction
    Du, Guanglong
    Chen, Mingxuan
    Liu, Caibing
    Zhang, Bo
    Zhang, Ping
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2018, 65 (12) : 9571 - 9581
  • [12] Online Learning of Varying Stiffness Through Physical Human-Robot Interaction
    Kronander, Klas
    Billard, Aude
    2012 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2012, : 1842 - 1849
  • [13] Human perception of intrinsically motivated autonomy in human-robot interaction
    Scheunemann, Marcus M.
    Salge, Christoph
    Polani, Daniel
    Dautenhahn, Kerstin
    ADAPTIVE BEHAVIOR, 2022, 30 (05) : 451 - 472
  • [14] Weakly-Supervised Object Detection Learning through Human-Robot Interaction
    Maiettini, Elisa
    Tikhanoff, Vadim
    Natale, Lorenzo
    PROCEEDINGS OF THE 2020 IEEE-RAS 20TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS 2020), 2021, : 392 - 399
  • [15] Autonomy mode suggestions for improving human-robot interaction
    Baker, M
    Yanco, HA
    2004 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN & CYBERNETICS, VOLS 1-7, 2004, : 2948 - 2953
  • [16] Autonomy in Physical Human-Robot Interaction: A Brief Survey
    Selvaggio, Mario
    Cognetti, Marco
    Nikolaidis, Stefanos
    Ivaldi, Serena
    Siciliano, Bruno
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (04) : 7989 - 7996
  • [17] Human-robot interaction through adjustable social autonomy
    Cantucci, Filippo
    Falcone, Rino
    Castelfranchi, Cristiano
    INTELLIGENZA ARTIFICIALE, 2022, 16 (01) : 69 - 79
  • [18] Online Nod Detection in Human-Robot Interaction
    Wall, Eduard
    Schillingmann, Lars
    Kummert, Franz
    2017 26TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2017, : 811 - 817
  • [19] Constructive learning for human-robot interaction
    Singh, Amarjot
    Karanam, Srikrishna
    Kumar, Devinder
    IEEE Potentials, 2013, 32 (04): : 13 - 19
  • [20] Touch-text answer for human-robot interaction via supervised adversarial learning
    Wang, Ya-Xin
    Meng, Qing-Hao
    Li, Yun-Kai
    Hou, Hui-Rang
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 242