A Closer Look at Reward Decomposition for High-level Robotic Explanations

被引:3
|
作者
Lu, Wenhao [1 ]
Zhao, Xufeng [1 ]
Magg, Sven [2 ]
Gromniak, Martin [1 ,3 ]
Li, Mengdi [1 ]
Wermter, Stefan [1 ]
机构
[1] Univ Hamburg, Dept Informat, Knowledge Technol Grp, Hamburg, Germany
[2] Hamburger Informat Technol Ctr HITeC, Hamburg, Germany
[3] ZAL Ctr Appl Aeronaut Res, Hamburg, Germany
关键词
D O I
10.1109/ICDL55364.2023.10364407
中图分类号
B84 [心理学]; C [社会科学总论]; Q98 [人类学];
学科分类号
03 ; 0303 ; 030303 ; 04 ; 0402 ;
摘要
Explaining the behaviour of intelligent agents learned by reinforcement learning (RL) to humans is challenging yet crucial due to their incomprehensible proprioceptive states, variational intermediate goals, and resultant unpredictability. Moreover, one-step explanations for RL agents can be ambiguous as they fail to account for the agent's future behaviour at each transition, adding to the complexity of explaining robot actions. By leveraging abstracted actions that map to task-specific primitives, we avoid explanations on the movement level. To further improve the transparency and explainability of robotic systems, we propose an explainable Q-Map learning framework that combines reward decomposition (RD) with abstracted action spaces, allowing for non-ambiguous and high-level explanations based on object properties in the task. We demonstrate the effectiveness of our framework through quantitative and qualitative analysis of two robotic scenarios, showcasing visual and textual explanations, from output artefacts of RD explanations, that are easy for humans to comprehend. Additionally, we demonstrate the versatility of integrating these artefacts with large language models (LLMs) for reasoning and interactive querying.
引用
收藏
页码:429 / 436
页数:8
相关论文
共 50 条
  • [1] Using Reward Machines for High-Level Task Specification and Decomposition in Reinforcement Learning
    Icarte, Rodrigo Toro
    Klassen, Toryn Q.
    Valenzano, Richard
    McHraith, Sheila A.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [2] High-Level Why-Not Explanations using Ontologies
    ten Cate, Balder
    Civili, Cristina
    Sherkhonov, Evgeny
    Tan, Wang-Chiew
    PODS'15: PROCEEDINGS OF THE 33RD ACM SYMPOSIUM ON PRINCIPLES OF DATABASE SYSTEMS, 2015, : 31 - 43
  • [3] Reinforcement Learning with Task Decomposition and Task-Specific Reward System for Automation of High-Level Tasks
    Kwon, Gunam
    Kim, Byeongjun
    Kwon, Nam Kyu
    BIOMIMETICS, 2024, 9 (04)
  • [4] Towards Minimal Explanations of Unsynthesizability for High-Level Robot Behaviors
    Raman, Vasumathi
    Kress-Gazit, Hadas
    2013 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2013, : 757 - 762
  • [5] A closer look at dark photon explanations of the excess radio background
    Acharya, Sandeep Kumar
    Chluba, Jens
    MONTHLY NOTICES OF THE ROYAL ASTRONOMICAL SOCIETY, 2023, 521 (03) : 3939 - 3950
  • [6] HIGH-LEVEL ACTIONS HURT DAIRY INDUSTRY, NEED CLOSER SCRUTINY
    SAAL, H
    AMERICAN DAIRY REVIEW, 1973, 35 (09): : 14 - &
  • [7] Hierarchical goals contextualize local reward decomposition explanations
    Rietz, Finn
    Magg, Sven
    Heintz, Fredrik
    Stoyanov, Todor
    Wermter, Stefan
    Stork, Johannes A.
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (23): : 16693 - 16704
  • [8] Hierarchical goals contextualize local reward decomposition explanations
    Finn Rietz
    Sven Magg
    Fredrik Heintz
    Todor Stoyanov
    Stefan Wermter
    Johannes A. Stork
    Neural Computing and Applications, 2023, 35 : 16693 - 16704
  • [9] THE NIOSH ACTION LEVEL - A CLOSER LOOK
    ROCK, JC
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 1980, 180 (AUG): : 34 - CHAS
  • [10] Optimizing Gate Decomposition for High-Level Quantum Programming
    Rosa, Evandro C. R.
    Duzzioni, Eduardo I.
    de Santiago, Rafael
    QUANTUM, 2025, 9