Learning state importance for preference-based reinforcement learning

被引:5
|
作者
Zhang, Guoxi [1 ]
Kashima, Hisashi [1 ,2 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Yoshida Honmachi, Kyoto 6068501, Japan
[2] RIKEN Guardian Robot Project, Kyoto, Japan
关键词
Interpretable reinforcement learning; Preference-based reinforcement learning; Human-in-the-loop reinforcement learning; Interpretability artificial intelligence;
D O I
10.1007/s10994-022-06295-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Preference-based reinforcement learning (PbRL) develops agents using human preferences. Due to its empirical success, it has prospect of benefiting human-centered applications. Meanwhile, previous work on PbRL overlooks interpretability, which is an indispensable element of ethical artificial intelligence (AI). While prior art for explainable AI offers some machinery, there lacks an approach to select samples to construct explanations. This becomes an issue for PbRL, as transitions relevant to task solving are often outnumbered by irrelevant ones. Thus, ad-hoc sample selection undermines the credibility of explanations. The present study proposes a framework for learning reward functions and state importance from preferences simultaneously. It offers a systematic approach for selecting samples when constructing explanations. Moreover, the present study proposes a perturbation analysis to evaluate the learned state importance quantitatively. Through experiments on discrete and continuous control tasks, the present study demonstrates the proposed framework's efficacy for providing interpretability without sacrificing task performance.
引用
收藏
页码:1885 / 1901
页数:17
相关论文
共 50 条
  • [21] Task Decoupling in Preference-based Reinforcement Learning for Personalized Human-Robot Interaction
    Liu, Mingjiang
    Chen, Chunlin
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 848 - 855
  • [22] Can You Rely on Synthetic Labellers in Preference-Based Reinforcement Learning? It's Complicated
    Metcalf, Katherine
    Sarabia, Miguel
    Fedzechkina, Masha
    Theobald, Barry-John
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 9, 2024, : 10128 - 10136
  • [23] A Leading Cruise Controller for Autonomous Vehicles in Mixed Autonomy Based on Preference-Based Reinforcement Learning
    Wen, Xiao
    Jian, Sisi
    He, Dengbo
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 752 - 757
  • [24] Mixing corrupted preferences for robust and feedback-efficient preference-based reinforcement learning
    Heo, Jongkook
    Lee, Young Jae
    Kim, Jaehoon
    Kwak, Min Gu
    Park, Young Joon
    Kim, Seoung Bum
    KNOWLEDGE-BASED SYSTEMS, 2025, 309
  • [25] A Preference-Based Online Reinforcement Learning With Embedded Communication Failure Solutions in Smart Grid
    Dang, Yibing
    Xu, Jiangjiao
    Li, Dongdong
    Sun, Hongjian
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2025, 21 (03) : 2422 - 2431
  • [26] Human-Guided Robot Behavior Learning: A GAN-Assisted Preference-Based Reinforcement Learning Approach
    Zhan, Huixin
    Tao, Feng
    Cao, Yongcan
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02): : 3545 - 3552
  • [27] Task Transfer by Preference-Based Cost Learning
    Jing, Mingxuan
    Ma, Xiaojian
    Huang, Wenbing
    Sun, Fuchun
    Liu, Huaping
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 2471 - 2478
  • [28] Preference-Based Learning for Exoskeleton Gait Optimization
    Tucker, Maegan
    Novoseller, Ellen
    Kann, Claudia
    Sui, Yanan
    Yue, Yisong
    Burdick, Joel W.
    Ames, Aaron D.
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 2351 - 2357
  • [29] Active Preference-Based Learning of Reward Functions
    Sadigh, Dorsa
    Dragan, Anca D.
    Sastry, Shankar
    Seshia, Sanjit A.
    ROBOTICS: SCIENCE AND SYSTEMS XIII, 2017,
  • [30] Learning solution similarity in preference-based CBR
    Abdel-Aziz, Amira
    Strickert, Marc
    Hüllermeier, Eyke
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2014, 8765 : 17 - 31