Learning state importance for preference-based reinforcement learning

被引:5
|
作者
Zhang, Guoxi [1 ]
Kashima, Hisashi [1 ,2 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Yoshida Honmachi, Kyoto 6068501, Japan
[2] RIKEN Guardian Robot Project, Kyoto, Japan
关键词
Interpretable reinforcement learning; Preference-based reinforcement learning; Human-in-the-loop reinforcement learning; Interpretability artificial intelligence;
D O I
10.1007/s10994-022-06295-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Preference-based reinforcement learning (PbRL) develops agents using human preferences. Due to its empirical success, it has prospect of benefiting human-centered applications. Meanwhile, previous work on PbRL overlooks interpretability, which is an indispensable element of ethical artificial intelligence (AI). While prior art for explainable AI offers some machinery, there lacks an approach to select samples to construct explanations. This becomes an issue for PbRL, as transitions relevant to task solving are often outnumbered by irrelevant ones. Thus, ad-hoc sample selection undermines the credibility of explanations. The present study proposes a framework for learning reward functions and state importance from preferences simultaneously. It offers a systematic approach for selecting samples when constructing explanations. Moreover, the present study proposes a perturbation analysis to evaluate the learned state importance quantitatively. Through experiments on discrete and continuous control tasks, the present study demonstrates the proposed framework's efficacy for providing interpretability without sacrificing task performance.
引用
收藏
页码:1885 / 1901
页数:17
相关论文
共 50 条
  • [41] Preference-Based Assistance Map Learning With Robust Adaptive Oscillators
    Li, Shilei
    Zou, Wulin
    Duan, Pu
    Shi, Ling
    IEEE TRANSACTIONS ON MEDICAL ROBOTICS AND BIONICS, 2022, 4 (04): : 1000 - 1009
  • [42] Preference-based decision making for personalised access to Learning Resources
    Department of Special Education, University of Thessaly, Argonafton and Filellinon Street, Volos, GR 38221, Greece
    不详
    不详
    Int. J. Auton. Adapt. Commun. Syst., 2008, 3 (356-369):
  • [43] Contextual Bandits and Imitation Learning with Preference-Based Active Queries
    Sekhari, Ayush
    Sridharan, Karthik
    Sun, Wen
    Wu, Runzhe
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [44] A Policy Iteration Algorithm for Learning from Preference-Based Feedback
    Wirth, Christian
    Furnkranz, Johannes
    ADVANCES IN INTELLIGENT DATA ANALYSIS XII, 2013, 8207 : 427 - 437
  • [45] Active Preference-Based Gaussian Process Regression for Reward Learning
    Biyik, Lirdem
    Huynh, Nicolas
    Kochenderfer, Mykel J.
    Sadigh, Dorsa
    ROBOTICS: SCIENCE AND SYSTEMS XVI, 2020,
  • [46] Preference-based experience sharing scheme for multi-agent reinforcement learning in multi-target environments
    Zuo, Xuan
    Zhang, Pu
    Li, Hui-Yan
    Liu, Zhun-Ga
    EVOLVING SYSTEMS, 2024, 15 (05) : 1681 - 1699
  • [47] Decoding Global Preferences: Temporal and Cooperative Dependency Modeling in Multi-Agent Preference-Based Reinforcement Learning
    Zhu, Tianchen
    Qiu, Yue
    Zhou, Haoyi
    Li, Jianxin
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 15, 2024, : 17202 - 17210
  • [48] Preference-learning based Inverse Reinforcement Learning for Dialog Control
    Sugiyama, Hiroaki
    Meguro, Toyomi
    Minami, Yasuhiro
    13TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2012 (INTERSPEECH 2012), VOLS 1-3, 2012, : 222 - 225
  • [49] The Method of Learning Personal Preference with Reinforcement Learning
    Park, Yong Hee
    Choi, Won Seok
    Choi, Seong Gon
    2022 24TH INTERNATIONAL CONFERENCE ON ADVANCED COMMUNICATION TECHNOLOGY (ICACT): ARITIFLCIAL INTELLIGENCE TECHNOLOGIES TOWARD CYBERSECURITY, 2022, : 73 - +
  • [50] Active preference-based Gaussian process regression for reward learning and optimization
    Biyik, Erdem
    Huynh, Nicolas
    Kochenderfer, Mykel J.
    Sadigh, Dorsa
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2024, 43 (05): : 665 - 684