Learning state importance for preference-based reinforcement learning

被引:5
|
作者
Zhang, Guoxi [1 ]
Kashima, Hisashi [1 ,2 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Yoshida Honmachi, Kyoto 6068501, Japan
[2] RIKEN Guardian Robot Project, Kyoto, Japan
关键词
Interpretable reinforcement learning; Preference-based reinforcement learning; Human-in-the-loop reinforcement learning; Interpretability artificial intelligence;
D O I
10.1007/s10994-022-06295-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Preference-based reinforcement learning (PbRL) develops agents using human preferences. Due to its empirical success, it has prospect of benefiting human-centered applications. Meanwhile, previous work on PbRL overlooks interpretability, which is an indispensable element of ethical artificial intelligence (AI). While prior art for explainable AI offers some machinery, there lacks an approach to select samples to construct explanations. This becomes an issue for PbRL, as transitions relevant to task solving are often outnumbered by irrelevant ones. Thus, ad-hoc sample selection undermines the credibility of explanations. The present study proposes a framework for learning reward functions and state importance from preferences simultaneously. It offers a systematic approach for selecting samples when constructing explanations. Moreover, the present study proposes a perturbation analysis to evaluate the learned state importance quantitatively. Through experiments on discrete and continuous control tasks, the present study demonstrates the proposed framework's efficacy for providing interpretability without sacrificing task performance.
引用
收藏
页码:1885 / 1901
页数:17
相关论文
共 50 条
  • [1] Learning state importance for preference-based reinforcement learning
    Guoxi Zhang
    Hisashi Kashima
    Machine Learning, 2024, 113 : 1885 - 1901
  • [2] Preference-Based Policy Iteration: Leveraging Preference Learning for Reinforcement Learning
    Cheng, Weiwei
    Fuernkranz, Johannes
    Huellermeier, Eyke
    Park, Sang-Hyeun
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT I, 2011, 6911 : 312 - 327
  • [4] A Survey of Preference-Based Reinforcement Learning Methods
    Wirth, Christian
    Akrour, Riad
    Neumann, Gerhard
    Fuernkranz, Johannes
    JOURNAL OF MACHINE LEARNING RESEARCH, 2017, 18
  • [5] Model-Free Preference-Based Reinforcement Learning
    Wirth, Christian
    Fuernkranz, Johannes
    Neumann, Gerhard
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2222 - 2228
  • [6] Dueling Posterior Sampling for Preference-Based Reinforcement Learning
    Novoseller, Ellen R.
    Wei, Yibing
    Sui, Yanan
    Yue, Yisong
    Burdick, Joel W.
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI 2020), 2020, 124 : 1029 - 1038
  • [7] Preference-based reinforcement learning: evolutionary direct policy search using a preference-based racing algorithm
    Róbert Busa-Fekete
    Balázs Szörényi
    Paul Weng
    Weiwei Cheng
    Eyke Hüllermeier
    Machine Learning, 2014, 97 : 327 - 351
  • [8] Preference-based reinforcement learning: evolutionary direct policy search using a preference-based racing algorithm
    Busa-Fekete, Robert
    Szoerenyi, Balazs
    Weng, Paul
    Cheng, Weiwei
    Huellermeier, Eyke
    MACHINE LEARNING, 2014, 97 (03) : 327 - 351
  • [9] Efficient Meta Reinforcement Learning for Preference-based Fast Adaptation
    Ren, Zhizhou
    Liu, Anji
    Liang, Yitao
    Peng, Jian
    Ma, Jianzhu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [10] Preference-based Reinforcement Learning with Finite-Time Guarantees
    Xu, Yichong
    Wang, Ruosong
    Yang, Lin F.
    Singh, Aarti
    Dubrawski, Artur
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33