Inverse Risk-Sensitive Reinforcement Learning

被引:16
|
作者
Ratliff, Lillian J. [1 ]
Mazumdar, Eric [2 ]
机构
[1] Univ Washington, Dept Elect Engn, Seattle, WA 98195 USA
[2] Univ Calif Berkeley, Dept Elect Engn & Comp Sci, Berkeley, CA 94720 USA
基金
美国国家科学基金会;
关键词
Autonomous systems; Markov processes; optimization; reinforcement learning; CHOICE;
D O I
10.1109/TAC.2019.2926674
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This work addresses the problem of inverse reinforcement learning in Markov decision processes where the decision-making agent is risk-sensitive. In particular, a risk-sensitive reinforcement learning algorithm with convergence guarantees that makes use of coherent risk metrics and models of human decision-making which have their origins in behavioral psychology and economics is presented. The risk-sensitive reinforcement learning algorithm provides the theoretical underpinning for a gradient-based inverse reinforcement learning algorithm that seeks to minimize a loss function defined on the observed behavior. It is shown that the gradient of the loss function with respect to the model parameters is well defined and computable via a contraction map argument. Evaluation of the proposed technique is performed on a Grid World example, a canonical benchmark problem.
引用
收藏
页码:1256 / 1263
页数:8
相关论文
共 50 条
  • [31] Risk-Sensitive Reinforcement Learning Via Entropic-VaR Optimization
    Ni, Xinyi
    Lai, Lifeng
    2022 56TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2022, : 953 - 959
  • [32] A Reinforcement Learning Look at Risk-Sensitive Linear Quadratic Gaussian Control
    Cui, Leilei
    Basar, Tamer
    Jiang, Zhong-Ping
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [33] Robust Reinforcement Learning for Risk-Sensitive Linear Quadratic Gaussian Control
    Cui, Leilei
    Basar, Tamer
    Jiang, Zhong-Ping
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (11) : 7678 - 7693
  • [34] Learning Bounds for Risk-sensitive Learning
    Lee, Jaeho
    Park, Sejun
    Shin, Jinwoo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [35] Risk-sensitive online learning
    Even-Dar, Eyal
    Kearns, Michael
    Wortman, Jennifer
    ALGORITHMIC LEARNING THEORY, PROCEEDINGS, 2006, 4264 : 199 - 213
  • [36] Exponential TD Learning: A Risk-Sensitive Actor-Critic Reinforcement Learning Algorithm
    Noorani, Erfaun
    Mavridis, Christos N.
    Baras, John S.
    2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 4104 - 4109
  • [37] Exponential Bellman Equation and Improved Regret Bounds for Risk-Sensitive Reinforcement Learning
    Fei, Yingjie
    Yang, Zhuoran
    Chen, Yudong
    Wang, Zhaoran
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [38] Mean-variance Based Risk-sensitive Reinforcement Learning with Interpretable Attention
    Kim, Woo Kyung
    Lee, Youngseok
    Woo, Honguk
    PROCEEDINGS OF THE 5TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, ICMVA 2022, 2022, : 104 - 109
  • [39] Sample-Efficient Multimodal Dynamics Modeling for Risk-Sensitive Reinforcement Learning
    Yashima, Ryota
    Yamaguchi, Akihiko
    Hashimoto, Koichi
    2022 8TH INTERNATIONAL CONFERENCE ON MECHATRONICS AND ROBOTICS ENGINEERING (ICMRE 2022), 2022, : 21 - 27
  • [40] RiskQ: Risk-sensitive Multi-Agent Reinforcement Learning Value Factorization
    Shen, Siqi
    Ma, Chennan
    Li, Chao
    Liu, Weiquan
    Fu, Yongquan
    Mei, Songzhu
    Liu, Xinwang
    Wang, Cheng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,