Reinforcement Learning Under Moral Uncertainty

被引:0
|
作者
Ecoffet, Adrien [1 ,2 ]
Lehman, Joel [1 ,2 ]
机构
[1] Uber AI Labs, San Francisco, CA 94103 USA
[2] OpenAI, San Francisco, CA 94110 USA
关键词
SOCIAL RATE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An ambitious goal for machine learning is to create agents that behave ethically: The capacity to abide by human moral norms would greatly expand the context in which autonomous agents could be practically and safely deployed, e.g. fully autonomous vehicles will encounter charged moral decisions that complicate their deployment. While ethical agents could be trained by rewarding correct behavior under a specific moral theory (e.g. utilitarianism), there remains widespread disagreement about the nature of morality. Acknowledging such disagreement, recent work in moral philosophy proposes that ethical behavior requires acting under moral uncertainty, i.e. to take into account when acting that one's credence is split across several plausible ethical theories. This paper translates such insights to the field of reinforcement learning, proposes two training methods that realize different points among competing desiderata, and trains agents in simple environments to act under moral uncertainty. The results illustrate (1) how such uncertainty can help curb extreme behavior from commitment to single theories and (2) several technical complications arising from attempting to ground moral philosophy in RL (e.g. how can a principled trade-off between two competing but incomparable reward functions be reached). The aim is to catalyze progress towards morally-competent agents and highlight the potential of RL to contribute towards the computational grounding of moral philosophy.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Capacity expansion under uncertainty in an oligopoly using indirect reinforcement-learning
    Oliveira, Fernando S.
    Costa, Manuel L. G.
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2018, 267 (03) : 1039 - 1050
  • [22] Hierarchical production control and distribution planning under retail uncertainty with reinforcement learning
    Deng, Yang
    Chow, Andy H. F.
    Yan, Yimo
    Su, Zicheng
    Zhou, Zhili
    Kuo, Yong-Hong
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2025,
  • [23] Uncertainty Quantification and Exploration for Reinforcement Learning
    Zhu, Yi
    Dong, Jing
    Lam, Henry
    OPERATIONS RESEARCH, 2024, 72 (04) : 1689 - 1709
  • [24] A model for system uncertainty in reinforcement learning
    Murray, Ryan
    Palladino, Michele
    SYSTEMS & CONTROL LETTERS, 2018, 122 : 24 - 31
  • [25] A model for system uncertainty in reinforcement learning
    Murray, Ryan
    Palladino, Michele
    Systems and Control Letters, 2018, 122 : 24 - 31
  • [26] A model-free, reinforcement learning algorithm for perceptual decision making under uncertainty
    Esmaily, Jamal
    Moran, Rani
    Roudi, Yasser
    Bahrami, Bahador
    JOURNAL OF COMPUTATIONAL NEUROSCIENCE, 2024, 52 : S21 - S21
  • [27] Large-scale dynamic surgical scheduling under uncertainty by hierarchical reinforcement learning
    Zhao, Lixiang
    Zhu, Han
    Zhang, Min
    Tang, Jiafu
    Wang, Yu
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2024,
  • [28] Controlling blood glucose variability under uncertainty using reinforcement learning and Gaussian processes
    De Paula, Mariano
    Avila, Luis Omar
    Martinez, Ernesto C.
    APPLIED SOFT COMPUTING, 2015, 35 : 310 - 332
  • [29] Belief state separated reinforcement learning for autonomous vehicle decision making under uncertainty
    Gu, Ziqing
    Yang, Yujie
    Duan, Jingliang
    Li, Shengbo Eben
    Chen, Jianyu
    Cao, Wenhan
    Zheng, Sifa
    2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2021, : 586 - 592
  • [30] Safe Multiagent Motion Planning Under Uncertainty for Drones Using Filtered Reinforcement Learning
    Safaoui, Sleiman
    Vinod, Abraham P.
    Chakrabarty, Ankush
    Quirynen, Rien
    Yoshikawa, Nobuyuki
    Di Cairano, Stefano
    IEEE TRANSACTIONS ON ROBOTICS, 2024, 40 (2529-2542) : 2529 - 2542