Reinforcement Learning Under Moral Uncertainty

被引:0
|
作者
Ecoffet, Adrien [1 ,2 ]
Lehman, Joel [1 ,2 ]
机构
[1] Uber AI Labs, San Francisco, CA 94103 USA
[2] OpenAI, San Francisco, CA 94110 USA
关键词
SOCIAL RATE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An ambitious goal for machine learning is to create agents that behave ethically: The capacity to abide by human moral norms would greatly expand the context in which autonomous agents could be practically and safely deployed, e.g. fully autonomous vehicles will encounter charged moral decisions that complicate their deployment. While ethical agents could be trained by rewarding correct behavior under a specific moral theory (e.g. utilitarianism), there remains widespread disagreement about the nature of morality. Acknowledging such disagreement, recent work in moral philosophy proposes that ethical behavior requires acting under moral uncertainty, i.e. to take into account when acting that one's credence is split across several plausible ethical theories. This paper translates such insights to the field of reinforcement learning, proposes two training methods that realize different points among competing desiderata, and trains agents in simple environments to act under moral uncertainty. The results illustrate (1) how such uncertainty can help curb extreme behavior from commitment to single theories and (2) several technical complications arising from attempting to ground moral philosophy in RL (e.g. how can a principled trade-off between two competing but incomparable reward functions be reached). The aim is to catalyze progress towards morally-competent agents and highlight the potential of RL to contribute towards the computational grounding of moral philosophy.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Reinforcement learning for humanitarian relief distribution with trucks and UAVs under travel time uncertainty
    Van Steenbergen, Robert
    Mes, Martijn
    Van Heeswijk, Wouter
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2023, 157
  • [32] A model-free, reinforcement learning algorithm for perceptual decision making under uncertainty
    Esmaily, Jamal
    Moran, Rani
    Roudi, Yasser
    Bahrami, Bahador
    JOURNAL OF COMPUTATIONAL NEUROSCIENCE, 2024, 52 : S21 - S21
  • [33] Deep Reinforcement Learning Approach for Capacitated Supply Chain Optimization under Demand Uncertainty
    Peng, Zedong
    Zhang, Yi
    Feng, Yiping
    Zhang, Tuchao
    Wu, Zhengguang
    Su, Hongye
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 3512 - 3517
  • [34] Towards Online Testing Under Uncertainty Using Model-Based Reinforcement Learning
    Camilli, Matteo
    Mirandola, Raffaela
    Scandurra, Patrizia
    Trubiani, Catia
    SOFTWARE ARCHITECTURE. ECSA 2022 TRACKS AND WORKSHOPS, 2023, 13928 : 233 - 245
  • [35] Deep Reinforcement Learning Based Autonomous Exploration under Uncertainty with Hybrid Network on Graph
    Zhang, Zhiwen
    Shi, Chenghao
    Zeng, Zhiwen
    Zhang, Hui
    2021 IEEE 9TH INTERNATIONAL CONFERENCE ON INFORMATION, COMMUNICATION AND NETWORKS (ICICN 2021), 2021, : 450 - 456
  • [36] Uncertainty quantification for operators in online reinforcement learning
    Wang, Bi
    Wu, Jianqing
    Li, Xuelian
    Shen, Jun
    Zhong, Yangjun
    KNOWLEDGE-BASED SYSTEMS, 2022, 258
  • [37] Online Robust Reinforcement Learning with Model Uncertainty
    Wang, Yue
    Zou, Shaofeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [38] Safe Reinforcement Learning with Model Uncertainty Estimates
    Lutjens, Bjorn
    Everett, Michael
    How, Jonathan P.
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 8662 - 8668
  • [39] Uncertainty Aware Model Integration on Reinforcement Learning
    Nagata, Takashi
    Xing, Jinwei
    Kumazawa, Tsutomu
    Neftci, Emre
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [40] Discrete Uncertainty Quantification For Offline Reinforcement Learning
    Perez, Jose Luis
    Corrochano, Javier
    Garcia, Javier
    Majadas, Ruben
    Ibanez-Llano, Cristina
    Perez, Sergio
    Fernandez, Fernando
    JOURNAL OF ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING RESEARCH, 2023, 13 (04) : 273 - 287