Artificial intelligence;
Moral AI;
Moral cognition;
Machine ethics;
Moral psychology;
Reinforcement learning;
Fairness;
D O I:
暂无
中图分类号:
学科分类号:
摘要:
I describe a suite of reinforcement learning environments in which artificial agents learn to value and respond to moral content and contexts. I illustrate the core principles of the framework by characterizing one such environment, or “gridworld,” in which an agent learns to trade-off between monetary profit and fair dealing, as applied in a standard behavioral economic paradigm. I then highlight the core technical and philosophical advantages of the learning approach for modeling moral cognition, and for addressing the so-called value alignment problem in AI.
机构:
Faculty of Philosophy, University of Bucharest, 204 Splaiul Independentei St., BucharestFaculty of Philosophy, University of Bucharest, 204 Splaiul Independentei St., Bucharest
Constantinescu M.
Vică C.
论文数: 0引用数: 0
h-index: 0
机构:
Faculty of Philosophy, University of Bucharest, 204 Splaiul Independentei St., BucharestFaculty of Philosophy, University of Bucharest, 204 Splaiul Independentei St., Bucharest
Vică C.
Uszkai R.
论文数: 0引用数: 0
h-index: 0
机构:
Department of Philosophy and Social Sciences, Bucharest University of Economic Studies, BucharestFaculty of Philosophy, University of Bucharest, 204 Splaiul Independentei St., Bucharest