Moral Philosophy of Artificial General Intelligence: Agency and Responsibility

被引:3
|
作者
Maruyama, Yoshihiro [1 ]
机构
[1] Australian Natl Univ, Sch Comp, Canberra, ACT, Australia
来源
关键词
Ethical intelligence; Artificial moral agency; Strong and weak moral AI; Artificial moral responsibility; Responsible AI; SCIENCE;
D O I
10.1007/978-3-030-93758-4_15
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The European Parliament recently proposed to grant the personhood of autonomous AI, which raises fundamental questions concerning the ethical nature of AI. Can they be moral agents? Can they be morally responsible for actions and their consequences? Here we address these questions, focusing upon, inter alia, the possibilities of moral agency and moral responsibility in artificial general intelligence; moral agency is a precondition for moral responsibility (which is, in turn, a precondition for legal punishment). In the first part of the paper we address the moral agency status of AI in light of traditional moral philosophy, especially Kant's, Hume's, and Strawson's, and clarify the possibility of Moral AI (i.e., AI with moral agency) by discussing the Ethical Turing Test, the Moral Chinese Room Argument, and Weak and Strong Moral AI. In the second part we address the moral responsibility status of AI, and thereby clarify the possibility of Responsible AI (i.e., AI with moral responsibility). These issues would be crucial for AI-pervasive technosociety in the (possibly near) future, especially for post-human society after the development of artificial general intelligence.
引用
收藏
页码:139 / 150
页数:12
相关论文
共 50 条