What governs attitudes toward artificial intelligence adoption and governance?

被引:25
|
作者
O'Shaughnessy, Matthew R. [1 ]
Schiff, Daniel S. [2 ]
Varshney, Lav R. [3 ]
Rozell, Christopher J. [1 ]
Davenport, Mark A. [1 ]
机构
[1] Georgia Inst Technol, Sch Elect & Comp Engn, 777 Atlantic Dr NW, Atlanta, GA 30332 USA
[2] Purdue Univ, Dept Polit Sci, 110 North Univ St, W Lafayette, IN 47907 USA
[3] Univ Illinois, Dept Elect & Comp Engn, 306 N Wright St MC 702, Urbana, IL 61801 USA
基金
美国国家科学基金会;
关键词
artificial intelligence policy; public opinion; public engagement; RISK PERCEPTION; TECHNOLOGY ACCEPTANCE; CULTURAL COGNITION; PERCEIVED RISK; GENDER; WORLDVIEWS; POLITICS; SCIENCE; MODEL; RACE;
D O I
10.1093/scipol/scac056
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Designing effective and inclusive governance and public communication strategies for artificial intelligence (AI) requires understanding how stakeholders reason about its use and governance. We examine underlying factors and mechanisms that drive attitudes toward the use and governance of AI across six policy-relevant applications using structural equation modeling and surveys of both US adults (N = 3,524) and technology workers enrolled in an online computer science master's degree program (N = 425). We find that the cultural values of individualism, egalitarianism, general risk aversion, and techno-skepticism are important drivers of AI attitudes. Perceived benefit drives attitudes toward AI use but not its governance. Experts hold more nuanced views than the public and are more supportive of AI use but not its regulation. Drawing on these findings, we discuss challenges and opportunities for participatory AI governance, and we recommend that trustworthy AI governance be emphasized as strongly as trustworthy AI.
引用
收藏
页码:161 / 176
页数:16
相关论文
共 50 条