Fighting Hate Speech, Silencing Drag Queens? Artificial Intelligence in Content Moderation and Risks to LGBTQ Voices Online

被引:53
|
作者
Dias Oliva, Thiago [1 ,2 ]
Antonialli, Dennys Marcelo [3 ,4 ,5 ,6 ]
Gomes, Alessandra [2 ,7 ,8 ]
机构
[1] Univ Sao Paulo, Sao Paulo, Brazil
[2] InternetLab, Sao Paulo, Brazil
[3] Univ Sao Paulo, Law Sch, Sao Paulo, Brazil
[4] Stanford Law Sch, Stanford, CA USA
[5] Bucerius Law Sch, Hamburg, Germany
[6] WHU Otto Von Beisheim Sch Management, Vallendar, Germany
[7] State Univ Campinas UNICAMP, Campinas, Brazil
[8] Fed Univ Para UFPA, Belem, Para, Brazil
来源
关键词
Artificial intelligence; Content moderation; Toxicity; Hate speech; Queer linguistics; Drag queens;
D O I
10.1007/s12119-020-09790-w
中图分类号
C [社会科学总论];
学科分类号
03 ; 0303 ;
摘要
Companies operating internet platforms are developing artificial intelligence tools for content moderation purposes. This paper discusses technologies developed to measure the 'toxicity' of text-based content. The research builds upon queer linguistic studies that have indicated the use of 'mock impoliteness' as a form of interaction employed by LGBTQ people to cope with hostility. Automated analyses that disregard such a pro-social function may, contrary to their intended design, actually reinforce harmful biases. This paper uses 'Perspective', an AI technology developed by Jigsaw (formerly Google Ideas), to measure the levels of toxicity of tweets from prominent drag queens in the United States. The research indicated that Perspective considered a significant number of drag queen Twitter accounts to have higher levels of toxicity than white nationalists. The qualitative analysis revealed that Perspective was not able to properly consider social context when measuring toxicity levels and failed to recognize cases in which words, that might conventionally be seen as offensive, conveyed different meanings in LGBTQ speech.
引用
收藏
页码:700 / 732
页数:33
相关论文
共 2 条