Fighting Hate Speech, Silencing Drag Queens? Artificial Intelligence in Content Moderation and Risks to LGBTQ Voices Online
被引:53
|
作者:
Dias Oliva, Thiago
论文数: 0引用数: 0
h-index: 0
机构:
Univ Sao Paulo, Sao Paulo, Brazil
InternetLab, Sao Paulo, BrazilUniv Sao Paulo, Sao Paulo, Brazil
Dias Oliva, Thiago
[1
,2
]
Antonialli, Dennys Marcelo
论文数: 0引用数: 0
h-index: 0
机构:
Univ Sao Paulo, Law Sch, Sao Paulo, Brazil
Stanford Law Sch, Stanford, CA USA
Bucerius Law Sch, Hamburg, Germany
WHU Otto Von Beisheim Sch Management, Vallendar, GermanyUniv Sao Paulo, Sao Paulo, Brazil
Antonialli, Dennys Marcelo
[3
,4
,5
,6
]
Gomes, Alessandra
论文数: 0引用数: 0
h-index: 0
机构:
InternetLab, Sao Paulo, Brazil
State Univ Campinas UNICAMP, Campinas, Brazil
Fed Univ Para UFPA, Belem, Para, BrazilUniv Sao Paulo, Sao Paulo, Brazil
Gomes, Alessandra
[2
,7
,8
]
机构:
[1] Univ Sao Paulo, Sao Paulo, Brazil
[2] InternetLab, Sao Paulo, Brazil
[3] Univ Sao Paulo, Law Sch, Sao Paulo, Brazil
[4] Stanford Law Sch, Stanford, CA USA
[5] Bucerius Law Sch, Hamburg, Germany
[6] WHU Otto Von Beisheim Sch Management, Vallendar, Germany
Companies operating internet platforms are developing artificial intelligence tools for content moderation purposes. This paper discusses technologies developed to measure the 'toxicity' of text-based content. The research builds upon queer linguistic studies that have indicated the use of 'mock impoliteness' as a form of interaction employed by LGBTQ people to cope with hostility. Automated analyses that disregard such a pro-social function may, contrary to their intended design, actually reinforce harmful biases. This paper uses 'Perspective', an AI technology developed by Jigsaw (formerly Google Ideas), to measure the levels of toxicity of tweets from prominent drag queens in the United States. The research indicated that Perspective considered a significant number of drag queen Twitter accounts to have higher levels of toxicity than white nationalists. The qualitative analysis revealed that Perspective was not able to properly consider social context when measuring toxicity levels and failed to recognize cases in which words, that might conventionally be seen as offensive, conveyed different meanings in LGBTQ speech.