Factuality challenges in the era of large language models and opportunities for fact-checking

被引:6
|
作者
Augenstein, Isabelle [1 ]
Baldwin, Timothy [2 ]
Cha, Meeyoung [3 ]
Chakraborty, Tanmoy [4 ]
Ciampaglia, Giovanni Luca [5 ]
Corney, David [6 ]
Diresta, Renee [7 ]
Ferrara, Emilio [8 ]
Hale, Scott [9 ]
Halevy, Alon [10 ]
Hovy, Eduard [11 ]
Ji, Heng [12 ]
Menczer, Filippo [13 ]
Miguez, Ruben [14 ]
Nakov, Preslav [2 ]
Scheufele, Dietram [15 ]
Sharma, Shivam [4 ]
Zagni, Giovanni [16 ]
机构
[1] Univ Copenhagen, Copenhagen, Denmark
[2] Mohamed bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
[3] Max Planck Inst Secur & Privacy, Univ Str, Bochum, Germany
[4] Indian Inst Technol Delhi, New Delhi, India
[5] Univ Maryland, College Pk, MD USA
[6] Full Fact, London, England
[7] Stanford Univ, Stanford, CA USA
[8] Univ Southern Calif, Los Angeles, CA USA
[9] Univ Oxford, Oxford, England
[10] Meta AI, Menlo Pk, CA USA
[11] Carnegie Mellon Univ, Pittsburgh, PA USA
[12] Univ Illinois, Champaign, IL USA
[13] Indiana Univ, Bloomington, IN USA
[14] Newtrales, Madrid, Spain
[15] Univ Wisconsin, Madison, WI USA
[16] GREMI, Milan, Italy
基金
新加坡国家研究基金会; 美国国家科学基金会; 欧洲研究理事会;
关键词
D O I
10.1038/s42256-024-00881-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of tools based on large language models (LLMs), such as OpenAI's ChatGPT and Google's Gemini, has garnered immense public attention owing to their advanced natural language generation capabilities. These remarkably natural-sounding tools have the potential to be highly useful for various tasks. However, they also tend to produce false, erroneous or misleading content-commonly referred to as hallucinations. Moreover, LLMs can be misused to generate convincing, yet false, content and profiles on a large scale, posing a substantial societal challenge by potentially deceiving users and spreading inaccurate information. This makes fact-checking increasingly important. Despite their issues with factual accuracy, LLMs have shown proficiency in various subtasks that support fact-checking, which is essential to ensure factually accurate responses. In light of these concerns, we explore issues related to factuality in LLMs and their impact on fact-checking. We identify key challenges, imminent threats and possible solutions to these factuality issues. We also thoroughly examine these challenges, existing solutions and potential prospects for fact-checking. By analysing the factuality constraints within LLMs and their impact on fact-checking, we aim to contribute to a path towards maintaining accuracy at a time of confluence of generative artificial intelligence and misinformation.
引用
收藏
页码:852 / 863
页数:12
相关论文
共 50 条
  • [1] The perils and promises of fact-checking with large language models
    Quelle, Dorian
    Bovet, Alexandre
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2024, 7
  • [2] Automated fact-checking of climate claims with large language models
    Leippold, Markus
    Vaghefi, Saeid Ashraf
    Stammbach, Dominik
    Muccione, Veruska
    Bingler, Julia
    Ni, Jingwei
    Senni, Chiara Colesanti
    Wekhof, Tobias
    Schimanski, Tobias
    Gostlow, Glen
    Yu, Tingyu
    Luterbacher, Juerg
    Huggel, Christian
    NPJ CLIMATE ACTION, 2025, 4 (01):
  • [3] Leveraging Large Language Models for Fact-Checking Farsi News Headlines
    Dehghani, Shirin
    Zahedivafa, Mohammadmasiha
    Baghshahi, Zahra
    Zare, Darya
    Yari, Sara
    Samei, Zeynab
    Aliahmadi, Mohammadhadi
    Abbasi, Mahdis
    Mirzamojtahedi, Sara
    Ebrahimi, Sarvenaz
    Alizadeh, Meysam
    DISINFORMATION IN OPEN ONLINE MEDIA, MISDOOM 2024, 2024, 15175 : 16 - 31
  • [4] Fact-checking information from large language models can decrease headline discernment
    Deverna, Matthew R.
    Yan, Harry Yaojun
    Yang, Kai-Cheng
    Menczer, Filippo
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2024, 121 (50)
  • [5] Scientific evidence and specific context: leveraging large language models for health fact-checking
    Ni, Zhenni
    Qian, Yuxing
    Chen, Shuaipu
    Jaulent, Marie-Christine
    Bousquet, Cedric
    ONLINE INFORMATION REVIEW, 2024, 48 (07) : 1488 - 1514
  • [6] The Social Opportunities and Challenges in the Era of Large Language Models
    Huimin C.
    Zhiyuan L.
    Maosong S.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2024, 61 (05): : 1094 - 1103
  • [7] The Impact and Opportunities of Generative AI in Fact-Checking
    Wolfe, Robert
    Mitra, Tanushree
    PROCEEDINGS OF THE 2024 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, ACM FACCT 2024, 2024, : 1531 - 1543
  • [8] Fact-Checking the Output of Large Language Models via Token-Level Uncertainty Quantification
    Fadeeva, Ekaterina
    Rubashevskii, Aleksandr
    Shelmanov, Artem
    Petrakov, Sergey
    Li, Haonan
    Mubarak, Hamdy
    Tsymbalov, Evgenii
    Kuzmin, Gleb
    Panchenko, Alexander
    Baldwin, Timothy
    Nakov, Preslav
    Panov, Maxim
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 9367 - 9385
  • [9] A Mystery for You: A fact-checking game enhanced by large language models (LLMs) and a tangible interface
    Tang, Haoheng
    Singha, Mrinalini
    EXTENDED ABSTRACTS OF THE 2024 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2024, 2024,
  • [10] A reflection on the epistemology of fact-checking journalism: challenges and dilemmas
    Rodriguez Perez, C.
    REVISTA DE COMUNICACION-PERU, 2020, 19 (01): : 243 - 258