Factuality challenges in the era of large language models and opportunities for fact-checking

被引:6
|
作者
Augenstein, Isabelle [1 ]
Baldwin, Timothy [2 ]
Cha, Meeyoung [3 ]
Chakraborty, Tanmoy [4 ]
Ciampaglia, Giovanni Luca [5 ]
Corney, David [6 ]
Diresta, Renee [7 ]
Ferrara, Emilio [8 ]
Hale, Scott [9 ]
Halevy, Alon [10 ]
Hovy, Eduard [11 ]
Ji, Heng [12 ]
Menczer, Filippo [13 ]
Miguez, Ruben [14 ]
Nakov, Preslav [2 ]
Scheufele, Dietram [15 ]
Sharma, Shivam [4 ]
Zagni, Giovanni [16 ]
机构
[1] Univ Copenhagen, Copenhagen, Denmark
[2] Mohamed bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
[3] Max Planck Inst Secur & Privacy, Univ Str, Bochum, Germany
[4] Indian Inst Technol Delhi, New Delhi, India
[5] Univ Maryland, College Pk, MD USA
[6] Full Fact, London, England
[7] Stanford Univ, Stanford, CA USA
[8] Univ Southern Calif, Los Angeles, CA USA
[9] Univ Oxford, Oxford, England
[10] Meta AI, Menlo Pk, CA USA
[11] Carnegie Mellon Univ, Pittsburgh, PA USA
[12] Univ Illinois, Champaign, IL USA
[13] Indiana Univ, Bloomington, IN USA
[14] Newtrales, Madrid, Spain
[15] Univ Wisconsin, Madison, WI USA
[16] GREMI, Milan, Italy
基金
新加坡国家研究基金会; 美国国家科学基金会; 欧洲研究理事会;
关键词
D O I
10.1038/s42256-024-00881-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of tools based on large language models (LLMs), such as OpenAI's ChatGPT and Google's Gemini, has garnered immense public attention owing to their advanced natural language generation capabilities. These remarkably natural-sounding tools have the potential to be highly useful for various tasks. However, they also tend to produce false, erroneous or misleading content-commonly referred to as hallucinations. Moreover, LLMs can be misused to generate convincing, yet false, content and profiles on a large scale, posing a substantial societal challenge by potentially deceiving users and spreading inaccurate information. This makes fact-checking increasingly important. Despite their issues with factual accuracy, LLMs have shown proficiency in various subtasks that support fact-checking, which is essential to ensure factually accurate responses. In light of these concerns, we explore issues related to factuality in LLMs and their impact on fact-checking. We identify key challenges, imminent threats and possible solutions to these factuality issues. We also thoroughly examine these challenges, existing solutions and potential prospects for fact-checking. By analysing the factuality constraints within LLMs and their impact on fact-checking, we aim to contribute to a path towards maintaining accuracy at a time of confluence of generative artificial intelligence and misinformation.
引用
收藏
页码:852 / 863
页数:12
相关论文
共 50 条
  • [31] Audiovisual fact-checking in the post-truth era: What does it mean to validate an image?
    Rodriguez Serrano, Aaron
    Soler-Campillo, Maria
    Marzal-Felici, Javier
    REVISTA LATINA DE COMUNICACION SOCIAL, 2021, 79
  • [32] Providing public access to archival materials and the problem of fact-checking in the post-truth era
    Wang, Hui
    SOCIAL SCIENCE QUARTERLY, 2022, 103 (07) : 1750 - 1764
  • [33] End-to-End Multimodal Fact-Checking and Explanation Generation: A Challenging Dataset and Models
    Yao, Barry Menglong
    Shah, Aditya
    Sun, Lichao
    Cho, Jin-Hee
    Huang, Lifu
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2733 - 2743
  • [34] Embracing Large Language Models for Medical Applications: Opportunities and Challenges
    Karabacak, Mert
    Margetis, Konstantinos
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2023, 15 (05)
  • [35] Large language models for building energy applications: Opportunities and challenges
    Liu, Mingzhe
    Zhang, Liang
    Chen, Jianli
    Chen, Wei-An
    Yang, Zhiyao
    Lo, L. James
    Wen, Jin
    O'Neill, Zheng
    BUILDING SIMULATION, 2025, 18 (02) : 225 - 234
  • [36] Large Language Models for Business Process Management: Opportunities and Challenges
    Vidgof, Maxim
    Bachhofner, Stefan
    Mendling, Jan
    BUSINESS PROCESS MANAGEMENT FORUM, BPM 2023 FORUM, 2023, 490 : 107 - 123
  • [37] Challenges and Opportunities of Moderating Usage of Large Language Models in Education
    Krupp, Lars
    Steinert, Steffen
    Kiefer-Emmanouilidis, Maximilian
    Avila, Karina E.
    Lukowicz, Paul
    Kuhn, Jochen
    Kuechemann, Stefan
    Karolus, Jakob
    AI FOR EDUCATION WORKSHOP, 2024, 257 : 9 - 17
  • [38] Latent Retrieval for Large-Scale Fact-Checking and Question Answering with NLI training
    Samarinas, Chris
    Hsu, Wynne
    Lee, Mong Li
    2020 IEEE 32ND INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2020, : 941 - 948
  • [39] Large Language Models and Future of Information Retrieval: Opportunities and Challenges
    Zhai, ChengXiang
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 481 - 490
  • [40] Opportunities and challenges for ChatGPT and large language models in biomedicine and health
    Tian, Shubo
    Jin, Qiao
    Yeganova, Lana
    Lai, Po-Ting
    Zhu, Qingqing
    Chen, Xiuying
    Yang, Yifan
    Chen, Qingyu
    Kim, Won
    Comeau, Donald C.
    Islamaj, Rezarta
    Kapoor, Aadit
    Gao, Xin
    Lu, Zhiyong
    BRIEFINGS IN BIOINFORMATICS, 2024, 25 (01)