Factuality challenges in the era of large language models and opportunities for fact-checking

被引:6
|
作者
Augenstein, Isabelle [1 ]
Baldwin, Timothy [2 ]
Cha, Meeyoung [3 ]
Chakraborty, Tanmoy [4 ]
Ciampaglia, Giovanni Luca [5 ]
Corney, David [6 ]
Diresta, Renee [7 ]
Ferrara, Emilio [8 ]
Hale, Scott [9 ]
Halevy, Alon [10 ]
Hovy, Eduard [11 ]
Ji, Heng [12 ]
Menczer, Filippo [13 ]
Miguez, Ruben [14 ]
Nakov, Preslav [2 ]
Scheufele, Dietram [15 ]
Sharma, Shivam [4 ]
Zagni, Giovanni [16 ]
机构
[1] Univ Copenhagen, Copenhagen, Denmark
[2] Mohamed bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
[3] Max Planck Inst Secur & Privacy, Univ Str, Bochum, Germany
[4] Indian Inst Technol Delhi, New Delhi, India
[5] Univ Maryland, College Pk, MD USA
[6] Full Fact, London, England
[7] Stanford Univ, Stanford, CA USA
[8] Univ Southern Calif, Los Angeles, CA USA
[9] Univ Oxford, Oxford, England
[10] Meta AI, Menlo Pk, CA USA
[11] Carnegie Mellon Univ, Pittsburgh, PA USA
[12] Univ Illinois, Champaign, IL USA
[13] Indiana Univ, Bloomington, IN USA
[14] Newtrales, Madrid, Spain
[15] Univ Wisconsin, Madison, WI USA
[16] GREMI, Milan, Italy
基金
新加坡国家研究基金会; 美国国家科学基金会; 欧洲研究理事会;
关键词
D O I
10.1038/s42256-024-00881-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of tools based on large language models (LLMs), such as OpenAI's ChatGPT and Google's Gemini, has garnered immense public attention owing to their advanced natural language generation capabilities. These remarkably natural-sounding tools have the potential to be highly useful for various tasks. However, they also tend to produce false, erroneous or misleading content-commonly referred to as hallucinations. Moreover, LLMs can be misused to generate convincing, yet false, content and profiles on a large scale, posing a substantial societal challenge by potentially deceiving users and spreading inaccurate information. This makes fact-checking increasingly important. Despite their issues with factual accuracy, LLMs have shown proficiency in various subtasks that support fact-checking, which is essential to ensure factually accurate responses. In light of these concerns, we explore issues related to factuality in LLMs and their impact on fact-checking. We identify key challenges, imminent threats and possible solutions to these factuality issues. We also thoroughly examine these challenges, existing solutions and potential prospects for fact-checking. By analysing the factuality constraints within LLMs and their impact on fact-checking, we aim to contribute to a path towards maintaining accuracy at a time of confluence of generative artificial intelligence and misinformation.
引用
收藏
页码:852 / 863
页数:12
相关论文
共 50 条
  • [21] Large language models in psychiatry: Opportunities and challenges
    Volkmer, Sebastian
    Meyer-Lindenberg, Andreas
    Schwarz, Emanuel
    PSYCHIATRY RESEARCH, 2024, 339
  • [22] Practitioner perceptions: Critical junctures and the global emergence and challenges of fact-checking
    Amazeen, Michelle A.
    INTERNATIONAL COMMUNICATION GAZETTE, 2019, 81 (6-8) : 541 - 561
  • [23] Automatic Fake News Detection: Are current models "fact-checking" or "gut-checking"?
    Kelk, Ian
    Basseri, Benjamin
    Lee, Wee Yi
    Qiu, Richard
    Tanner, Chris
    PROCEEDINGS OF THE FIFTH FACT EXTRACTION AND VERIFICATION WORKSHOP (FEVER 2022), 2022, : 29 - 36
  • [24] ChatGPT and large language models in academia: opportunities and challenges
    Jesse G. Meyer
    Ryan J. Urbanowicz
    Patrick C. N. Martin
    Karen O’Connor
    Ruowang Li
    Pei-Chen Peng
    Tiffani J. Bright
    Nicholas Tatonetti
    Kyoung Jae Won
    Graciela Gonzalez-Hernandez
    Jason H. Moore
    BioData Mining, 16
  • [25] ChatGPT and large language models in academia: opportunities and challenges
    Meyer, Jesse G.
    Urbanowicz, Ryan J.
    Martin, Patrick C. N.
    O'Connor, Karen
    Li, Ruowang
    Peng, Pei-Chen
    Bright, Tiffani J.
    Tatonetti, Nicholas
    Won, Kyoung Jae
    Gonzalez-Hernandez, Graciela
    Moore, Jason H.
    BIODATA MINING, 2023, 16 (01)
  • [26] Large Language Models: Opportunities and Challenges For Cognitive Assessment
    Efremova, Maria
    Kubiak, Emeric
    Baron, Simon
    Bernard, David
    EUROPEAN JOURNAL OF PSYCHOLOGY OPEN, 2023, 82 : 133 - 134
  • [27] Semantic similarity models for automated fact-checking: ClaimCheck as a claim matching tool
    Larraz, Irene
    Miguez, Ruben
    Sallicati, Francesca
    PROFESIONAL DE LA INFORMACION, 2023, 32 (03):
  • [28] FaxPlainAC: A Fact-Checking Tool Based on EXPLAINable Models with HumAn Correction in the Loop
    Zhang, Zijian
    Rudra, Koustav
    Anand, Avishek
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 4823 - 4827
  • [29] FELM: Benchmarking Factuality Evaluation of Large Language Models
    Chen, Shiqi
    Zhao, Yiran
    Zhang, Jinghan
    Chern, I-Chun
    Gao, Siyang
    Liu, Pengfei
    He, Junxian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [30] Transparency in AI usage within fact-checking platforms in Spain and its ethical challenges
    Cuartielles, Roger
    Mauri-Rios, Marcel
    Rodriguez-Martinez, Ruth
    COMMUNICATION & SOCIETY-SPAIN, 2024, 37 (04): : 257 - 271