Factuality challenges in the era of large language models and opportunities for fact-checking

被引:6
|
作者
Augenstein, Isabelle [1 ]
Baldwin, Timothy [2 ]
Cha, Meeyoung [3 ]
Chakraborty, Tanmoy [4 ]
Ciampaglia, Giovanni Luca [5 ]
Corney, David [6 ]
Diresta, Renee [7 ]
Ferrara, Emilio [8 ]
Hale, Scott [9 ]
Halevy, Alon [10 ]
Hovy, Eduard [11 ]
Ji, Heng [12 ]
Menczer, Filippo [13 ]
Miguez, Ruben [14 ]
Nakov, Preslav [2 ]
Scheufele, Dietram [15 ]
Sharma, Shivam [4 ]
Zagni, Giovanni [16 ]
机构
[1] Univ Copenhagen, Copenhagen, Denmark
[2] Mohamed bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
[3] Max Planck Inst Secur & Privacy, Univ Str, Bochum, Germany
[4] Indian Inst Technol Delhi, New Delhi, India
[5] Univ Maryland, College Pk, MD USA
[6] Full Fact, London, England
[7] Stanford Univ, Stanford, CA USA
[8] Univ Southern Calif, Los Angeles, CA USA
[9] Univ Oxford, Oxford, England
[10] Meta AI, Menlo Pk, CA USA
[11] Carnegie Mellon Univ, Pittsburgh, PA USA
[12] Univ Illinois, Champaign, IL USA
[13] Indiana Univ, Bloomington, IN USA
[14] Newtrales, Madrid, Spain
[15] Univ Wisconsin, Madison, WI USA
[16] GREMI, Milan, Italy
基金
新加坡国家研究基金会; 美国国家科学基金会; 欧洲研究理事会;
关键词
D O I
10.1038/s42256-024-00881-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of tools based on large language models (LLMs), such as OpenAI's ChatGPT and Google's Gemini, has garnered immense public attention owing to their advanced natural language generation capabilities. These remarkably natural-sounding tools have the potential to be highly useful for various tasks. However, they also tend to produce false, erroneous or misleading content-commonly referred to as hallucinations. Moreover, LLMs can be misused to generate convincing, yet false, content and profiles on a large scale, posing a substantial societal challenge by potentially deceiving users and spreading inaccurate information. This makes fact-checking increasingly important. Despite their issues with factual accuracy, LLMs have shown proficiency in various subtasks that support fact-checking, which is essential to ensure factually accurate responses. In light of these concerns, we explore issues related to factuality in LLMs and their impact on fact-checking. We identify key challenges, imminent threats and possible solutions to these factuality issues. We also thoroughly examine these challenges, existing solutions and potential prospects for fact-checking. By analysing the factuality constraints within LLMs and their impact on fact-checking, we aim to contribute to a path towards maintaining accuracy at a time of confluence of generative artificial intelligence and misinformation.
引用
收藏
页码:852 / 863
页数:12
相关论文
共 50 条
  • [41] ChatGPT for good? On opportunities and challenges of large language models for education
    Kasneci, Enkelejda
    Sessler, Kathrin
    Kuechemann, Stefan
    Bannert, Maria
    Dementieva, Daryna
    Fischer, Frank
    Gasser, Urs
    Groh, Georg
    Guennemann, Stephan
    Huellermeier, Eyke
    Krusche, Stepha
    Kutyniok, Gitta
    Michaeli, Tilman
    Nerdel, Claudia
    Pfeffer, Juergen
    Poquet, Oleksandra
    Sailer, Michael
    Schmidt, Albrecht
    Seidel, Tina
    Stadler, Matthias
    Weller, Jochen
    Kuhn, Jochen
    Kasneci, Gjergji
    LEARNING AND INDIVIDUAL DIFFERENCES, 2023, 103
  • [42] Fake news agenda in the era of COVID-19: Identifying trends through fact-checking content
    Ceron W.
    de-Lima-Santos M.-F.
    Quiles M.G.
    Online Social Networks and Media, 2021, 21
  • [43] Blockchain in the agrifood sector: From storytelling to traceability fact-checking up to new economic models
    Rampone, Francesco
    Lecca, Fabio
    Giolito, Paolo
    Romano, Massimo
    ECONOMIA AGRO-ALIMENTARE, 2023, 25 (02) : 97 - 114
  • [44] Foundation and large language models: fundamentals, challenges, opportunities, and social impacts
    Myers, Devon
    Mohawesh, Rami
    Chellaboina, Venkata Ishwarya
    Sathvik, Anantha Lakshmi
    Venkatesh, Praveen
    Ho, Yi-Hui
    Henshaw, Hanna
    Alhawawreh, Muna
    Berdik, David
    Jararweh, Yaser
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2024, 27 (01): : 1 - 26
  • [45] Large language models for biomedicine: foundations, opportunities, challenges, and best practices
    Sahoo, Satya S.
    Plasek, Joseph M.
    Xu, Hua
    Uzuner, Ozlem
    Cohen, Trevor
    Yetisgen, Meliha
    Liu, Hongfang
    Meystre, Stephane
    Wang, Yanshan
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 2114 - 2124
  • [46] When large language models meet personalization: perspectives of challenges and opportunities
    Chen, Jin
    Liu, Zheng
    Huang, Xu
    Wu, Chenwang
    Liu, Qi
    Jiang, Gangwei
    Pu, Yuanhao
    Lei, Yuxuan
    Chen, Xiaolong
    Wang, Xingmei
    Zheng, Kai
    Lian, Defu
    Chen, Enhong
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2024, 27 (04):
  • [47] Large Language Models in Medical Education: Opportunities, Challenges, and Future Directions
    Abd-alrazaq, Alaa
    AlSaad, Rawan
    Alhuwail, Dari
    Ahmed, Arfan
    Healy, Padraig Mark
    Latifi, Syed
    Aziz, Sarah
    Damseh, Rafat
    Alrazak, Sadam Alabed
    Sheikh, Javaid
    JMIR MEDICAL EDUCATION, 2023, 9
  • [48] The usefulness of large language models for patient information on melanoma: challenges and opportunities
    Sangers, Tobias E.
    van Doorn, Remco
    BRITISH JOURNAL OF DERMATOLOGY, 2024, 192 (02) : 185 - 185
  • [49] Large language models for whole-learner support: opportunities and challenges
    Mannekote, Amogh
    Davies, Adam
    Pinto, Juan D.
    Zhang, Shan
    Olds, Daniel
    Schroeder, Noah L.
    Lehman, Blair
    Zapata-Rivera, Diego
    Zhai, Chengxiang
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2024, 7
  • [50] Opportunities and challenges of human large language models in surgery: a bibliometric analysis
    Luo, Ziyue
    Wei, Jingwen
    Zhou, Ruihao
    INTERNATIONAL JOURNAL OF SURGERY, 2024, 110 (03) : 1817 - 1819