HILL: A Hallucination Identifier for Large Language Models

被引:2
|
作者
Leiser, Florian [1 ]
Eckhardt, Sven [2 ]
Leuthe, Valentin [1 ]
Knaeble, Merlin [3 ]
Maedche, Alexander [3 ]
Schwabe, Gerhard [2 ]
Sunyaev, Ali [1 ]
机构
[1] Karlsruhe Inst Technol, Inst Appl Informat & Formal Descript Methods, Karlsruhe, Germany
[2] Univ Zurich, Dept Informat, Zurich, Switzerland
[3] Karlsruhe Inst Technol, Human Ctr Syst Lab, Karlsruhe, Germany
关键词
ChatGPT; Large Language Models; Artificial Hallucinations; Wizard of Oz; Artifact Development; AUTOMATION; WIZARD; OZ;
D O I
10.1145/3613904.3642428
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) are prone to hallucinations, i.e., non-sensical, unfaithful, and undesirable text. Users tend to overrely on LLMs and corresponding hallucinations which can lead to misinterpretations and errors. To tackle the problem of overreliance, we propose HILL, the "Hallucination Identifier for Large Language Models". First, we identified design features for HILL with a Wizard of Oz approach with nine participants. Subsequently, we implemented HILL based on the identified design features and evaluated HILL's interface design by surveying 17 participants. Further, we investigated HILL's functionality to identify hallucinations based on an existing question-answering dataset and five user interviews. We find that HILL can correctly identify and highlight hallucinations in LLM responses which enables users to handle LLM responses with more caution. With that, we propose an easy-to-implement adaptation to existing LLMs and demonstrate the relevance of user-centered designs of AI artifacts.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Beyond Probabilities: Unveiling the Delicate Dance of Large Language Models (LLMs) and AI-Hallucination
    Hamid, Oussama H.
    2024 IEEE CONFERENCE ON COGNITIVE AND COMPUTATIONAL ASPECTS OF SITUATION MANAGEMENT, COGSIMA, 2024, : 85 - 90
  • [22] Unsupervised Real-Time Hallucination Detection based on the Internal States of Large Language Models
    Su, Weihang
    Wang, Changyue
    Ai, Qingyao
    Hu, Yiran
    Wu, Zhijing
    Zhou, Yujia
    Liu, Yiqun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 14379 - 14391
  • [23] SELFCHECKGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models
    Manakul, Potsawee
    Liusie, Adian
    Gales, Mark J. F.
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9004 - 9017
  • [24] Dual Process Theory for Large Language Models: An overview of using Psychology to address hallucination and reliability issues
    Bellini-Leite, Samuel C.
    ADAPTIVE BEHAVIOR, 2024, 32 (04) : 329 - 343
  • [25] A novel system for strengthening security in large language models against hallucination and injection attacks with effective strategies
    Gokcimen, Tunahan
    Das, Bihter
    ALEXANDRIA ENGINEERING JOURNAL, 2025, 123 : 71 - 90
  • [26] Accuracy, Consistency, and Hallucination of Large Language Models When Analyzing Unstructured Clinical Notes in Electronic Medical Records
    Shah, Savyasachi V.
    JAMA NETWORK OPEN, 2024, 7 (08)
  • [27] THRONE: An Object-based Hallucination Benchmark for the Free-form Generations of Large Vision-Language Models
    Kaul, Prannay
    Li, Zhizhong
    Yang, Hao
    Dukler, Yonatan
    Swaminathan, Ashwin
    Taylor, C. J.
    Soatto, Stefano
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 27218 - +
  • [28] Large Language Models are Not Models of Natural Language: They are Corpus Models
    Veres, Csaba
    IEEE ACCESS, 2022, 10 : 61970 - 61979
  • [29] Large Language Models
    Vargas, Diego Collarana
    Katsamanis, Nassos
    ERCIM NEWS, 2024, (136): : 12 - 13
  • [30] Large Language Models
    Cerf, Vinton G.
    COMMUNICATIONS OF THE ACM, 2023, 66 (08) : 7 - 7