HILL: A Hallucination Identifier for Large Language Models

被引:2
|
作者
Leiser, Florian [1 ]
Eckhardt, Sven [2 ]
Leuthe, Valentin [1 ]
Knaeble, Merlin [3 ]
Maedche, Alexander [3 ]
Schwabe, Gerhard [2 ]
Sunyaev, Ali [1 ]
机构
[1] Karlsruhe Inst Technol, Inst Appl Informat & Formal Descript Methods, Karlsruhe, Germany
[2] Univ Zurich, Dept Informat, Zurich, Switzerland
[3] Karlsruhe Inst Technol, Human Ctr Syst Lab, Karlsruhe, Germany
关键词
ChatGPT; Large Language Models; Artificial Hallucinations; Wizard of Oz; Artifact Development; AUTOMATION; WIZARD; OZ;
D O I
10.1145/3613904.3642428
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) are prone to hallucinations, i.e., non-sensical, unfaithful, and undesirable text. Users tend to overrely on LLMs and corresponding hallucinations which can lead to misinterpretations and errors. To tackle the problem of overreliance, we propose HILL, the "Hallucination Identifier for Large Language Models". First, we identified design features for HILL with a Wizard of Oz approach with nine participants. Subsequently, we implemented HILL based on the identified design features and evaluated HILL's interface design by surveying 17 participants. Further, we investigated HILL's functionality to identify hallucinations based on an existing question-answering dataset and five user interviews. We find that HILL can correctly identify and highlight hallucinations in LLM responses which enables users to handle LLM responses with more caution. With that, we propose an easy-to-implement adaptation to existing LLMs and demonstrate the relevance of user-centered designs of AI artifacts.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Large Language Models in Cyberattacks
    S. V. Lebed
    D. E. Namiot
    E. V. Zubareva
    P. V. Khenkin
    A. A. Vorobeva
    D. A. Svichkar
    Doklady Mathematics, 2024, 110 (Suppl 2) : S510 - S520
  • [42] Large language models and psychiatry
    Orru, Graziella
    Melis, Giulia
    Sartori, Giuseppe
    INTERNATIONAL JOURNAL OF LAW AND PSYCHIATRY, 2025, 101
  • [43] Autoformalization with Large Language Models
    Wu, Yuhuai
    Jiang, Albert Q.
    Li, Wenda
    Rabe, Markus N.
    Staats, Charles
    Jamnik, Mateja
    Szegedy, Christian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [44] Imitation and Large Language Models
    Boisseau, Eloise
    MINDS AND MACHINES, 2024, 34 (04)
  • [45] The Smallness of Large Language Models
    Denning, Peter J.
    COMMUNICATIONS OF THE ACM, 2023, 66 (09) : 24 - 27
  • [46] Large language models in medicine
    Thirunavukarasu, Arun James
    Ting, Darren Shu Jeng
    Elangovan, Kabilan
    Gutierrez, Laura
    Tan, Ting Fang
    Ting, Daniel Shu Wei
    NATURE MEDICINE, 2023, 29 (08) : 1930 - 1940
  • [47] Editorial on large language models
    Deibert, Christopher M.
    TRANSLATIONAL ANDROLOGY AND UROLOGY, 2024, 13 (05) : 897 - 898
  • [48] Large language models in medicine
    Arun James Thirunavukarasu
    Darren Shu Jeng Ting
    Kabilan Elangovan
    Laura Gutierrez
    Ting Fang Tan
    Daniel Shu Wei Ting
    Nature Medicine, 2023, 29 : 1930 - 1940
  • [49] On Finetuning Large Language Models
    Wang, Yu
    POLITICAL ANALYSIS, 2023,
  • [50] Large Language Models and Psychoeducation
    Kleebayoon, Amnuay
    Wiwanitkit, Viroj
    JOURNAL OF ECT, 2024, 40 (01) : e1 - e1