HILL: A Hallucination Identifier for Large Language Models

要旨

Large language models (LLMs) are prone to hallucinations, i.e., nonsensical, unfaithful, and undesirable text. Users tend to overrely on LLMs and corresponding hallucinations which can lead to misinterpretations and errors. To tackle the problem of overreliance, we propose HILL, the Hallucination Identifier for Large Language Models. First, we identified design features for HILL with a Wizard of Oz approach with nine participants. Subsequently, we implemented HILL based on the identified design features and evaluated HILL's interface design by surveying 17 participants. Further, we investigated HILL's functionality to identify hallucinations based on an existing question-answering dataset and five user interviews. We find that HILL can correctly identify and highlight hallucinations in LLM responses which enables users to handle LLM responses with more caution. With that, we propose an easy-to-implement adaptation to existing LLMs and demonstrate the relevance of user-centered designs of AI artifacts.

著者
Florian Leiser
Karlsruhe Institute of Technology, Karlsruhe, Germany
Sven Eckhardt
University of Zurich, Zurich, Switzerland
Valentin Leuthe
Karlsruhe Institute of Technology, Karlsruhe, Germany
Merlin Knaeble
Karlsruhe Institute of Technology (KIT), Karlsruhe, Germany
Alexander Mädche
Karlsruhe Institute of Technology (KIT), Karlsruhe, DEUTSCHLAND, Germany
Gerhard Schwabe
University of Zurich, Zurich, Switzerland
Ali Sunyaev
Karlsruhe Institute of Technology, Karlsruhe, Germany
論文URL

https://doi.org/10.1145/3613904.3642428

動画

会議: CHI 2024

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2024.acm.org/)

セッション: Remote Presentations: Highlight on Chatbots and LLMs

Remote Sessions
4 件の発表
2024-05-15 18:00:00
2024-05-16 02:20:00