HALoGEN: Fantastic LLM Hallucinations and Where to Find Them
Abstract
Despite their impressive ability to generate high-quality and fluent text, generative large language models (LLMs) also produce hallucinations: statements that are misaligned with established world knowledge or provided input context. However, measuring hallucination can be challenging, as having humans verify model generations on-the-fly is both expensive and time-consuming. In this work, we release HALoGEN, a comprehensive hallucination benchmark consisting of: (1) 10,923 prompts for generative models spanning nine domains including programming, scientific attribution, and summarization, and (2) automatic high-precision verifiers for each use case that decompose LLM generations into atomic units, and verify each unit against a high-quality knowledge source. We use this framework to evaluate ~150,000 generations from 14 language models, finding that even the best-performing models are riddled with hallucinations (sometimes up to 86% of generated atomic facts depending on the domain). We further define a novel error classification for LLM hallucinations based on whether they likely stem from incorrect recollection of training data (Type A errors), or incorrect knowledge in training data (Type B errors), or are fabrication (Type C errors). We hope our framework provides a foundation to enable the principled study of why generative models hallucinate, and advances the development of trustworthy large language models.
Community
LLMs hallucinate differently in each task, so we need different fixes for different problems
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- DHCP: Detecting Hallucinations by Cross-modal Attention Pattern in Large Vision-Language Models (2024)
- VidHalluc: Evaluating Temporal Hallucinations in Multimodal Large Language Models for Video Understanding (2024)
- VidHal: Benchmarking Temporal Hallucinations in Vision LLMs (2024)
- Fine-tuning Large Language Models for Improving Factuality in Legal Question Answering (2025)
- HalluCana: Fixing LLM Hallucination with A Canary Lookahead (2024)
- MedHallBench: A New Benchmark for Assessing Hallucination in Medical Large Language Models (2024)
- Verb Mirage: Unveiling and Assessing Verb Concept Hallucinations in Multimodal Large Language Models (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper