logo
ResearchBunny Logo
Detecting hallucinations in large language models using semantic entropy

Computer Science

Detecting hallucinations in large language models using semantic entropy

S. Farquhar, J. Kossen, et al.

Discover how researchers Sebastian Farquhar, Jannik Kossen, Lorenz Kuhn, and Yarin Gal are tackling the reliability of large language models with an innovative entropy-based method. This approach enables users to identify confabulations—incorrect outputs—without needing prior task knowledge, paving the way for safer applications in various fields.

00:00
00:00
Playback language: English
Abstract
Large language models (LLMs) often produce false or unsubstantiated outputs, hindering their adoption in various fields. This paper introduces entropy-based uncertainty estimators to detect a subset of these hallucinations—confabulations—which are arbitrary and incorrect generations. The method focuses on semantic meaning rather than word sequences, working across datasets and tasks without prior knowledge and generalizing well to new tasks. By identifying prompts likely to produce confabulations, it helps users assess LLM reliability and expands the potential uses of these otherwise unreliable models.
Publisher
Nature
Published On
Jun 20, 2024
Authors
Sebastian Farquhar, Jannik Kossen, Lorenz Kuhn, Yarin Gal
Tags
large language models
confabulations
entropy-based uncertainty
reliability
semantic meaning
hallucinations
user assessment
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs, just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny