logo
ResearchBunny Logo
Abstract
Large language models (LLMs) often produce false or unsubstantiated outputs, hindering their adoption in various fields. This paper introduces entropy-based uncertainty estimators to detect a subset of these hallucinations—confabulations—which are arbitrary and incorrect generations. The method focuses on semantic meaning rather than word sequences, working across datasets and tasks without prior knowledge and generalizing well to new tasks. By identifying prompts likely to produce confabulations, it helps users assess LLM reliability and expands the potential uses of these otherwise unreliable models.
Publisher
Nature
Published On
Jun 20, 2024
Authors
Sebastian Farquhar, Jannik Kossen, Lorenz Kuhn, Yarin Gal
Tags
large language models
confabulations
entropy-based uncertainty
reliability
semantic meaning
hallucinations
user assessment
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs, just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny