Hallucination Detection in LLMs: Fast and Memory-Efficient Fine-Tuned Models

Published: 06 Nov 2024, Last Modified: 06 Jan 2025NLDL 2025 OralEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Large Language Models, Uncertainty Estimation, Hallucination Detection, Memory-Efficient Deep Ensembles
TL;DR: Hallucination Detection using ensembles of LLMs
Abstract: Uncertainty estimation is a necessary component when implementing AI in high-risk settings, such as autonomous cars, medicine, or insurances. Large Language Models (LLMs) have seen a surge in popularity in recent years, but they are subject to hallucinations, which may cause serious harm in high-risk settings. Despite their success, LLMs are expensive to train and run: they need a large amount of computations and memory, preventing the use of ensembling methods in practice. In this work, we present a novel method that allows for fast and memory-friendly training of LLM ensembles. We show that the resulting ensembles can detect hallucinations and are a viable approach in practice as only one GPU is needed for training and inference.
Git: https://github.com/Gabriel-Arteaga/LLM-Ensemble
Submission Number: 16
Loading