73
2

Enhancing Hallucination Detection through Noise Injection

Abstract

Large Language Models (LLMs) are prone to generating plausible yet incorrect responses, known as hallucinations. Effectively detecting hallucinations is therefore crucial for the safe deployment of LLMs. Recent research has linked hallucinations to model uncertainty, suggesting that hallucinations can be detected by measuring dispersion over answer distributions obtained from a set of samples drawn from a model. While drawing from the distribution over tokens defined by the model is a natural way to obtain samples, in this work, we argue that it is sub-optimal for the purpose of detecting hallucinations. We show that detection can be improved significantly by taking into account model uncertainty in the Bayesian sense. To this end, we propose a very simple and efficient approach that perturbs an appropriate subset of model parameters, or equivalently hidden unit activations, during sampling. We demonstrate its effectiveness across a wide range of datasets and model architectures.

View on arXiv
@article{liu2025_2502.03799,
  title={ Enhancing Hallucination Detection through Noise Injection },
  author={ Litian Liu and Reza Pourreza and Sunny Panchal and Apratim Bhattacharyya and Yao Qin and Roland Memisevic },
  journal={arXiv preprint arXiv:2502.03799},
  year={ 2025 }
}
Comments on this paper