In-Context Learning for Scalable and Online Hallucination Detection in RAGS

Article Status
Published
Author/contributor
Title
In-Context Learning for Scalable and Online Hallucination Detection in RAGS
Abstract
Ensuring fidelity to source documents is crucial for the responsible use of Large Language Models (LLMs) in Retrieval Augmented Generation (RAG) systems. We propose a lightweight method for real-time hallucination detection, with potential to be deployed as a model-agnostic microservice to bolster reliability. Using in-context learning, our approach evaluates response factuality at the sentence level without annotated data, promoting transparency and user trust. Compared to other prompt-based and semantic similarity baselines from recent literature, our method improves hallucination detection F1 scores by at least 11%, with consistent performance across different models. This research offers a practical solution for real-time validation of response accuracy in RAG systems, fostering responsible adoption, especially in critical domains where document fidelity is paramount.
Date
2024-04-20
Proceedings Title
Software Engineering
Conference Name
10th International Conference on Software Engineering
Publisher
Academy & Industry Research Collaboration Center
Pages
31-42
ISBN
978-1-923107-23-6
Accessed
10/06/2024, 18:07
Library Catalogue
DOI.org (Crossref)
Citation
Albanese, N. C. (2024). In-Context Learning for Scalable and Online Hallucination Detection in RAGS. Software Engineering, 31–42. https://doi.org/10.5121/csit.2024.140703
Technical methods
Powered by Zotero and Kerko.