ExplorerData ScienceMachine Learning
Research PaperResearchia:202605.07020

Low-Cost Black-Box Detection of LLM Hallucinations via Dynamical System Prediction

Dan Wilson

Abstract

Large Language Models (LLMs) frequently generate plausible but non-factual content, a phenomenon known as hallucination. While existing detection methods typically rely on computationally expensive sampling-based consistency checks or external knowledge retrieval, we propose a new method that treats the LLM as a black-box dynamical system. By projecting LLM responses into a high-dimensional manifold via an embedding model, we characterize the resulting vector sequences as observable realizations...

Submitted: May 7, 2026Subjects: Machine Learning; Data Science

Description / Details

Large Language Models (LLMs) frequently generate plausible but non-factual content, a phenomenon known as hallucination. While existing detection methods typically rely on computationally expensive sampling-based consistency checks or external knowledge retrieval, we propose a new method that treats the LLM as a black-box dynamical system. By projecting LLM responses into a high-dimensional manifold via an embedding model, we characterize the resulting vector sequences as observable realizations of the model's latent state-space dynamics. Leveraging Koopman operator theory, we fit the transition operators for both factual and hallucinated regimes and define a differential residual score based on their respective prediction errors. To accommodate varying user requirements and domain-specific sensitivities, we introduce a preference-aware calibration mechanism that optimizes the classification threshold based on a small set of demonstrations. This approach enables low-cost hallucination detection in a single-sample pass, avoiding the need for secondary sampling or external grounding. Extensive testing across three data benchmarks demonstrates that our method achieves state-of-the-art performance with reduced resource overhead.


Source: arXiv:2605.05134v1 - http://arxiv.org/abs/2605.05134v1 PDF: https://arxiv.org/pdf/2605.05134v1 Original Link: http://arxiv.org/abs/2605.05134v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
May 7, 2026
Topic:
Data Science
Area:
Machine Learning
Comments:
0
Bookmark
Low-Cost Black-Box Detection of LLM Hallucinations via Dynamical System Prediction | Researchia