Multi-perspective consistency checking for large language model hallucination detection: a black-box zero-resource approach
Linggang KONG , Xiaofeng ZHONG , Jie CHEN , Haoran FU , Yongjie WANG
Eng Inform Technol Electron Eng ›› 2025, Vol. 26 ›› Issue (11) : 2298 -2309.
Multi-perspective consistency checking for large language model hallucination detection: a black-box zero-resource approach
Large language models (LLMs) have been applied across various domains due to their superior natural language processing and generation capabilities. Nonetheless, LLMs occasionally generate content that contradicts real-world facts, known as hallucinations, posing significant challenges for real-world applications. To enhance the reliability of LLMs, it is imperative to detect hallucinations within LLM generations. Approaches that retrieve external knowledge or inspect the internal states of the model are frequently used to detect hallucinations; however, this requires either white-box access to the LLM or reliable expert knowledge resources, raising a high barrier for end-users. To address these challenges, we propose a black-box zero-resource approach for detecting LLM hallucinations, which primarily leverages multi-perspective consistency checking. The proposed approach mitigates the LLM overconfidence phenomenon by integrating multi-perspective consistency scores from both queries and responses. In comparison to the single-perspective detection approach, our proposed approach demonstrates superior performance in detecting hallucinations across multiple datasets and LLMs. Notably, in one experiment, where the hallucination rate reaches 94.7%, our approach improves the balanced accuracy (B-ACC) by 2.3 percentage points compared with the single consistency approach and achieves an area under the curve (AUC) of 0.832, all without depending on any external resources.
Large language models (LLMs) / LLM hallucination detection / Consistency checking / LLM security
Zhejiang University Press
/
| 〈 |
|
〉 |