Large Language Models (LLMs) have made revolutionary progress in the field of natural language processing, capable of generating fluent, coherent, and seemingly reasonable text. However, these models have a fatal flaw: hallucination—generating information that appears real but is actually incorrect or fictional.
The hallucination problem poses serious risks in multiple scenarios:
- Medical consultation: AI may provide incorrect medical advice, endangering patients' health
- Legal consultation: Inaccurate legal interpretations may lead to serious consequences
- Financial analysis: Incorrect market information may cause investment losses
- News reporting: The spread of false information can mislead public opinion
Existing hallucination detection methods often rely on a single source of signals, such as only model-internal confidence or only external knowledge base retrieval. This single-perspective approach struggles to handle the diversity and complexity of hallucinations.