Section 01
[Introduction] TokenHD: A New Fine-Grained Detection Method for Hallucinations in Large Language Models
TokenHD proposes a new token-level method for detecting hallucinations in large language models. By using a scalable data synthesis engine and an importance-weighted training strategy, it addresses the limitations of existing step-level detection methods, such as restricted granularity and poor scalability. Experiments show that even a small model with only 0.6B parameters can outperform the reasoning capabilities of large models with 32B parameters, achieving excellent performance in hallucination detection tasks and providing more precise solutions for scenarios like AI safety and content moderation.