Section 01
Introduction: Evidence Misalignment Project—An Open-Source Framework for Quantifying LLM Fact Anchoring Capability
This open-source project addresses the "hallucination" issue of large language models (LLMs), studying the impact of parameter scale (from 8B to 405B) on fact anchoring capability. It proposes the Evidence Alignment Score (EAS) as a hybrid evaluation metric, supports multi-backend model evaluation (local Ollama, cloud-based NVIDIA NIM/OpenAI), uses the FEVER benchmark dataset and rigorous processes, and provides a systematic, reproducible framework for LLM factuality evaluation.