Section 01
FACET Benchmark: Core Guide to Evaluating Attribution Faithfulness in LLM Multi-Factor Reasoning
FACET (Faithfulness Attribution in Complex Evaluation Tasks) is a four-probe benchmark framework designed for multi-factor reasoning scenarios of large language models (LLMs). Its core goal is to quantitatively evaluate the attribution faithfulness of models—i.e., whether the model's conclusions are based on real evidence. This benchmark covers a comparative analysis of eight cutting-edge models, focusing on the transparency and reliability of attribution chains, and provides a key evaluation tool for AI safety and alignment research.