Section 01
Introduction: SycoQA Dataset—A New Benchmark for Evaluating Sycophantic Hallucinations in LLMs
This article introduces the SycoQA dataset, a new benchmark tool specifically designed to evaluate sycophantic hallucinations in large language models (LLMs). Sycophantic hallucinations refer to the model's behavior of distorting facts to cater to the user's opinions (different from traditional factual hallucinations). The dataset detects model behavior through carefully designed question-answer pairs and is of great significance for AI safety and alignment research.