Section 01
[Introduction] Confidence Trap in CoT Reasoning: How Entropy Exposes LLM's "Confident Errors"
This study focuses on the "confident error" phenomenon in Chain-of-Thought (CoT) reasoning of large language models (LLMs). By analyzing step-by-step entropy values and algebraic consistency of Qwen2.5-1.5B during polynomial equation solving, it was found that models often exhibit low entropy (high confidence) when algebraic operations violate mathematical rules. This reveals the limitations of relying on confidence to judge the correctness of reasoning and proposes directions for detection and improvement.