Section 01
MetaProbe: A Comprehensive Benchmark for Evaluating LLM Metacognitive Capabilities (Introduction)
MetaProbe is a benchmark framework specifically for evaluating the metacognitive capabilities of large language models (LLMs). Through four core dimensions—confidence calibration, error detection, knowledge boundary, and confidence stability—it tests whether models truly "know what they know" and "know what they don't know". This framework fills a gap in the LLM evaluation field and is of great significance for improving the reliability of AI systems and reducing the risk of hallucinations.