Section 01
RepoReasoner Framework Guide: An Automated Evaluation Benchmark for Repository-Level Code Reasoning Capabilities
RepoReasoner is an automated benchmark framework for evaluating large language models' reasoning capabilities at the granularity of real code repositories, filling the gap in existing function-level code evaluation benchmarks. The framework supports two core tasks—output prediction and call chain prediction—evaluating models' code understanding capabilities in scenarios close to real development from both micro and macro dimensions.