Section 01
Introduction: InteractComp—A Systematic Evaluation Framework for Interactive Reasoning Capabilities of Large Language Models
This article introduces InteractComp, an evaluation framework specifically designed to assess the interactive reasoning capabilities of large language models. It supports multiple interaction modes, includes a built-in ReAct-style agent, and provides an asynchronous evaluation pipeline, offering a standardized tool for the systematic analysis of model decision-making abilities. It fills the gap where traditional single-turn question-answering benchmarks fail to evaluate interactive reasoning capabilities.