Section 01
[Introduction] R2ABench: Evaluating LLMs' Ability to Generate Software Architecture from Requirements
R2ABench is a new benchmark designed to evaluate large language models (LLMs) ability to generate software architecture diagrams from requirements documents. This study reveals fundamental flaws in LLMs' relational reasoning capabilities and provides a standardized evaluation foundation for LLM-driven software architecture generation research. This article will discuss aspects including background, methodology, evaluation results, and practical implications.