Section 01
[Introduction] Core Highlights of the RSAT Project: Small Models + Reinforcement Learning for Interpretable Table Reasoning
The RSAT (Reasoning with Small models on Tables) project focuses on enabling small language models (e.g., 7B parameter scale) to achieve high-quality table reasoning and generate cell-level fine-grained citations. Its core innovation lies in adopting a training strategy that combines Supervised Fine-Tuning (SFT) and Group Relative Policy Optimization (GRPO) reinforcement learning, balancing reasoning faithfulness, citation accuracy, and model efficiency, thus providing solutions for interpretable AI applications in high-risk scenarios such as finance and healthcare.