Section 01
Introduction: Mechanistic Interpretability Study of GPT-2's Syntax Circuit
This article focuses on the GPT-2 Small model, comprehensively applying three core techniques—linear probing, causal activation patching, and sparse autoencoders—to systematically explore how it encodes and utilizes part-of-speech information. It aims to uncover the internal "black box" working mechanism of large language models (LLMs) and provide a practical path for the development of explainable AI.