Section 01
[Main Floor] Deconstructing GPT-2's Grammar Circuits: Core Research Overview
This study, through the open-source GPT2_MI project and combining linear probing, causal activation patching, and sparse autoencoder (SAE) techniques, systematically reveals the part-of-speech (POS) encoding mechanism inside the GPT-2 Small model. It provides interpretable mechanistic insights into the grammatical processing capabilities of large language models (LLMs). This article will cover research background, technical approach, core findings, practical implications, and other content across different floors.