Section 01
【Introduction】In-Place TTT: A New Framework for Enabling LLMs to Evolve During Inference
This article proposes the In-Place Test-Time Training (TTT) framework, which enables large language models (LLMs) to dynamically update parameters during inference by using the final projection matrix of MLP blocks as adaptable fast weights and designing an objective function optimized for autoregressive language modeling. Experiments show that a 4B-parameter model achieves excellent performance on tasks with up to 128k context, opening a new path for the continuous learning of LLMs.