Section 01
[Introduction] Fast-Slow Training (FST) Framework: Solving the Core Dilemma of LLM Continuous Learning
Institutions including the University of California, Berkeley, proposed the Fast-Slow Training (FST) framework, which treats model parameters as "slow weights" (to maintain general reasoning abilities) and optimized context as "fast weights" (to absorb task-specific information). It achieves task specialization while preserving general capabilities. Experiments show a 3x improvement in sample efficiency, a 70% reduction in KL divergence, significantly better performance than traditional RL in continuous learning, and effective mitigation of catastrophic forgetting and loss of plasticity.