Section 01
SSD: Simple Self-Distillation Significantly Improves Code Generation Capability (Introduction)
Simple Self-Distillation (SSD) improves code generation capability through sampling with specific temperature configurations plus standard supervised fine-tuning training, without needing validators, teacher models, or reinforcement learning. On LiveCodeBench, SSD increases the pass@1 of Qwen3-30B-Instruct from 42.4% to 55.3%. The method is concise and general, applicable to various models and scales.