Section 01
Introduction: Training a 200M-Parameter Large Model on Mobile Becomes a Reality, Butterfly Transform Brings Breakthrough
An open-source project named "on-device-butterfly-llm" enables direct training of a 200M-parameter large language model on Android phones without cloud support. Using the Diagonal-Interleaved Butterfly (DIB) attention mechanism and NEON SIMD optimization, it achieves 10x faster inference speed than traditional methods and reduces memory usage by over 50x, opening up new possibilities for privacy-sensitive applications and offline scenarios.