Section 01
XTuner V1: Introduction to the Next-Generation Training Engine for Ultra-Large-Scale MoE Models
XTuner V1 is a next-generation LLM training engine developed by Shanghai AI Laboratory, specifically designed for ultra-large-scale Mixture-of-Experts (MoE) models. It breaks through the limitations of traditional 3D parallel architecture, supports training models up to 1 trillion parameters, and achieves training efficiency exceeding H800 on Ascend NPUs. Its core advantages include simplified parallel strategies, support for long-sequence training, cross-hardware platform compatibility, and full-link algorithm capabilities, aiming to lower the research threshold for ultra-large-scale MoE models and promote the construction of domestic computing power ecosystems.