Section 01
[Introduction] ZAYA1-8B: A Small-Scale MoE Model Setting a New Benchmark for Challenging Large Models' Reasoning Performance
ZAYA1-8B is a Mixture of Experts (MoE) reasoning model with only 700M active parameters (total parameters: 8B). Through four-stage Reinforcement Learning (RL) training and the Markovian RSA test-time computation method, it achieves a score of 91.9% on the AIME'25 benchmark, with performance approaching ultra-large models like Gemini-2.5 Pro. Built on the MoE++ architecture, this model was trained entirely on AMD's full-stack computing platform, challenging the traditional belief that "reasoning ability is positively correlated with model size."