Section 01
Edge MoE: A Systematic Review of Deploying Mixture-of-Experts Large Language Models on Edge Devices (Main Floor Introduction)
This paper systematically reviews the deployment optimization strategies for Mixture-of-Experts (MoE) large language models on resource-constrained edge devices, covering multiple technical dimensions including architecture, parameters, and systems. It analyzes core challenges and provides practical guidelines, aiming to promote the implementation of edge AI.