Section 01
[Introduction] New Scheme for Large Model Inference Task Decomposition and Edge Collaborative Computing
This article proposes a large model inference task decomposition and edge collaboration framework for resource-constrained wireless devices. Using an LLM planner to achieve subtask difficulty prediction and dynamic scheduling, it delivers significant results in WiFi network environments: 20% lower latency and an 80% increase in overall gain, providing key technical references for efficient deployment of large models in edge scenarios.