Section 01
Introduction: Tsinghua Open-Sources Chitu Inference Engine, Fully Supporting Domestic Chips and Full-Scenario Deployment
The Chitu (赤兔) inference framework, open-sourced by Tsinghua University's PACMAN Lab, is positioned as a production-grade large model inference engine with both high performance and stability. Its core advantages include: supporting the full range of NVIDIA GPUs and domestic chips such as Huawei Ascend, Moore Threads, Muxi, and Hygon; covering full-scenario deployment from pure CPU, single-card GPU to large-scale clusters; being compatible with mainstream large models like DeepSeek, Qwen, and GLM; and having technical highlights such as FP4/FP8 quantization and CPU+GPU heterogeneous hybrid inference, which can handle real concurrent business traffic.