章节 01
mlx-serve: Zig-Native LLM Inference Server for Apple Silicon (Main Guide)
mlx-serve is a pure Zig-language LLM inference server optimized for Apple Silicon (M1/M2/M3/M4), free of Python dependencies. It provides OpenAI and Anthropic compatible APIs, and includes a macOS GUI app MLX Core. Key features include lightweight deployment, high performance, tool calling, and multi-modal support.