Section 01
Uzu: High-Performance Local LLM Inference Engine for Apple Silicon (Main Guide)
Uzu is a local AI inference engine specifically designed for Apple Silicon (M1/M2/M3 series) to solve the trade-off between cloud and local AI deployment. It enables zero-latency inference, full data privacy, and easy integration via the TypeScript library uzu-ts. Key features include speculative decoding for speed enhancement, dynamic/static context management, scene-specific presets, cloud hybrid inference, and structured output support.