Section 01
Hoosh: Introduction to the High-Performance AI Inference Gateway Built with Rust
Hoosh is a high-performance AI inference gateway written in Rust, designed to solve problems in AI application development such as switching between multiple LLM providers and balancing cost and performance between local inference and cloud APIs. It supports unified routing and scheduling for 14 LLM providers, covering both local (e.g., Ollama, llama.cpp) and cloud (e.g., OpenAI, Anthropic) resources, provides an OpenAI-compatible API, and has enterprise-grade features like security and observability required for production environments.