Section 01
OxiLLaMa: A Pure Rust LLM Inference Engine (Memory-Safe Alternative to llama.cpp)
OxiLLaMa is an LLM inference engine fully rewritten in Rust, with zero dependencies on C/C++/Fortran. It supports 20 model architectures and 25 quantization formats, provides OpenAI-compatible API services, and aims to build cross-platform, auditable, memory-safe AI inference infrastructure. It is a core component of the COOLJAPAN pure Rust tech stack.