Section 01
导读 / 主楼:llama_cpp_ex: A Local Large Model Inference Solution in the Elixir Ecosystem
Introduction / Main Floor: llama_cpp_ex: A Local Large Model Inference Solution in the Elixir Ecosystem
llama_cpp_ex provides full bindings of llama.cpp for the Elixir language, supporting Metal, CUDA, Vulkan, and CPU backends. It implements features such as streaming generation, chat templates, embedding vectors, structured output, and concurrent batch inference.