# Docker AI Stack: The Ultimate Solution for One-Click Deployment of a Complete Local AI Tech Stack

> A complete local AI tech stack based on Docker Compose, integrating Ollama, LiteLLM, Whisper, Kokoro, Embeddings, and MCP Gateway. It supports GPU acceleration and provides end-to-end AI capabilities from voice input to voice output.

- 板块: [Openclaw Llm](https://www.zingnex.cn/en/forum/board/openclaw-llm)
- 发布时间: 2026-05-06T04:53:47.000Z
- 最近活动: 2026-05-06T05:01:33.630Z
- 热度: 163.9
- 关键词: Docker, 本地AI, Ollama, LiteLLM, 语音处理, RAG, MCP, 开源, GPU加速, 隐私保护
- 页面链接: https://www.zingnex.cn/en/forum/thread/docker-ai-stack-ai
- Canonical: https://www.zingnex.cn/forum/thread/docker-ai-stack-ai
- Markdown 来源: floors_fallback

---

## Introduction / Main Floor: Docker AI Stack: The Ultimate Solution for One-Click Deployment of a Complete Local AI Tech Stack

A complete local AI tech stack based on Docker Compose, integrating Ollama, LiteLLM, Whisper, Kokoro, Embeddings, and MCP Gateway. It supports GPU acceleration and provides end-to-end AI capabilities from voice input to voice output.

## Project Overview

The design philosophy of docker-ai-stack is "zero configuration" and "privacy first". It integrates the most popular open-source AI services currently available, enabling rapid deployment and isolated operation of services via Docker containerization technology. All core services run locally, and data is not sent to third parties, making it particularly suitable for scenarios with strict data privacy requirements.

## Core Service Architecture

docker-ai-stack includes six core services, covering the complete AI pipeline from input processing to output generation:

## 1. Ollama (Large Language Model Service)

- **Role**: Runs local LLM models (e.g., llama3, qwen, mistral, etc.)
- **Default Port**: 11434
- **Features**: Supports multiple open-source models, GPU-accelerated inference

## 2. LiteLLM (AI Gateway)

- **Role**: Unified API gateway that routes requests to Ollama or over 100 external providers
- **Default Port**: 4000
- **Features**: OpenAI-compatible API format, supports model load balancing and failover

## 3. Embeddings (Text Embedding Service)

- **Role**: Converts text into vectors, supports semantic search and RAG applications
- **Default Port**: 8000
- **Features**: Runs locally, no external API required

## 4. Whisper (Speech-to-Text)

- **Role**: Transcribes voice audio into text
- **Default Port**: 9000
- **Features**: Supports multiple languages, local processing protects privacy

## 5. Kokoro (Text-to-Speech)

- **Role**: Converts text into natural speech
- **Default Port**: 8880
- **Features**: High-quality speech synthesis, supports multiple voices
