Zing Forum

Reading

Millrace: An Autonomous Software Delivery Runtime for Long-Running AI Programming Tasks

Millrace is a local autonomous software delivery runtime designed specifically for long-running coding tasks. It provides reliable support for unattended end-to-end agent workflows through file-driven governance, validated execution phases, and self-healing recovery mechanisms.

AI编程智能体工作流软件交付运行时OpenClaw长时任务自治系统
Published 2026-04-05 17:43Recent activity 2026-04-05 17:49Estimated read 5 min
Millrace: An Autonomous Software Delivery Runtime for Long-Running AI Programming Tasks
1

Section 01

[Main Floor] Millrace: An Autonomous Software Delivery Runtime Filling the Gap in Long-Running AI Programming Tasks

Millrace is a local autonomous software delivery runtime designed specifically for long-running coding tasks. It addresses the shortcomings of mainstream AI programming tools in long-running tasks (such as lack of persistent state, recoverable execution, and governance handover) through file-driven governance, validated execution phases, and self-healing recovery mechanisms, providing reliable support for unattended end-to-end agent workflows.

2

Section 02

Background: Shortcomings of Mainstream AI Programming Tools in Long-Running Tasks

Current mainstream interactive programming tools like GitHub Copilot and Cursor focus on short-term conversational interactions, but long-running tasks (such as cross-hour/day requirement analysis and architecture design) lack capabilities like queue discipline, persistent state, and recoverable execution, relying on unreliable implicit mechanisms like shell history and chat logs. Millrace's core insight: Long-running AI programming work requires a true runtime environment with lifecycle management, queue changes, and persistent state.

3

Section 03

Architecture and Core Mechanisms: Control Plane Separation and Self-Healing Design

Millrace adopts a layered architecture: the underlying runtime engine manages lifecycle, state, etc.; the upper layer is divided into an execution plane (delivery work) and a research plane (requirement analysis), which share the engine but have independent states. The file-driven design (millrace.toml configuration + agents directory storing queues, logs, etc.) ensures the workspace is inspectable and backupable. Core mechanisms include: frozen plan (fixing the execution plan before running to avoid configuration drift), self-healing recovery (disk-persisted state allows recovery of unfinished tasks after crashes).

4

Section 04

Research-Execution Handover and External Supervision: Governance Process and Security Contract

The research plane handles requirement analysis via the add-idea command, generates specifications and execution tasks, then hands them over to the execution plane, separating requirement review from code implementation. It supports OpenClaw-compatible external supervision: supervisors obtain status via supervisor report --json, and use CLI operations with signers (like pause/resume) to ensure action traceability, facilitating audit for multi-agent collaboration.

5

Section 05

Tool Experience and Installation Guide: TUI Interface and Simple Deployment

The TUI is based on the Textual framework, providing multi-panel operations (overview, queue, logs, etc.), automatic health checks before startup, and support for operator/debug modes. Installation via PyPI: python3 -m pip install millrace-ai; create workspace millrace init; check readiness doctor; add task add-task; start execution start --once/--daemon; start TUI python3 -m millrace_engine.tui.

6

Section 06

Design Philosophy and Summary: Explicit Boundaries and Production-Grade Application Scenarios

Design principles: Explicit boundaries (independent surfaces for lifecycle, state, etc.), recoverability first, honest failure (exposing blockages/partial progress). Application scenarios: Long-running AI programming tasks like large-scale code migration and multi-module refactoring. The project is MIT open-source with clear code and complete documentation, providing a reference architecture and infrastructure for production-grade AI programming systems.