Zing Forum

Reading

Local Proxy Bridges Cursor and Domestic Large Models: ai-ide-local-proxer Enables DeepSeek, Kimi to Support Full Reasoning Chains

ai-ide-local-proxer is a zero-dependency local OpenAI-compatible proxy that allows Cursor IDE and Codex CLI to seamlessly call domestic reasoning models like DeepSeek, Moonshot/Kimi. It addresses the pain points of API format incompatibility and loss of reasoning content, enabling developers to enjoy the full reasoning capabilities of domestic large models without modifying their IDE.

Cursor IDEDeepSeekKimiMoonshotOpenAI API代理国产大模型AI 编程工具Codex CLI推理链
Published 2026-04-28 00:16Recent activity 2026-04-28 01:48Estimated read 6 min
Local Proxy Bridges Cursor and Domestic Large Models: ai-ide-local-proxer Enables DeepSeek, Kimi to Support Full Reasoning Chains
1

Section 01

Introduction: ai-ide-local-proxer — The Bridge Connecting Cursor and Domestic Large Models

ai-ide-local-proxer is a zero-dependency local OpenAI-compatible proxy designed to solve the issues of API format incompatibility and reasoning chain loss between Cursor IDE/Codex CLI and domestic reasoning models like DeepSeek, Moonshot/Kimi. It implements format conversion and field mapping via a local lightweight proxy, allowing developers to enjoy the full reasoning capabilities of domestic large models without modifying IDE configurations.

2

Section 02

Project Background: The Interface Gap Between IDEs and Domestic Models

With the rise of domestic large models, developers want to integrate them into tools like Cursor IDE and Codex CLI. However, these tools natively only support OpenAI-format APIs, and domestic models have different interfaces—leading to failed calls or loss of key reasoning chains (reasoning_content). Additionally, OpenAI's newly launched Responses API is incompatible with the old format, and domestic models haven't fully aligned with the new specifications. Developers face the dilemma of either abandoning their IDE or sacrificing reasoning transparency.

3

Section 03

Core Mechanisms: Format Conversion and Reasoning Chain Re-injection

OpenAI Compatibility Layer

The proxy implements endpoints compliant with OpenAI API specifications, receives chat/completions format requests from Cursor, and converts them into the request body required by domestic models (handling authentication headers, parameter mapping, and response restructuring).

Reasoning Content Re-injection

Re-injects the reasoning_content field returned by domestic models into the conversation history, enabling developers to view the model's thinking process.

Responses API Translation

Provides bidirectional translation for tools like Codex CLI: converts Responses API calls into chat/completions format to send to domestic models, then wraps the response back into Responses API format.

4

Section 04

Technical Highlights: Single-File, Zero-Dependency, and Open-Source License

Single-File Architecture

The entire proxy is a standalone executable file with no npm dependencies, which can be directly copied and run to avoid dependency conflicts.

Zero External Dependencies

Implemented using pure Node.js built-in modules, reducing the attack surface and ensuring long-term maintainability.

MIT Open-Source License

Allows free use, modification, and distribution, making it suitable for integration by individuals and enterprises.

5

Section 05

Usage Guide: Integration Methods for Cursor and Codex CLI

Cursor IDE Integration

In Cursor settings, point the OpenAI API base URL to the local proxy port (default localhost:3000), then enter the domestic model's API key.

Codex CLI Adaptation

Make Codex CLI point to the local proxy via environment variables or configuration files, intercept API calls, and handle format conversion.

Multi-Model Switching

Supports configuring multiple domestic model endpoints; you can switch backends in Cursor via model names (e.g., use DeepSeek for code completion, Kimi for conversations).

6

Section 06

Ecosystem Value: Lowering the Barrier to Using Domestic Models and Promoting Tool Integration

ai-ide-local-proxer lowers the barrier to using domestic large models—developers don't need to wait for official IDE support or write adaptation code. It keeps the toolchain open, promoting integration between the domestic AI ecosystem and international development tools; allows Cursor/Codex CLI users to access cost-effective models while maintaining their original development experience; and helps domestic model vendors expand their user base and accelerate ecosystem maturity.

7

Section 07

Limitations and Outlook: To-Be-Optimized Advanced Features and Community Contribution Directions

Current Limitations

Focuses mainly on core function stability; advanced features like full handling of streaming responses and multi-modal input support need improvement; needs to keep up with the continuous evolution of the OpenAI API.

Future Directions

The community can participate in: supporting more proprietary features of domestic models, optimizing the display format of reasoning content, providing more granular configuration options, etc.