Zing Forum

Reading

Tencent Hunyuan Hy3-preview Open Source: A Cost-Effective Reasoning and Agent Model with 295B MoE Architecture

The Tencent Hunyuan team has open-sourced the Hy3-preview model, which uses a MoE architecture with 295B total parameters and 21B active parameters. It performs excellently in STEM reasoning, code generation, and Agent tasks while maintaining outstanding cost-effectiveness.

腾讯混元Hy3-previewMoE大语言模型开源Agent代码生成推理模型
Published 2026-04-23 23:09Recent activity 2026-04-23 23:25Estimated read 6 min
Tencent Hunyuan Hy3-preview Open Source: A Cost-Effective Reasoning and Agent Model with 295B MoE Architecture
1

Section 01

【Introduction】Tencent Hunyuan Hy3-preview Open Source: A Cost-Effective Reasoning and Agent Model with 295B MoE Architecture

The Tencent Hunyuan team officially open-sourced the Hy3-preview model on April 23, 2026. This model uses a MoE architecture with 295B total parameters and 21B active parameters. It performs excellently in STEM reasoning, code generation, and Agent tasks while maintaining outstanding cost-effectiveness. The model supports a 256K long context window, offers multiple deployment methods, and has been open-sourced on multiple platforms, providing a new option for cost-effective reasoning solutions.

2

Section 02

Model Background and Core Parameter Analysis

Hy3-preview is the strongest model released by the Tencent Hunyuan team to date, using a Mixture of Experts (MoE) architecture: total parameters of 295 billion, only 21 billion parameters are activated per inference; equipped with an MTP layer of 3.8 billion parameters, supporting a 256K context window, with a vocabulary size of 120,832; the expert routing design selects 8 out of 192 experts for computation, and sparse activation significantly reduces inference costs.

3

Section 03

Breakthrough Performance in STEM and Complex Reasoning Capabilities

Hy3-preview performs outstandingly in the STEM field: it achieved excellent results in high-difficulty benchmark tests such as FrontierScience-Olympiad and IMOAnswerBench; it obtained impressive scores in the 2026 Tsinghua Yau Mathematical Sciences Center PhD Qualification Exam and the 2025 CHSBO competition; its accuracy rate on the MATH benchmark (mathematical reasoning) reached 76.28%, and 95.37% on GSM8K (primary school math problems). This is attributed to the team's reconstruction of reinforcement learning training infrastructure and expansion of training task scale.

4

Section 04

Improvements in In-Context Learning and Instruction Following Capabilities

To evaluate real-scenario capabilities, the team built the CL-bench and CL-bench-Life evaluation benchmarks. Hy3-preview's 256K context window can handle entire books, long technical documents, etc., without segmentation to avoid information loss; it has achieved significant improvements in both in-context learning and instruction following.

5

Section 05

Practical Progress in Code Generation and Agent Capabilities

Hy3-preview has made significant progress in code generation and Agent fields: it has strong competitiveness in code Agent benchmarks such as SWE-bench Verified and Terminal-Bench 2.0; it performs well in search Agent benchmarks like BrowseComp and WideSearch; it achieved high scores in complex Agent scenarios like ClawEval and WildClawBench, making Agent capabilities practical; internal evaluations (Hy3-backend, Hy-Vibe Bench, etc.) also show strong competitiveness.

6

Section 06

Deployment and Usage Guide

Hy3-preview supports deployment via vLLM and SGLang, and provides an OpenAI-compatible API; inference modes include no_think (default, simple Q&A), low (light chain of thought), and high (deep chain of thought, complex reasoning); recommended generation parameters are temperature=0.9, top_p=1.0, supporting BF16 precision, and can run efficiently on mainstream frameworks.

7

Section 07

Summary of Open Source Ecosystem and Technical Value

The Hy3-preview model weights have been synchronously open-sourced on platforms such as Hugging Face and ModelScope, including Instruct and Base pre-trained models; the core innovation is a 14x parameter sparsity (only 21B of the 295B total parameters are activated), maintaining performance comparable to Kimi-K2 (1043B) and DeepSeek-V3 (671B) but with only 1/3 to 1/2 of the active parameters, providing a new path for the popularization of large models.