Tencent Hy3 Preview
Tencent Hunyuan's most powerful open-source model — also known as Hunyuan 3 (Hunyuan 3.0). A 295B-parameter MoE architecture with 21B active parameters and 256K context window. Rebuilt from scratch in 90 days by the Hy team, built for real products like Yuanbao. Now available via Tencent Cloud and OpenRouter.
Why Hy3 Preview Matters
Three core capabilities fully upgraded in Hy3 Preview — from complex reasoning to coding agents, every capability refined through real product scenarios. A leap forward for the Hunyuan family.
Hy3 Preview: STEM & Reasoning
Excels on challenging STEM benchmarks like FrontierScience-Olympiad and IMOAnswerBench. Achieved the highest domestic score on the Tsinghua Qiuzhen College Math PhD qualifying exam, demonstrating strong generalizable reasoning that rivals OpenAI and DeepSeek models.
Hy3 Preview: Context Learning
Real-world tasks demand parsing messy, lengthy contexts and following complex rules. The Hunyuan team built CL-bench from real business scenarios to measure context learning. Solid gains in both context learning and instruction following — a key differentiator from Ling 2.6 Flash and other competing models.
Hy3 Preview: Code & Agent
Coding and agents saw the biggest gains. Competitive scores on mainstream coding agent benchmarks (SWE-bench Verified, Terminal-Bench 2.0) and search agent benchmarks (BrowseComp, WideSearch). Powers Tencent's OpenCode workflows and integrates seamlessly with tools like CodeBuddy.
Hy3 Preview Architecture
The Hunyuan 3.0 Mixture-of-Experts architecture fuses fast and slow thinking, achieving optimal balance between parameter scale and performance.
Hy3 Preview Mixture-of-Experts
Hy3 Preview (Tencent Hy3) uses a MoE architecture with 295B total parameters, activating only 21B per forward pass. Top-8 out of 192 experts are activated, routing routine queries to fast pattern-matching experts and complex problems to deeper reasoning chains. Chief AI Scientist Shunyu Yao describes the team as "exploring non-homogeneous capabilities" — features shaped by specific products.
This is not a compromise — it is a deliberate ceiling. Beyond roughly one trillion parameters, multi-node deployment erodes latency and throughput faster than marginal capability gains justify. While competitors like Nemotron 3 Super and DeepSeek-V3 push toward larger scales, the Hunyuan team's 300B range is intentional for cost-performance.
| Property | Value |
|---|---|
| Architecture | MoE |
| Total Parameters | 295B |
| Activated Parameters | 21B |
| MTP Layer Params | 3.8B |
| Layers | 80 |
| Attention Heads | 64 (GQA, 8 KV) |
| Hidden Size | 4096 |
| Intermediate Size | 13312 |
| Context Length | 256K |
| Vocab Size | 120832 |
| Experts | 192, top-8 |
| Precision | BF16 |
Hy3 Preview Benchmark Results
Leading performance across multiple benchmarks, outpacing DeepSeek-V3 and GLM-4.5 in math, coding, and multilingual tasks.
Math — MATH (4-shot)
Math — GSM8K (4-shot)
Code — LiveCodeBench-v6
Coding Agent — SWE-bench Verified
Terminal Agent — Terminal-Bench 2.0
Multilingual — MMMLU (5-shot)
Hy3 Preview Pre-trained Model Comparison
| Benchmark | Kimi-K2 32B / 1043B |
DeepSeek-V3 37B / 671B |
GLM-4.5 32B / 355B |
Hy3 Preview 21B / 295B |
|---|---|---|---|---|
| MMLU (5-shot) | 88.24 | 87.68 | 87.73 | 87.42 |
| MMLU-Pro (5-shot) | 65.98 | 63.98 | 63.67 | 65.76 |
| SuperGPQA (5-shot) | 51.10 | 46.17 | 49.64 | 51.60 |
| MATH (4-shot) | 71.20 | 59.37 | 61.00 | 76.28 |
| GSM8K (4-shot) | 93.46 | 88.15 | 90.06 | 95.37 |
| LiveCodeBench-v6 | 30.86 | 29.31 | 27.43 | 34.86 |
| MMMLU (5-shot) | 77.63 | 79.54 | 79.26 | 80.15 |
| INCLUDE (5-shot) | 75.66 | 77.86 | 76.27 | 78.64 |
Hy3 Preview: 90 Days, From Scratch
In February 2026, Tencent tore down the Hunyuan infrastructure and rebuilt from scratch. Three core principles drove the birth of Hy3 Preview (Hunyuan 3).
Infrastructure Rebuild
Tore down pre-training and RL infrastructure, rebuilt from scratch around three principles: capability systematisation, evaluation authenticity, and cost-performance.
Training Begins
New infrastructure ready. Training began on the rebuilt framework. The Hy model team merged with Yuanbao, CodeBuddy, WorkBuddy, and other product teams into a single development loop.
Model Goes Live
Hy3 Preview went live, integrated into Yuanbao, CodeBuddy, WorkBuddy and other products. Real user feedback began driving the optimization loop.
Open-Source Release
Model weights open-sourced on Hugging Face, ModelScope, and GitCode. API hosted on Tencent Cloud and available via OpenRouter, priced at roughly one-tenth of OpenAI GPT-4-class rates.
Hy3 Preview Product Ecosystem
Not just built for products — built with them. Tencent Hunyuan's live product metrics directly shape Hy3 Preview training priorities.
Yuanbao
AI chatbot powered by Hy3 Preview, interactions that feel like a real friend
CodeBuddy
AI coding assistant for code generation and debugging
WorkBuddy
AI productivity assistant for documents and scheduling
ima
AI creative tool for multimodal content generation
QQ Browser
Smart browser with AI-powered search and summarization
Deploy Hy3 Preview
Deploy Hy3 Preview with vLLM or SGLang, OpenAI-compatible API, ready out of the box. Also available on Tencent Cloud and OpenRouter.
from openai import OpenAI client = OpenAI(base_url="http://localhost:8000/v1", api_key="EMPTY") response = client.chat.completions.create( model="hy3-preview", messages=[ {"role": "user", "content": "Hello! Can you briefly introduce yourself?"}, ], temperature=0.9, top_p=1.0, # reasoning_effort: "no_think" (default) | "low" | "high" (deep CoT) extra_body={"chat_template_kwargs": {"reasoning_effort": "no_think"}}, ) print(response.choices[0].message.content)
# Build from source uv venv --python 3.12 --seed --managed-python source .venv/bin/activate git clone https://github.com/vllm-project/vllm.git cd vllm && uv pip install -e . --torch-backend=auto # Launch with MTP vllm serve tencent/Hy3-preview \ --tensor-parallel-size 8 \ --speculative-config.method mtp \ --speculative-config.num_speculative_tokens 1 \ --tool-call-parser hy_v3 \ --reasoning-parser hy_v3 \ --served-model-name hy3-preview
# Build from source git clone https://github.com/sgl-project/sglang cd sglang pip3 install pip --upgrade pip3 install "transformers>=5.6.0" pip3 install -e "python" # Launch with MTP python3 -m sglang.launch_server \ --model tencent/Hy3-preview \ --tp 8 \ --tool-call-parser hunyuan \ --reasoning-parser hunyuan \ --speculative-algorithm EAGLE \ --served-model-name hy3-preview
Hy3 Preview API Pricing
Roughly one-tenth of OpenAI GPT-4-class rates. Exceptional cost efficiency vs DeepSeek, Nemotron 3 Super, and Ling 2.6 Flash.
Hy3 Preview Input (0-16K)
per million tokens
Output
per million tokens
Hy3 Preview vs GPT-4
of the cost
Hy3 Preview FAQ
Start Building with Hy3 Preview
Open-source, cost-efficient, built for products. Whether you're building chatbots like Yuanbao, coding assistants via OpenCode, or complex agent workflows — Hy3 Preview by Tencent Hunyuan is your starting point.