Hunyuan Vision

Hunyuan Vision
A multi-modal visual large model developed by Tencent Hunyuan Team, supporting video generation, image editing, and cross-modal association reasoning, integrating 3D VAE compression technology and MLLM text encoder
Intelligence(Strong)
Speed(Slow)
Input Supported Modalities
Yes
Is Reasoning Model
32,000
Context Window
16,000
Maximum Output Tokens
2024-12-31
Knowledge Cutoff
Pricing
¥18 /M tokens
Input
¥18 /M tokens
Output
¥18 /M tokens
Blended Price
Quick Simple Comparison
Hunyuan-T1-20250403
¥0.14
Hunyuan-Vision
¥2.5
HunYuan-TurboS
¥0.11
Basic Parameters
GPT-4.1 Technical Parameters
Parameter Count
Not Announced
Context Length
32.00k tokens
Training Data Cutoff
2024-12-31
Open Source Category
Open Source
Multimodal Support
Text, Image
Throughput
1,650
Release Date
2025-01-03
Response Speed
22.3 tokens/s
Benchmark Scores
Below is the performance of claude-monet in various standard benchmark tests. These tests evaluate the model's capabilities in different tasks and domains.
Intelligence Index
8870
Large Language Model Intelligence Level
Coding Index
-
Indicator of AI model performance on coding tasks
Math Index
-
Capability indicator in solving mathematical problems, mathematical reasoning, or performing math-related tasks
MMLU Pro
-
Massive Multitask Multimodal Understanding - Testing understanding of text, images, audio, and video
GPQA
-
Graduate Physics Questions Assessment - Testing advanced physics knowledge with diamond science-level questions
HLE
62.7
The model's comprehensive average score on the Hugging Face Open LLM Leaderboard
LiveCodeBench
-
Specific evaluation focused on assessing large language models' ability in real-world code writing and solving programming competition problems
SciCode
-
The model's capability in code generation for scientific computing or specific scientific domains
HumanEval
-
Score achieved by the AI model on the specific HumanEval benchmark test set
Math 500 Score
-
Score on the first 500 larger, more well-known mathematical benchmark tests
AIME Score
-
An indicator measuring an AI model's ability to solve high-difficulty mathematical competition problems (specifically AIME level)
Featured Recommended AI Models
Gemini 2.0 Flash Lite (Preview)
google

¥0.58
Input tokens/million
¥2.16
Output tokens/million
1M
Context Length
Gemini 1.0 Pro
google

¥3.6
Input tokens/million
¥10.8
Output tokens/million
33k
Context Length
Qwen2.5 Coder Instruct 32B
alibaba

¥0.65
Input tokens/million
¥0.65
Output tokens/million
131k
Context Length
GPT 4
openai

¥216
Input tokens/million
¥432
Output tokens/million
8192
Context Length
Gemini 1.5 Flash 8B
google

¥0.58
Input tokens/million
¥2.16
Output tokens/million
1M
Context Length
Gemma 3 4B Instruct
google

-
Input tokens/million
-
Output tokens/million
128k
Context Length
Gemini 2.0 Pro Experimental (Feb '25)
google

-
Input tokens/million
-
Output tokens/million
2M
Context Length
Llama 3.2 Instruct 11B (Vision)
meta

¥0.43
Input tokens/million
¥0.43
Output tokens/million
128k
Context Length