DeepSeek V4 Pro Thinking enables DeepSeek's chain-of-thought mode on the large-scale Mixture-of-Experts model with a 1M-token context window, built for advanced reasoning, coding, long-horizon agent workflows, knowledge, math, and software engineering tasks.
Added Apr 24, 2026
Context Window
1.0M
Max Output
384.0K
Input Price
$1.74/1M
Output Price
$3.48/1M
Capabilities
Performance metrics and benchmarks
Sourced from Artificial Analysis.
Intelligence Index
51.5
Coding Index
47.5
Agentic Index
67.2
GPQA Diamond
Graduate-level scientific reasoning
88.8%
Better than 96% of models compared
HLE
Humanity's Last Exam
35.9%
Better than 98% of models compared
IFBench
Instruction-following benchmark
76.5%
Better than 96% of models compared
T²-Bench Telecom
Conversational AI agents in dual-control scenarios
96.2%
Better than 97% of models compared
AA-LCR
Long context reasoning evaluation
66.3%
Better than 90% of models compared
GDPval-AA
Economically valuable tasks
52.9%
Better than 96% of models compared
CritPt
Research-level physics reasoning
12.9%
Better than 97% of models compared
SciCode
Python programming for scientific computing
50.0%
Better than 96% of models compared
Terminal-Bench Hard
Agentic coding and terminal use
46.2%
Better than 95% of models compared
AA-Omniscience Accuracy
Proportion of correctly answered questions
43.3%
Better than 94% of models compared
AA-Omniscience Hallucination Rate
Rate of incorrect answers among non-correct responses
88.6%
Better than 32% of models compared
Last updated May 11, 2026
Artificial Analysis