We use essential cookies to make our site work. With your consent, we may also use non-essential cookies to improve user experience and analyze website traffic…

🚀 New models by Bria.ai, generate and edit images at scale 🚀

deepseek-ai logo

deepseek-ai/

DeepSeek-V3.2-Exp

$0.27

in

$0.40

out

DeepSeek-V3.2-Exp is an intermediate step toward the next-generation architecture of the DeepSeek models by introducing DeepSeek Sparse Attention—a sparse attention mechanism designed to explore and validate optimizations for training and inference efficiency in long-context scenarios.

deepseek-ai/DeepSeek-V3.2-Exp cover image
deepseek-ai/DeepSeek-V3.2-Exp cover image
DeepSeek-V3.2-Exp

Ask me anything

0.00s

Settings

Model Information

We are excited to announce the official release of DeepSeek-V3.2-Exp, an experimental version of our model. As an intermediate step toward our next-generation architecture, V3.2-Exp builds upon V3.1-Terminus by introducing DeepSeek Sparse Attention—a sparse attention mechanism designed to explore and validate optimizations for training and inference efficiency in long-context scenarios.

This experimental release represents our ongoing research into more efficient transformer architectures, particularly focusing on improving computational efficiency when processing extended text sequences.

  • DeepSeek Sparse Attention (DSA) achieves fine-grained sparse attention for the first time, delivering substantial improvements in long-context training and inference efficiency while maintaining virtually identical model output quality.

  • To rigorously evaluate the impact of introducing sparse attention, we deliberately aligned the training configurations of DeepSeek-V3.2-Exp with V3.1-Terminus. Across public benchmarks in various domains, DeepSeek-V3.2-Exp demonstrates performance on par with V3.1-Terminus.

BenchmarkDeepSeek-V3.1-TerminusDeepSeek-V3.2-Exp
Reasoning Mode w/o Tool Use
MMLU-Pro85.085.0
GPQA-Diamond80.779.9
Humanity's Last Exam21.719.8
LiveCodeBench74.974.1
AIME 202588.489.3
HMMT 202586.183.6
Codeforces20462121
Aider-Polyglot76.174.5
Agentic Tool Use
BrowseComp38.540.1
BrowseComp-zh45.047.9
SimpleQA96.897.1
SWE Verified68.467.8
SWE-bench Multilingual57.857.9
Terminal-bench36.737.7