DeepSeek

DeepSeek v3.1

DeepSeekBalanced
ThinkingTool UseStructured Output

About this model

DeepSeek-V3.1 is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes via prompt templates. It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) The model improves tool use, code generation, and reasoning efficiency, achieving performance comparable to DeepSeek-R1 on difficult benchmarks while responding more quickly. It supports structured tool calling, code agents, and search agents, making it suitable for research, coding, and agentic workflows. It succeeds the [DeepSeek V3-0324](/deepseek/deepseek-chat-v3-0324) model and performs well on a variety of tasks.

Performance Tier

Balanced

DeepSeek v3.1 is a balanced model from DeepSeek : strong performance at a reasonable price.

Strong cost-performance ratio. Reliable for most professional use cases without premium pricing.

Pricing

This model is included in Elosia plans
Typeper 1M tokens
Input (prompt)$0.150
Output (completion)$0.750

Capabilities

Context Length33K
Max Output Tokens7K
TokenizerDeepSeek
Inputtext
Outputtext
Release DateAugust 21, 2025

Benchmarks

General Intelligence
MMLU
87.5%
GPQA Diamond
72.8%
Mathematics
MATH-500
89.2%
Programming
HumanEval
89%

Recommended Use Cases

CodingMathematicsAnalysisGeneral Chat

Strengths

  • Strong math and coding performance at a very low price
  • Open-weight model enabling self-hosting
  • Good general reasoning for its cost tier

Limitations

  • Superseded by DeepSeek v3.2 on most benchmarks
  • Weaker creative writing compared to proprietary models

Resources

This model may use your data for training

Similar Models