DeepSeek V3 0324
DeepSeek V3 0324 is DeepSeek's open-source 671B Mixture-of-Experts language model released December 26, 2024. It achieves 3x the inference throughput of DeepSeek-V2 while matching closed-source models in published benchmark evaluations.
import { streamText } from 'ai'
const result = streamText({ model: 'deepseek/deepseek-v3', prompt: 'Why is the sky blue?'})Frequently Asked Questions
What is the architecture of DeepSeek V3 0324?
A sparse Mixture-of-Experts (MoE) model with 671B total parameters, activating 37B per forward pass. The context window is 163.8K tokens.
What is the inference speed of DeepSeek V3 0324?
Roughly 3x faster than DeepSeek-V2. Live throughput metrics on this page update based on real traffic.
How does DeepSeek V3 0324 differ from DeepSeek-R1?
DeepSeek V3 0324 is a general-purpose chat and instruction model. DeepSeek-R1 is a reasoning specialist trained with reinforcement learning to generate extended chain-of-thought for math, code, and formal reasoning tasks.
Is DeepSeek V3 0324 open-source?
Yes. Model weights and the research paper are openly published.
Does DeepSeek V3 0324 maintain API compatibility with DeepSeek-V2?
Yes. It maintains backward API compatibility, so upgrading from V2 requires minimal migration effort.
What context window does DeepSeek V3 0324 support?
163.8K tokens, validated through Needle In A Haystack evaluations across the full range.