Grok 3 Beta was released February 17, 2025 as xAI's full-scale Grok 3 reasoning model, trained on the Colossus supercomputer cluster. It's a step up from the Grok 2 generation on reasoning-heavy tasks across mathematics, science, and code.
Grok 3 Beta supports a context window of 131.1K tokens and up to 131.1K tokens per response. It reports scores on benchmarks including AIME, MATH-500, and GPQA. The model handles complex multi-step reasoning, extended analysis, and nuanced instruction following. It's designed for tasks where depth of reasoning matters more than raw response speed.
The model is accessible through Vercel AI Gateway at $3.0 per million input tokens and $15.0 per million output tokens. For latency-sensitive applications that can tolerate some quality tradeoff, the Grok 3 Fast variant offers faster inference at the same context window size.