Grok 3 Mini Beta
Grok 3 Mini Beta is xAI's compact reasoning model in the Grok 3 family. It provides efficient inference for tasks that need solid reasoning without the full computational overhead of the full-scale Grok 3, within a context window of 131.1K tokens.
import { streamText } from 'ai'
const result = streamText({ model: 'xai/grok-3-mini', prompt: 'Why is the sky blue?'})Frequently Asked Questions
How does Grok 3 Mini Beta compare to the full Grok 3?
Grok 3 Mini Beta is a smaller, more efficient model that trades some reasoning depth for lower latency and cost. The full Grok 3 still fits better when you need the longest multi-step reasoning chains.
What is the difference between Grok 3 Mini Beta and Grok 3 Mini Fast?
Both share the same compact architecture, but Grok 3 Mini Fast adds further latency optimization for applications where response speed is the top priority.
What is the context window for Grok 3 Mini Beta?
The context window is 131.1K tokens.
What does Grok 3 Mini Beta cost?
See the pricing section on this page for today's rates. AI Gateway exposes each provider's pricing for Grok 3 Mini Beta.
How do I authenticate with Grok 3 Mini Beta through Vercel AI Gateway?
Use your Vercel AI Gateway API key with
xai/grok-3-minias the model identifier. AI Gateway handles routing and provider management automatically.Is Grok 3 Mini Beta suitable for high-volume batch processing?
Yes. Its lower per-token cost makes it well-suited for batch workloads like document summarization, data extraction, and content classification at scale.
Does Vercel AI Gateway support Zero Data Retention for Grok 3 Mini Beta?
Zero Data Retention is not currently available for this model. ZDR on AI Gateway applies to direct gateway requests; BYOK flows aren't covered. See https://vercel.com/docs/ai-gateway/capabilities/zdr for details.