Skip to content

Grok 3 Mini Beta

Grok 3 Mini Beta is xAI's compact reasoning model in the Grok 3 family. It provides efficient inference for tasks that need solid reasoning without the full computational overhead of the full-scale Grok 3, within a context window of 131.1K tokens.

Tool Use
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'xai/grok-3-mini',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • How does Grok 3 Mini Beta compare to the full Grok 3?

    Grok 3 Mini Beta is a smaller, more efficient model that trades some reasoning depth for lower latency and cost. The full Grok 3 still fits better when you need the longest multi-step reasoning chains.

  • What is the difference between Grok 3 Mini Beta and Grok 3 Mini Fast?

    Both share the same compact architecture, but Grok 3 Mini Fast adds further latency optimization for applications where response speed is the top priority.

  • What is the context window for Grok 3 Mini Beta?

    The context window is 131.1K tokens.

  • What does Grok 3 Mini Beta cost?

    See the pricing section on this page for today's rates. AI Gateway exposes each provider's pricing for Grok 3 Mini Beta.

  • How do I authenticate with Grok 3 Mini Beta through Vercel AI Gateway?

    Use your Vercel AI Gateway API key with xai/grok-3-mini as the model identifier. AI Gateway handles routing and provider management automatically.

  • Is Grok 3 Mini Beta suitable for high-volume batch processing?

    Yes. Its lower per-token cost makes it well-suited for batch workloads like document summarization, data extraction, and content classification at scale.

  • Does Vercel AI Gateway support Zero Data Retention for Grok 3 Mini Beta?

    Zero Data Retention is not currently available for this model. ZDR on AI Gateway applies to direct gateway requests; BYOK flows aren't covered. See https://vercel.com/docs/ai-gateway/capabilities/zdr for details.