Skip to content

o4-mini

o4-mini advances OpenAI's compact reasoning model line with stronger performance and greater efficiency than o3-mini, adding native tool use and image reasoning.

File InputReasoningTool UseVision (Image)Implicit Caching
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'openai/o4-mini',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • How does o4-mini improve over o3-mini?

    It delivers stronger reasoning performance with greater efficiency, adds native vision support, and includes improved tool use capabilities.

  • Does o4-mini support image input?

    Yes. Unlike earlier mini reasoning models, it natively processes images, diagrams, and visual content as part of its chain-of-thought reasoning.

  • What is the reasoning_effort parameter?

    It controls how deeply the model reasons per request. Low effort for simple queries saves cost; high effort for hard problems enables thorough deliberation.

  • What context window does o4-mini support?

    200K tokens, providing ample capacity for complex reasoning tasks.

  • How does AI Gateway handle authentication for o4-mini?

    AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.

  • When should I use o3 instead of o4-mini?

    When the hardest problems require maximum reasoning depth and the quality gap between o4-mini and o3 is consequential for your application.

  • What are typical latency characteristics?

    This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.