o4-mini
o4-mini advances OpenAI's compact reasoning model line with stronger performance and greater efficiency than o3-mini, adding native tool use and image reasoning.
import { streamText } from 'ai'
const result = streamText({ model: 'openai/o4-mini', prompt: 'Why is the sky blue?'})Frequently Asked Questions
How does o4-mini improve over o3-mini?
It delivers stronger reasoning performance with greater efficiency, adds native vision support, and includes improved tool use capabilities.
Does o4-mini support image input?
Yes. Unlike earlier mini reasoning models, it natively processes images, diagrams, and visual content as part of its chain-of-thought reasoning.
What is the
reasoning_effortparameter?It controls how deeply the model reasons per request. Low effort for simple queries saves cost; high effort for hard problems enables thorough deliberation.
What context window does o4-mini support?
200K tokens, providing ample capacity for complex reasoning tasks.
How does AI Gateway handle authentication for o4-mini?
AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.
When should I use o3 instead of o4-mini?
When the hardest problems require maximum reasoning depth and the quality gap between o4-mini and o3 is consequential for your application.
What are typical latency characteristics?
This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.