o3 Pro
o3 Pro applies maximum reasoning compute to the o3 architecture, delivering the deepest chain-of-thought analysis available for the most demanding mathematical, scientific, and analytical problems.
import { streamText } from 'ai'
const result = streamText({ model: 'openai/o3-pro', prompt: 'Why is the sky blue?'})Frequently Asked Questions
How does o3 Pro differ from standard o3?
It applies more compute per request, enabling deeper and more exhaustive chain-of-thought reasoning. The improvement is most significant on the hardest problems.
When is the premium cost justified?
For frontier mathematical proofs, critical scientific analysis, security audits, and other problems where the deepest possible reasoning produces meaningfully better outcomes.
What context window does o3 Pro support?
200K tokens, matching the o3 family.
Is o3 Pro slower than o3?
Yes. The additional reasoning compute increases response time. Reserve it for problems where depth justifies the wait.
How does AI Gateway handle authentication for o3 Pro?
AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.
What are typical latency characteristics?
This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.