GPT-5 pro
GPT-5 pro is the highest-capability tier in the GPT-5 family, offering extended compute for the most demanding tasks, deeper reasoning, more thorough analysis, and higher-quality output for problems where standard GPT-5 reaches its limits.
import { streamText } from 'ai'
const result = streamText({ model: 'openai/gpt-5-pro', prompt: 'Why is the sky blue?'})Frequently Asked Questions
How does GPT-5 pro differ from standard GPT-5?
It applies more compute per request, enabling deeper reasoning and more thorough analysis. Improvements are most apparent on genuinely hard problems; on simple tasks the difference may be minimal.
When should I use GPT-5 pro versus GPT-5?
Use it selectively for your hardest queries, complex analysis, critical code review, high-stakes content, and tasks where quality directly impacts outcomes. Use standard GPT-5 for routine traffic.
What context window does GPT-5 pro support?
400K tokens, matching the GPT-5 family.
Is GPT-5 pro slower than GPT-5?
It may take longer per request due to the additional compute applied. For latency-sensitive applications, consider routing only specific queries to the pro tier.
How does AI Gateway handle authentication for GPT-5 pro?
AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.
What are typical latency characteristics?
This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.