Skip to content

GPT 5.5 Pro

GPT 5.5 Pro is the higher-capability tier of the GPT-5.5 family, applying more compute per request for deeper analysis on the hardest tasks while keeping the same modalities and APIs as standard GPT-5.5.

ReasoningTool UseImplicit CachingFile InputWeb SearchVision (Image)
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'openai/gpt-5.5-pro',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • How does GPT 5.5 Pro differ from standard GPT-5.5?

    It applies more compute per request, producing deeper analysis on hard problems. Modalities, tool support, and APIs are the same; the difference is reasoning depth and list price.

  • When should I use GPT 5.5 Pro versus standard GPT-5.5?

    Route your hardest queries to GPT 5.5 Pro (critical code review, complex research, high-stakes analysis) and use standard GPT-5.5 for routine traffic. A routing layer keyed on query complexity is the practical pattern.

  • What context window does GPT 5.5 Pro support?

    1M tokens, with up to 128K tokens output tokens per request, matching the rest of the GPT-5.5 family.

  • Which APIs can I use to call GPT 5.5 Pro?

    Call GPT 5.5 Pro through the AI SDK, the Chat Completions API, or the Responses API. AI Gateway accepts requests in each format and routes them to the model.

  • What does GPT 5.5 Pro cost?

    List pricing is $30 per million input tokens and $180 per million output tokens, with cached input at N/A. Compare these rates against standard GPT-5.5 when deciding which tier each request belongs in. Pricing on this page updates when providers change list prices.

  • Does GPT 5.5 Pro support zero data retention through AI Gateway?

    Zero Data Retention is not currently available for this model. Zero Data Retention is offered on a per-provider basis. See https://vercel.com/docs/ai-gateway/capabilities/zdr for details.

  • How does AI Gateway handle authentication for GPT 5.5 Pro?

    AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.

  • What are typical latency characteristics?

    This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic. Additional compute per request typically increases time-to-first-token compared to standard GPT-5.5.