Skip to content

GPT 5.5

GPT 5.5 is the standard tier of the GPT-5.5 model family, advancing the GPT-5 series with stronger intent understanding, deeper autonomous work, and improvements across coding, research, data analysis, and document creation.

ReasoningTool UseWeb SearchImplicit CachingFile InputVision (Image)
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'openai/gpt-5.5',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • How does GPT 5.5 improve over earlier GPT-5.x models?

    It understands what you're trying to do faster and carries more of the work itself, with measurable gains in code writing and debugging, online research, data analysis, and document creation.

  • What context window does GPT 5.5 support?

    1M tokens, with up to 272K tokens output tokens per request. That is enough for full codebases, long research dossiers, and extended conversation histories in a single call.

  • Which APIs can I use to call GPT 5.5?

    Call GPT 5.5 through the AI SDK, the Chat Completions API, or the Responses API. AI Gateway accepts requests in each format and routes them to the model.

  • Does GPT 5.5 support tool use and web search?

    Yes. Tags include reasoning, tool use, web search, implicit caching, file input, and vision. You can wire up function calling, browsing tools, and file or image inputs through the AI SDK or the Responses API.

  • What does GPT 5.5 cost?

    Standard list pricing is $5 per million input tokens and $30 per million output tokens, with cached input at $0.5. Pricing on this page is sourced from each provider routed through AI Gateway and updates when those providers change list prices.

  • Does GPT 5.5 support zero data retention through AI Gateway?

    Yes, Zero Data Retention is available for this model. Zero Data Retention is offered on a per-provider basis. See https://vercel.com/docs/ai-gateway/capabilities/zdr for details.

  • How does AI Gateway handle authentication for GPT 5.5?

    AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.

  • What are typical latency characteristics?

    This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.