Skip to content

GPT-4 Turbo

GPT-4 Turbo launched at OpenAI DevDay 2023 with a context window of 128K tokens, built-in vision, JSON mode, and a knowledge cutoff of April 2023, all at reduced input prices compared to the original GPT-4.

Tool UseVision (Image)
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'openai/gpt-4-turbo',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • What made GPT-4 Turbo's context window significant?

    At 128K tokens it holds roughly 300 pages of text, enabling workflows like full-codebase review, long legal document analysis, and extended multi-session conversation history without chunking.

  • How does JSON mode in GPT-4 Turbo work?

    Set response_format to { type: "json_object" } and the model constrains itself to produce valid JSON. This differs from the stricter JSON Schema-based Structured Outputs introduced later with gpt-4o-2024-08-06.

  • Does GPT-4 Turbo support image inputs?

    Yes. You can pass image URLs or base64-encoded images alongside text in the messages array. The model can analyze photographs, diagrams, screenshots, and documents with embedded figures.

  • What is GPT-4 Turbo's knowledge cutoff?

    April 2023. This was an update from earlier GPT-4 models and makes the model aware of events from the first half of 2023.

  • How does GPT-4 Turbo pricing compare to the original GPT-4?

    Rates are listed on this page. They reflect the providers routing through AI Gateway and shift when providers update their pricing.

  • Can I route GPT-4 Turbo requests through AI Gateway without storing provider API keys?

    Yes. AI Gateway handles authentication using its own API key or OIDC token system, so you don't need to embed OpenAI credentials in your deployment environment.

  • What are typical latency characteristics?

    This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.