Skip to content

GPT-3.5 Turbo Instruct

openai/gpt-3.5-turbo-instruct

GPT-3.5 Turbo Instruct is an instruction-tuned completion model designed for the legacy Completions endpoint, offering a direct prompt-in, text-out format suited to few-shot tasks, templated generation, and workflows that predate the chat message structure.

index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'openai/gpt-3.5-turbo-instruct',
prompt: 'Why is the sky blue?'
})

What To Consider When Choosing a Provider

  • Zero Data Retention

    AI Gateway does not currently support Zero Data Retention for this model. See the documentation for models that support ZDR.

    Authentication

    AI Gateway authenticates requests using an API key or OIDC token. You do not need to manage provider credentials directly.

If you're migrating legacy Completions-based integrations or maintaining backwards-compatible pipelines, note that this model's prompt format differs from chat models. Preserving your existing prompt templates avoids rework.

When to Use GPT-3.5 Turbo Instruct

Best For

  • Legacy completions integrations:

    Existing Completions-endpoint deployments where migrating to the chat message format isn't yet feasible

  • Few-shot text transformation:

    Prompt templates with examples baked in for pattern completion

  • Code completion:

    Inline suggestion features that use a fill-in-the-middle or suffix-based prompt style

  • Structured extraction:

    Well-crafted prompts that reliably produce a specific output format

  • Legacy chatbot pipelines:

    Automation pipelines built before the Chat Completions API became standard

Consider Alternatives When

  • New projects:

    Starting a new project where GPT-3.5 Turbo or GPT-4o mini will generally offer better multi-turn coherence

  • Chat-only features:

    The task requires vision input, function calling, or JSON mode, which are only available on chat-format models

  • Deeper reasoning:

    You need the extended context or reasoning depth of the GPT-4 family

Conclusion

GPT-3.5 Turbo Instruct occupies a specific but important position: it is the only GPT-3.5 Turbo variant designed for the Completions endpoint, making it the right choice when a completion-style prompt contract is required or preferred. Through AI Gateway, teams can continue using this model alongside newer chat-format models under a single authentication layer.

FAQ

GPT-3.5 Turbo uses the Chat Completions endpoint with a messages array. GPT-3.5 Turbo Instruct uses the legacy Completions endpoint with a single prompt string, a fundamental structural difference that affects how you construct requests.

No. Function calling and JSON mode are features of the Chat Completions API. GPT-3.5 Turbo Instruct targets the Completions endpoint and doesn't support these capabilities.

Use it when your prompt structure works best as a single string with few-shot examples inline, when you're maintaining an existing integration, or when a direct prompt-response contract is semantically simpler than a roles-based message list.

Yes. Single-turn code completion and fill-in-the-middle tasks map naturally to the Completions format, and the model's instruction tuning makes it responsive to explicit directives within the prompt.

Authenticate with an AI Gateway API key or OIDC token and route requests to the AI Gateway endpoint specifying this model's slug. No direct OpenAI credentials are required in your application.

Technically you can simulate turns by concatenating prior exchanges into a single prompt string, but the chat-format models handle multi-turn context more naturally and efficiently.

This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.