Skip to content

GPT OSS 120B

GPT OSS 120B is OpenAI's open-source 20-billion parameter language model, providing a lightweight yet capable open-weights option suitable for cost-efficient deployment.

ReasoningTool Use
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'openai/gpt-oss-20b',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • How does GPT OSS 120B compare to gpt-oss-120b?

    It's more compact (20B vs 120B parameters), making it cheaper to run and easier to self-host, with correspondingly lower capability on complex tasks.

  • What tasks can GPT OSS 120B handle?

    Chat, content generation, summarization, analysis, and other general-purpose language tasks where 20B parameter scale provides sufficient quality.

  • What context window does GPT OSS 120B support?

    131.1K tokens.

  • How does AI Gateway handle authentication for GPT OSS 120B?

    AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.

  • What are typical latency characteristics?

    This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.