GPT OSS 120B
GPT OSS 120B is OpenAI's open-source 20-billion parameter language model, providing a lightweight yet capable open-weights option suitable for cost-efficient deployment.
import { streamText } from 'ai'
const result = streamText({ model: 'openai/gpt-oss-20b', prompt: 'Why is the sky blue?'})Frequently Asked Questions
How does GPT OSS 120B compare to gpt-oss-120b?
It's more compact (20B vs 120B parameters), making it cheaper to run and easier to self-host, with correspondingly lower capability on complex tasks.
What tasks can GPT OSS 120B handle?
Chat, content generation, summarization, analysis, and other general-purpose language tasks where 20B parameter scale provides sufficient quality.
What context window does GPT OSS 120B support?
131.1K tokens.
How does AI Gateway handle authentication for GPT OSS 120B?
AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.
What are typical latency characteristics?
This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.