GPT 5.4 Mini
GPT 5.4 Mini is the cost-efficient member of the GPT-5.4 family, delivering strong performance in code generation, tool orchestration, and multi-step browser interactions at a price point designed for agentic production workloads.
import { streamText } from 'ai'
const result = streamText({ model: 'openai/gpt-5.4-mini', prompt: 'Why is the sky blue?'})Frequently Asked Questions
How does GPT 5.4 Mini compare to GPT-5 mini?
It handles code generation, tool orchestration, and multi-step browser interactions more reliably. It also supports verbosity and reasoning level parameters for tunable output.
What context window does GPT 5.4 Mini support?
400K tokens, supporting extended inputs for agentic workflows.
What are the verbosity and reasoning level parameters?
They give you control over response detail and how much the model reasons before answering, letting you tune the cost-quality tradeoff per request.
Is GPT 5.4 Mini suitable for sub-agent workflows?
Yes. It's built for sub-agent architectures where multiple smaller models coordinate on parts of a larger task.
When should I use GPT-5.4 Nano instead?
When cost is the dominant concern and you're running high-volume parallel calls. GPT-5.4 Nano performs close to mini in evaluations at a lower price point.
How does AI Gateway handle authentication for GPT 5.4 Mini?
AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.
What are typical latency characteristics?
This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.