GPT-4o mini
GPT-4o mini is OpenAI's cost-efficient multimodal model, priced at $0.15 per million input tokens, at reduced cost compared to GPT-3.5 Turbo, while outperforming GPT-4 on chat preference benchmarks and supporting vision and function calling.
import { streamText } from 'ai'
const result = streamText({ model: 'openai/gpt-4o-mini', prompt: 'Why is the sky blue?'})Frequently Asked Questions
How does GPT-4o mini compare to GPT-3.5 Turbo on price?
Pricing appears on this page and updates as providers adjust their rates. AI Gateway routes traffic through the configured provider.
Does GPT-4o mini support image input?
Yes. It supports vision alongside text, enabling image analysis, document processing, and visual classification at the small-model cost tier.
What benchmark scores did GPT-4o mini achieve?
82.0% on MMLU, outperforming comparable small models and topping GPT-4 on the LMSYS Chatbot Arena chat preference leaderboard at launch.
Is GPT-4o mini suitable for function calling and tool use?
Yes. Function calling is supported, and OpenAI highlighted agentic pipelines that call external APIs as one of the key intended use cases.
What is the context window for GPT-4o mini?
128K tokens, providing ample space for conversation histories, long codebases, and extended document processing.
How does gpt-4o-mini (the alias) differ from gpt-4o-mini-2024-07-18?
The alias
gpt-4o-minipoints to the current recommended version and may be updated. The dated snapshotgpt-4o-mini-2024-07-18is pinned to the specific July 18, 2024 release.What are typical latency characteristics?
This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.