Skip to content

GPT-4o mini

GPT-4o mini is OpenAI's cost-efficient multimodal model, priced at $0.15 per million input tokens, at reduced cost compared to GPT-3.5 Turbo, while outperforming GPT-4 on chat preference benchmarks and supporting vision and function calling.

File InputTool UseVision (Image)Implicit Caching
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'openai/gpt-4o-mini',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • How does GPT-4o mini compare to GPT-3.5 Turbo on price?

    Pricing appears on this page and updates as providers adjust their rates. AI Gateway routes traffic through the configured provider.

  • Does GPT-4o mini support image input?

    Yes. It supports vision alongside text, enabling image analysis, document processing, and visual classification at the small-model cost tier.

  • What benchmark scores did GPT-4o mini achieve?

    82.0% on MMLU, outperforming comparable small models and topping GPT-4 on the LMSYS Chatbot Arena chat preference leaderboard at launch.

  • Is GPT-4o mini suitable for function calling and tool use?

    Yes. Function calling is supported, and OpenAI highlighted agentic pipelines that call external APIs as one of the key intended use cases.

  • What is the context window for GPT-4o mini?

    128K tokens, providing ample space for conversation histories, long codebases, and extended document processing.

  • How does gpt-4o-mini (the alias) differ from gpt-4o-mini-2024-07-18?

    The alias gpt-4o-mini points to the current recommended version and may be updated. The dated snapshot gpt-4o-mini-2024-07-18 is pinned to the specific July 18, 2024 release.

  • What are typical latency characteristics?

    This page shows live throughput and time-to-first-token metrics measured across real AI Gateway traffic.