Gemini 2.0 Flash
Gemini 2.0 Flash is Google's workhorse model for the agentic era. It delivers low-latency multimodal output, including natively generated images and steerable text-to-speech (TTS) audio, alongside native tool use and a Multimodal Live API for real-time streaming.
import { streamText } from 'ai'
const result = streamText({ model: 'google/gemini-2.0-flash', prompt: 'Why is the sky blue?'})Playground
Try out Gemini 2.0 Flash by Google. Usage is billed to your team at API rates. Free users (those who haven't made a payment) get $5 of credits every 30 days.
About Gemini 2.0 Flash
Google released Gemini 2.0 Flash on December 11, 2024 as the first model in the Gemini 2.0 generation, optimized for high-volume, high-frequency tasks at scale. It outperforms Gemini 1.5 Pro on key benchmarks at twice the speed, a significant capability leap over the previous Flash generation.
Gemini 2.0 Flash stands out through native multimodal output. Beyond accepting text, images, video, and audio as input, it produces natively generated images mixed with text and steerable text-to-speech (TTS) multilingual audio. This eliminates the need for separate image-generation or speech-synthesis calls, enabling tighter integration within a single request.
Google simultaneously released the Multimodal Live API alongside 2.0 Flash for real-time and interactive applications. This API adds streaming real-time audio and video input, combined tool use, and the low-latency response characteristics conversational agents and live-session experiences need. Gemini 2.0 Flash also supports native tool use including Google Search, code execution, and user-defined functions for multi-step agentic workflows.
The context window of 1.0M tokens handles tasks that require reasoning over large codebases, lengthy documents, or extended conversation histories in a single pass.
Providers
Route requests across multiple providers. Copy a provider slug to set your preference. Visit the docs for more info. Using a provider means you agree to their terms, listed under Legal.
| Provider |
|---|
P50 throughput on live AI Gateway traffic, in tokens per second (TPS). Visit the docs for more info.
P50 time to first token (TTFT) on live AI Gateway traffic, in milliseconds. View the docs for more info.
Direct request success rate on AI Gateway and per-provider. Visit the docs for more info.
More models by Google
| Model |
|---|
What To Consider When Choosing a Provider
- Configuration: When selecting a provider variant, consider whether your application requires the Multimodal Live API for real-time audio and video streaming, as that capability may vary across provider endpoints.
- Zero Data Retention: AI Gateway supports Zero Data Retention for this model via direct gateway requests (BYOK is not included). To configure this, check the documentation.
- Authentication: AI Gateway authenticates requests using an API key or OIDC token. You do not need to manage provider credentials directly.
When to Use Gemini 2.0 Flash
Best For
- High-frequency production workloads: You need strong benchmark performance at low latency and competitive cost
- Agentic applications: Require compositional function-calling, native Google Search tool use, and multi-step planning
- Applications generating mixed media: Benefit from native image and TTS audio output within a single model call rather than chained services
- Real-time interactive experiences: Using the Multimodal Live API for streaming audio/video input with sub-second response loops
- Long-context analysis: Processing up to 1.0M tokens of text, video, images, audio, or code in a single context window
Consider Alternatives When
- Deep extended reasoning: Your task demands deliberate chain-of-thought thinking, which is more central to the 2.5 generation of models
- Lowest cost per token: For very simple classification or captioning tasks, a lighter model like Gemini 2.0 Flash Lite may be more appropriate
- Dedicated embedding workloads: Your application requires only text embeddings and semantic retrieval, where a dedicated embedding model is a better architectural fit
- Strict budget constraints: Per-request cost is above capability and quality parity with 1.5 Flash is sufficient for your use case
Conclusion
Gemini 2.0 Flash marks a generational upgrade in what a workhorse model can do. It brings native multimodal output (images and audio) and real-time streaming into a single, high-throughput package. Teams building agentic pipelines, interactive media applications, or large-scale inference workloads get a model designed from the ground up for production AI in the agentic era.
Frequently Asked Questions
What makes Gemini 2.0 Flash different from 1.5 Flash?
Gemini 2.0 Flash adds native multimodal output (images and steerable TTS audio), native tool use (Google Search, code execution, user-defined functions), and the Multimodal Live API for real-time streaming, while maintaining similar latency to 1.5 Flash and outperforming 1.5 Pro on key benchmarks.
What is the Multimodal Live API and does AI Gateway support it?
The Multimodal Live API is a streaming interface released alongside 2.0 Flash. It supports real-time audio and video input with combined tool use. Check the AI Gateway documentation and your provider in vertex, google for current Live API support.
Can Gemini 2.0 Flash generate images and audio in the same response as text?
Yes. Gemini 2.0 Flash produces natively generated images and steerable text-to-speech audio alongside text in a single response, without requiring separate generation calls.
How does the context window of 1.0M tokens affect prompt construction?
With 1.0M tokens, you can pass entire codebases, long PDF documents, hours of transcripts, or extended conversation histories in a single context, eliminating the need to chunk or summarize inputs for most practical workloads.
What native tools can Gemini 2.0 Flash call?
Gemini 2.0 Flash supports Google Search, code execution, and third-party user-defined functions natively, enabling it to fetch live information, run and test code, and call external APIs within a single inference pass.
Is Gemini 2.0 Flash suitable for building Project Astra-style universal assistant experiences?
Yes. Google uses Gemini 2.0 Flash as the foundation for Project Astra prototypes, which rely on its multimodal reasoning, native tool use, low latency, and multi-language conversational capabilities.
How does Zero Data Retention work with this model through AI Gateway?
Yes, Zero Data Retention is available for this model. ZDR on AI Gateway applies to direct gateway requests; BYOK flows aren't covered. See https://vercel.com/docs/ai-gateway/capabilities/zdr for details.
What safety measures are built into Gemini 2.0 Flash?
Gemini 2.0 Flash uses reinforcement learning to critique its own responses and improve handling of sensitive prompts. Google also runs automated red teaming to assess risks including indirect prompt injection attacks.