Google introduced Nano Banana (Gemini 2.5 Flash Image) (internally codenamed nano-banana) on March 20, 2025 as a direct response to developer feedback on the earlier native image generation in Gemini 2.0 Flash. Users valued 2.0 Flash's low latency and ease of use but requested higher image quality and more capable creative control.
Four capabilities define the model. First, character consistency: it can place the same character or object into different environments, generate a product from multiple angles in new settings, or produce consistent brand assets across a series of prompts while preserving subject appearance. Second, prompt-based image editing: you use natural language to perform targeted local edits (blurring a background, removing a stain, altering a pose, colorizing a black-and-white photo) in a single call. Third, native world knowledge: unlike prior image generation models that excelled at aesthetics but lacked semantic grounding, Nano Banana (Gemini 2.5 Flash Image) draws on Gemini's world knowledge to interpret hand-drawn diagrams, answer questions grounded in real-world understanding, and follow complex editing instructions in one step. Fourth, multi-image fusion: the model accepts multiple input images and merges them, enabling product placement into new scenes, room restyling from a reference texture, and image-to-image blending.
All images created or edited with Nano Banana (Gemini 2.5 Flash Image) include an invisible SynthID digital watermark for downstream identification of AI-generated or AI-edited content.