In December, we first introduced Gemini 2.0 Flash’s native image output to our trusted testers. Today, we’re making it available for developer experimentation in all regions currently supported by Google AI Studio. You can test this new feature using Google AI Studio’s Gemini 2.0 Flash experimental version (gemini-2.0-flash-exp) and the Gemini API.
Gemini 2.0 Flash combines multimodal input, enhanced inference, and natural language understanding to create images.
2.0 Here are some examples of how Flash’s multimodal output shines.
1. Text and images together
Use Gemini 2.0 Flash to tell your story, illustrate it with pictures, and keep your characters and settings consistent. When you give feedback, the model retells the story or changes its drawing style.
Sorry, your browser does not support playing this video
Generating stories and illustrations in Google AI Studio
2. Conversational image editing
Gemini 2.0 Flash helps you edit images with natural language dialogue over and over again, perfect for iterating on the perfect image or exploring different ideas together.
Sorry, your browser does not support playing this video
Multi-turn conversation image editing that maintains the context of the entire conversation in Google AI Studio
3. World understanding
Unlike many other image generation models, Gemini 2.0 Flash leverages world knowledge and advanced reasoning to create the right images. This is perfect for creating realistic, detailed images, such as those illustrating recipes. As with all language models, we strive for accuracy, but our knowledge is broad and general, not absolute or complete.
Sorry, your browser does not support playing this video
Interleaved text and image output for recipes in Google AI Studio
4. Rendering text
Most image generation models struggle to accurately render long text sequences, often resulting in poorly formatted, illegible, or misspelled characters. Internal benchmarks show that 2.0 Flash has more powerful rendering compared to leading competing models, making it ideal for creating ads, social posts, and even invitations.
Sorry, your browser does not support playing this video
Image output by rendering long text in Google AI Studio
Start creating images with Gemini today
Get started with Gemini 2.0 Flash via the Gemini API. For more information on image generation, please see the documentation.
from google import genai from google.genai import tables client = genai.Client(api_key=”GEMINI_API_KEY”) response = client.models.generate_content( model=”gemini-2.0-flash-exp”, content=( “Generate a story about cute baby turtles in 3D digital art style.” “Generate images for each scene.” ), config=types.GenerateContentConfig( response_modalities=(“text”, “image”) ), )
python
Whether you’re building an AI agent, developing an app with beautiful visuals like illustrated interactive stories, or brainstorming visual ideas during conversations, Gemini 2.0 Flash lets you add text and image generation in a single model. We look forward to seeing what developers create with native image output. Your feedback will help us quickly finalize a production-ready version.

