In December, we kicked off the agent era by releasing an experimental version of Gemini 2.0 Flash. This is a highly efficient workhorse model for developers with low latency and enhanced performance. Earlier this year, we updated Google AI Studio’s 2.0 Flash Thinking Experimental. This combines the speed of Flash with the ability to reason about more complex problems to improve performance.
And last week, we made the updated 2.0 Flash available to all users of the Gemini app on desktop and mobile, allowing everyone to discover new ways to create, interact, and collaborate with Gemini.
Today, we are making the updated Gemini 2.0 Flash generally available via Google AI Studio and Vertex AI’s Gemini API. Developers can now use 2.0 Flash to build production applications.
We’re also releasing an experimental version of Gemini 2.0 Pro, our best model to date in terms of coding performance and complex prompts. This is available in Google AI Studio and Vertex AI, as well as the Gemini app for Gemini Advanced users.
We are releasing Gemini 2.0 Flash-Lite, our most cost-effective new model to date, in public preview for Google AI Studio and Vertex AI.
Finally, the 2.0 Flash Thinking Experimental will be available to Gemini app users in the model dropdown on desktop and mobile.
All of these models will feature multimodal inputs with text output at release, with more modalities expected to be generally available in the coming months. For more information, including pricing details, please visit the Google for Developers blog. Looking ahead, we are working on further updates and enhancements to the Gemini 2.0 family of models.
2.0 Flash: New update for general availability
First introduced at I/O 2024, the Flash series of models is popular with developers as a powerful workhorse model, ideal for large-scale, high-volume, high-frequency tasks, with advanced capabilities for multimodal inference across vast amounts of information in a million-token context window. We are pleased to see the acceptance of this product by the developer community.
2.0 Flash is now generally available to more people across our AI products, with improved performance on key benchmarks, and image generation and text-to-speech coming soon.
Try Gemini 2.0 Flash in the Gemini app or Gemini API in Google AI Studio and Vertex AI. For pricing details, see the Google for Developers blog.
2.0 Pro Experimental: Best model yet for coding performance and complex prompts
As we continue to share early experimental versions of Gemini 2.0 like Gemini-Exp-1206, we’ve gotten great feedback from developers about its strengths and best use cases, including coding.
Today, we’re releasing an experimental version of Gemini 2.0 Pro in response to that feedback. It has the strongest coding performance and ability to better understand and reason about the world’s knowledge and handle complex prompts than any model we’ve released to date. It comes with our largest context window at 2 million tokens, allowing you to comprehensively analyze and understand vast amounts of information, as well as the ability to invoke tools like Google Search and Code Execution.

