Gemini 2.5 Flash: Our most cost-efficient thinking model
We’re also announcing Gemini 2.5 Flash, our low latency and most cost efficient thinking model. With 2.5 Flash, you can control how much the model reasons, and balance performance with your budget.
Gemini 2.5 Flash is coming soon in Google AI Studio, Vertex AI and in the Gemini app. We’ll be sharing more details on the model and its performance soon.
Products and platforms powered by world-class AI
Our goal is to always bring our latest AI advances into the fourth layer of our stack: products and platforms. Today all 15 of our half-billion user products — including seven with 2 billion users — are using our Gemini models. AI deployed at this scale requires world-class inference, which enterprises can benefit from to build their own AI-powered applications.
Gemini is also helping us create net-new products and experiences. NotebookLM is one example, used by 100,000 businesses. It uses long context, multimodality and our latest thinking models to show information in powerful ways. Veo 2 is a leading video generation model. Major film studios, entertainment companies, as well as the top advertising agencies in the world are using it to bring their stories to life.
Getting these advancements into the hands of both consumers and enterprises is something we’re really focused on. This is why we are able to innovate at the cutting edge, and push the boundaries of what’s possible, for us — and for you. The result: better, faster, and more innovation for everyone.