Google has introduced Gemini 2.0, a new AI model designed for the “agentic era,” according to CEO Sundar Pichai. This is considered the company’s most capable AI yet, with advances in multimodality, such as native image and audio output, and native tool use.
Agentic AI refers to systems capable of reasoning, planning, and taking informed actions. Gemini 2.0 is designed to enable a new class of agentic experiences, including Project Astra, a universal AI assistant; and Project Mariner, which explores the future of human-agent interaction.
The new AI agents can understand more about the world, think multiple steps ahead, and take action on behalf of the user, with their supervision.
Gemini 2.0's research assistant uses advanced reasoning and long context capabilities to explore complex topics and compile reports on behalf of the user. Project Astra can converse in multiple languages, use Google Search, Lens, and Maps, and has improved memory.
Gemini 2.0 Flash, the first experimental model from the family, is available in Google AI Studio and Vertex AI. It features enhanced performance, multimodal input, and native support for tools like Google Search, Lens, and Maps.
General availability for Gemini 2.0 is planned for January. The model will be expanded to more Google products in early 2025.
"Over the past year, we have continued to make incredible progress in artificial intelligence," Demis Hassabis, CEO of Google DeepMind and Koray Kavukcuoglu, CTO of Google DeepMind wrote in a blog Wednesday, announcing the new model. "Today, we are releasing the first model in the Gemini 2.0 family of models: an experimental version of Gemini 2.0 Flash. It’s our workhorse model with low latency and enhanced performance at the cutting edge of our technology, at scale."