Why it matters: Google’s latest AI model represents a strategic pivot toward autonomous AI agents, potentially reshaping how users interact with technology across Google’s ecosystem.
The big picture: Gemini 2.0 integrates advanced multimodal capabilities and native tool usage, positioning Google to compete directly with OpenAI’s ChatGPT and Microsoft’s Copilot.
By the numbers:
- Gemini 2.0 Flash operates at double the speed of its predecessor
- Available immediately through Google Search and the Gemini app
- Accessible to all developers via Vertex AI and Google AI Studio
Key features include:
- Enhanced multimodal processing across text, images, audio and video
- Deep Research Tool for synthesizing data with source citations
- Native integration with Google Search and code execution
- Multimodal Live API supporting real-time video/audio applications
What’s next: Google plans broader rollouts in 2025, including new output modalities enabling AI responses with text, images, and audio through a single API call. All audio and image outputs will include SynthID invisible watermarks to combat deepfakes.
What they’re saying: Google CEO Sundar Pichai calls this the “new agentic era” in technology.