New Google AI Updates Are INSANE!
Google announced a sweeping set of AI updates at their Cloud Next event, including new TPU chips, an enterprise agent platform, advanced research agents, and new embedding and training technologies. These updates are positioned as tools that will make AI faster, cheaper, and more accessible for businesses of all sizes. The presenter argues that companies adopting these tools quickly will gain a significant competitive advantage.
Summary
The video covers a series of major AI announcements from Google's Cloud Next event, framed around their potential business impact. The presenter begins with Google's eighth-generation TPU chips — the TPU 8T (for training) and TPU 8I (for inference) — noting that the training chip is nearly three times more powerful than its predecessor. The key takeaway is that cheaper, faster AI hardware reduces costs for end-users across all AI tools, not just those building models.
The next major announcement is the Gemini Enterprise Agent Platform, described as a one-stop shop for deploying AI agents in business workflows. It includes Agent Designer, a no-code tool for building custom agents; long-running agents that can operate autonomously for hours or days; and an Agent Inbox for monitoring all active agents in one dashboard. The platform supports over 200 AI models, including Gemini 3.1 Pro and models from Anthropic, giving users flexibility to choose the best model per task.
Google also launched Deep Research and Deep Research Max — two AI agents designed for automated research. Deep Research prioritizes speed, while Deep Research Max is built for depth, capable of running for hours and producing comprehensive reports with charts and citations. The presenter highlights that Deep Research Max scored 93.3% on a rigorous research benchmark, compared to 66% for the previous version — a significant jump. It can also search private internal files and company data.
Gemini Embedding 2 was made generally available, enabling multimodal search across text, images, video, and audio. The presenter gives examples like searching hours of video footage by typing a natural language query, or allowing e-commerce customers to upload a photo to find matching products without keywords.
Google also open-sourced a format called design.md, which acts as a brand style guide that any AI tool can read and apply consistently — covering colors, fonts, and visual style. Because it's open-source, it works across tools like Claude Code, Cursor, and Copilot, not just Google's ecosystem.
For Google AI Pro and Ultra subscribers, Google expanded access to higher usage limits and the Nano Banana Pro image generation model inside Google AI Studio. Finally, Google DeepMind introduced Decoupled Dial-A-Co, a distributed AI training technology that allows model training to be split across multiple data centers globally, eliminating single points of failure and making AI development faster and more resilient. It was tested on the Gemma 4 model family across four US regions successfully.
Key Insights
- Google's new TPU 8T training chip is claimed to be nearly three times more powerful than its predecessor, which the presenter argues will reduce the cost and time of AI development and ultimately lower prices for end-users of consumer AI tools.
- The Gemini Enterprise Agent Platform includes 'long-running agents' capable of operating autonomously for hours or even days, enabling workflows like overnight lead generation, outreach, and calendar booking without human involvement.
- Deep Research Max scored 93.3% on a rigorous research benchmark compared to 66% for the previous version — a jump the presenter characterizes as massive — and can additionally search a user's own private files and internal company data.
- Gemini Embedding 2 supports multimodal search across text, images, video, and audio in a single model, enabling use cases like natural language search across hours of video footage or photo-based product search in e-commerce without keywords.
- Google DeepMind's Decoupled Dial-A-Co technology decouples AI model training from a single cluster of chips, distributing it across multiple data centers globally so that a failure in one location does not halt the entire training process — validated on a 12-billion parameter Gemma 4 model across four US regions.
Topics
Full transcript available for MurmurCast members
Sign Up to Access