@www.bigdatawire.com
//
References:
www.bigdatawire.com
, www.itpro.com
Google Cloud Next 2025 has wrapped up, showcasing Google Cloud's commitment to simplifying AI adoption for enterprises. The conference highlighted advancements in integrating AI into its data analytics and database services, with a strong focus on flexibility across deployment environments. Key announcements centered around enhancing BigQuery, its flagship database for analytics, and AlloyDB, as well as leveraging the power of Vertex AI and the new Ironwood TPU-powered AI Hypercomputer for demanding AI workloads. Google Cloud emphasized making AI more accessible to businesses of all sizes.
New AI-powered tools and features were unveiled, designed to streamline data management, enhance analytics capabilities, and enable AI-driven workflows. Google Cloud is embedding pre-built AI agents into its own software services, including BigQuery, to assist with tasks such as data engineering, data science, building data pipelines, data transformation, and anomaly detection. These agents, powered by Gemini, Google’s flagship foundation model, will provide suggestions and assistance to data analysts, scientists, and engineers, transforming the way they work with data. The Agent Development Kit (ADK) supports the creation of interconnected AI agents that can communicate regardless of their vendor or framework. These innovations mark a significant push towards agentic AI, with Google aiming to automate traditionally automation-resistant workflows. New generative AI features within AlloyDB for PostgreSQL will help data professionals build agent-ready data views without compromising privacy and security requirements. Additionally, with Gen AI Toolbox for Databases, built in partnership with LangChain, Google intends to connect GenAI applications to several Google data sources efficiently. By incorporating the Model Context Protocol (MCP) into its new Agent Development Kit, Google is helping developers integrate external data sources into their agentic solutions. Recommended read:
References :
@cloud.google.com
//
Google Cloud Next 2025 showcased a new direction for the company, focusing on an application-centric, AI-powered cloud environment for developers and operators. The conference highlighted Google's commitment to simplifying AI adoption for enterprises, emphasizing flexibility across deployments. Key announcements included AI assistance features within Gemini Code Assist and Gemini Cloud Assist, designed to streamline the application development lifecycle. These tools introduce new agents capable of handling complex workflows directly within the IDE, aiming to offload development tasks and improve overall productivity.
Google Cloud is putting applications at the center of its cloud experience, abstracting away traditional infrastructure complexities. This application-centric approach enables developers to design, observe, secure, and optimize at the application level, rather than at the infrastructure level. To support this shift, Google introduced the Application Design Center, a service designed to streamline the design, deployment, and evolution of cloud applications. The Application Design Center provides a visual, canvas-style approach to designing and modifying application templates. It also allows users to configure application templates for deployment, view infrastructure as code in-line, and collaborate with teammates on designs. The event also highlighted Cloud Hub, a service that unifies visibility and control over applications, providing insights into deployments, health, resource optimization, and support cases. Gemini Code Assist and Cloud Assist aim to accelerate application development and streamline cloud operations by offering agents that translate natural language requests into multi-step solutions and tools for connecting Code Assist to external services. Google's vision is to make the entire application journey smarter and more productive through AI-driven assistance and simplified cloud management. Recommended read:
References :
staff@insideAI News
//
Google Cloud has unveiled its seventh-generation Tensor Processing Unit (TPU), named Ironwood, at the recent Google Cloud Next 2025 conference. This new custom AI accelerator is specifically designed for inference workloads, marking a shift in Google's AI chip development strategy. Ironwood aims to meet the growing demands of "thinking models" like Gemini 2.5, addressing the increasing shift from model training to inference observed across the industry. According to Amin Vahdat, Google's Vice President and General Manager of ML, Systems, and Cloud AI, the aim is to enter the "age of inference" where AI agents proactively retrieve and generate data for insights.
Ironwood's technical specifications are impressive, offering substantial computational power and efficiency. When scaled to a pod of 9,216 chips, it can deliver 42.5 exaflops of compute, surpassing the world's fastest supercomputer, El Capitan, by more than 24 times. Each individual Ironwood chip boasts a peak compute of 4,614 teraflops. To manage the communication demands of modern AI, each Ironwood setup features Inter-Chip Interconnect (ICI) networking spanning nearly 10 MW and each chip is equipped with 192GB of High Bandwidth Memory (HBM) and memory bandwidth that reaches 7.2 terabits per second. This focus on inference is a response to the evolving AI landscape where proactive AI agents are becoming more prevalent. Ironwood is engineered to minimize data movement and latency on-chip while executing massive tensor manipulations, crucial for handling large language models and advanced reasoning tasks. Google emphasizes that Ironwood offers twice the performance per watt compared to its predecessor, Trillium, and is nearly 30 times more power efficient than Google’s first Cloud TPU from 2018, addressing the critical need for power efficiency in modern data centers. Recommended read:
References :
@www.analyticsvidhya.com
//
Google Cloud Next '25 saw a major push into agentic AI with the unveiling of several key technologies and initiatives aimed at fostering the development and interoperability of AI agents. Google announced the Agent Development Kit (ADK), an open-source framework designed to simplify the creation and management of AI agents. The ADK, written in Python, allows developers to build both simple and complex multi-agent systems. Complementing the ADK is Agent Garden, a collection of pre-built agent patterns and components to accelerate development. Additionally, Google introduced Agent Engine, a fully managed runtime in Vertex AI, enabling secure and reliable deployment of custom agents at a global scale.
Google is also addressing the challenge of AI agent interoperability with the introduction of the Agent2Agent (A2A) protocol. A2A is an open standard intended to provide a common language for AI agents to communicate, regardless of the frameworks or vendors used to build them. This protocol allows agents to collaborate and share information securely, streamlining workflows and reducing integration costs. The A2A initiative has garnered support from over 50 industry leaders, including Atlassian, Box, Cohere, Intuit, and Salesforce, signaling a collaborative effort to advance multi-agent systems. These advancements are integrated within Vertex AI, Google's comprehensive platform for managing models, data, and agents. Enhancements to Vertex AI include supporting Model Context Protocol (MCP) to ensure secure data connections for agents. In addition to software advancements, Google unveiled its seventh-generation Tensor Processing Unit (TPU), named Ironwood, designed to optimize AI inferencing. Ironwood offers significantly increased compute capacity and high-bandwidth memory, further empowering AI applications within the Google Cloud ecosystem. Recommended read:
References :
staff@insideAI News
//
Google has unveiled its seventh-generation Tensor Processing Unit (TPU), named "Ironwood," marking a significant advancement in AI accelerator technology. Designed specifically for AI inference workloads, Ironwood is Google's most performant and scalable custom AI accelerator to date. This launch is part of Google Cloud's strategy to lead in supplying AI models, applications, and infrastructure, capitalizing on its own substantial AI needs to drive homegrown infrastructure development. Google Cloud is positioning itself for the "agentic AI era," with Ironwood playing a pivotal role in enabling multiple AI systems to work together across platforms.
Google's Ironwood TPU comes in configurations of up to 9,216 liquid-cooled chips interconnected via Inter-Chip Interconnect (ICI) networking, spanning nearly 10 MW. The architecture of the Ironwood TPU is designed to optimize hardware and software for AI workloads, allowing developers to leverage Google's Pathways software stack to harness tens of thousands of Ironwood TPUs. The chip design philosophy shift is from models that provide real-time information for interpretation, to models that proactively generate insights and interpretations, which Google calls the "age of inference." Ironwood is designed to manage the computational and communication demands of "thinking models" like large language models and advanced reasoning tasks. A configuration with 9,216 chips per pod can support more than 24 times the compute power of the world’s no. 1 supercomputer, El Capitan. NVIDIA is collaborating with Google Cloud to bring agentic AI to enterprises seeking to locally harness the Google Gemini family of AI models using the NVIDIA Blackwell HGX and DGX platforms and NVIDIA Confidential Computing for data safety. This collaboration enables enterprises to innovate securely while maintaining data privacy. Recommended read:
References :
|
BenchmarksBlogsResearch Tools |