News from the AI & ML world

DeeperML

@developers.googleblog.com //
Google is aggressively advancing AI agent interoperability with its new Agent2Agent (A2A) protocol and development kit. Unveiled at Google Cloud Next '25, the A2A protocol aims to standardize how AI agents communicate, collaborate, and discover each other across different platforms and tasks. This initiative is designed to streamline the exchange of tasks, streaming updates, and sharing of artifacts, fostering a more connected and efficient AI ecosystem. The A2A protocol complements existing efforts by providing a common language for agents, enabling them to seamlessly integrate and normalize various frameworks like LangChain, AutoGen, and Pydantic.

The Agent2Agent protocol introduces the concept of an "Agent Card" (agent.json), which describes an agent's capabilities and how to reach it. Agents communicate through structured messages, indicating task states such as working, input-required, or completed. By establishing this open standard, Google, along with partners like SAP, seeks to enable agents from different vendors to interact, share context, and collaborate effectively. This move represents a significant step beyond simple API integrations, laying the groundwork for interoperability and automation across traditionally disconnected systems.

The development of A2A aligns with Google's broader strategy to solidify its position in the competitive AI landscape, challenging rivals like Microsoft and Amazon. Google is not only introducing new AI chips, such as the Ironwood TPUs, but also updating its Vertex AI platform with Gemini 2.5 models and releasing an agent development kit. This comprehensive approach aims to empower businesses to turn AI potential into real-world impact by facilitating open agent collaboration, model choice, and multimodal intelligence. The collaboration with SAP to enable AI agents to securely interact and collaborate across platforms through A2A exemplifies this commitment to enterprise-ready AI that is open, flexible, and deeply grounded in business context.
Original img attribution: https://storage.googleapis.com/gweb-developer-goog-blog-assets/images/Agent2Agent-Interoperability.2e16d0ba.fill-1200x600.png
ImgSrc: storage.googlea

Share: bluesky twitterx--v2 facebook--v1 threads


References :
  • Search Engine Land: Google AI Mode lets you ask questions with images
  • Search Engine Journal: Google AI mode now understands images, allowing you to upload photos and ask questions about them. AI Mode is rolling out to more people.
  • The Verge: Google is adding multimodal capabilities to its search-centric AI Mode chatbot that enable it to “see†and answer questions about images, as it expands access to AI Mode to “millions more†users.
  • Glenn Gabe: AI Mode expands with multimodal functionality and it's rolling out to millions of more users -> Google AI Mode lets you ask questions with images “With AI Mode’s new multimodal understanding, you can snap a photo or upload an image, ask a question about it and get a rich, comprehensive response with links to dive deeper,†Robby Stein, VP of Product, Google Search wrote."
  • PCMag Middle East ai: Google is also adding AI Mode to the Lens feature of its Google app for Android and iOS. Google is opening up , the web-search chatbot it , to 'millions more Labs users in the US.'
  • www.searchenginejournal.com: Google AI mode now understands images, allowing you to upload photos and ask questions about them. AI Mode is rolling out to more people. The post appeared first on .
  • www.tomsguide.com: Google's Search just got a whole lot more intuitive with the integration of Google Lens in AI Mode.
  • www.zdnet.com: Google Search just got an AI upgrade that you might actually find useful - and it's free
  • www.searchenginejournal.com: Google Maps content moderation now uses Gemini to detect fake reviews and suspicious profile edits.
  • SAP News Center: How SAP and Google Cloud Are Advancing Enterprise AI Through Open Agent Collaboration, Model Choice, and Multimodal Intelligence
  • Ken Yeung: Google Pushes Agent Interoperability With New Dev Kit and Agent2Agent Standard
  • Thomas Roccia :verified:: Google just dropped A2A, a new protocol for agents to talk to each other.
  • AI & Machine Learning: Delivering an application-centric, AI-powered cloud for developers and operators
  • AI News | VentureBeat: Google’s Agent2Agent interoperability protocol aims to standardize agentic communication
  • www.marktechpost.com: Google Introduces Agent2Agent (A2A): A New Open Protocol that Allows AI Agents Securely Collaborate Across Ecosystems Regardless of Framework or Vendor
  • Maginative: Google just Launched Agent2Agent, an Open Protocol for AI agents to Work Directly with Each Other
  • Analytics Vidhya: In today’s fast moving world, many businesses use AI agents to handle their tasks autonomously. However, these agents often operate in isolation, unable to communicate across different systems or vendors.
  • www.analyticsvidhya.com: Agent-to-Agent Protocol: Helping AI Agents Work Together Across Systems
  • developers.googleblog.com: Google's A2A Protocol for Seamless AI Agent Communication
  • TestingCatalog: Google's new Agent2Agent (A2A) protocol enables seamless AI agent collaboration across diverse frameworks, enhancing enterprise productivity and automating complex workflows.
  • bdtechtalks.com: Google's new A2A framework lets different AI agents chat and work together seamlessly, breaking down silos and improving productivity across platforms. The post first appeared on .
  • TheSequence: Google just pushed the boundaries of multi agent communications
Classification:
  • HashTags: #Agent2Agent #AIinteroperability #GoogleAI
  • Company: Google
  • Target: AI developers
  • Product: Agent2Agent
  • Feature: agent interoperability
  • Malware: Agent2Agent
  • Type: AI
  • Severity: Informative