Google’s Agent2Agent (A2A) Protocol v0.3 is out — and it’s the upgrade that turns AI agent interoperability from a proof-of-concept into production infrastructure. With gRPC transport, cryptographically signed agent cards, and an expanded Python SDK, A2A v0.3 is the release that serious multi-agent system builders have been waiting for.
The Agent2Agent (A2A) Protocol v0.3 is the upgrade that finally makes multi-agent AI communication enterprise-grade. Released in early 2026 under Linux Foundation governance, A2A v0.3 adds three capabilities that close the gap between proof-of-concept agent demos and production deployments: gRPC transport support, cryptographically signed agent cards, and a substantially expanded Python SDK. With 50+ technology partners including Atlassian, Salesforce, SAP, Workday, and PayPal, and major consulting firms from Accenture to Deloitte and McKinsey building implementation practices around it, A2A is establishing itself as the HTTP of the AI agent layer — the foundational protocol that all multi-agent systems will be expected to speak.
Why Agent Communication Standards Matter More Than the Models
In 2024 and early 2025, the AI industry’s attention was almost entirely focused on model capabilities: which frontier model scored highest on benchmarks, which had the largest context window, which produced the most coherent code. By Q1 2026, the conversation has shifted. The models are capable enough. The bottleneck is now infrastructure: how do you make independent AI agents from different vendors, built on different frameworks, running in different environments, actually work together as a coherent system?
The Model Context Protocol (MCP) solved the agent-to-tool connectivity problem: a standardized way for AI agents to discover and invoke external tools, APIs, and data sources. A2A solves the complementary problem: standardized agent-to-agent communication. Where MCP is the protocol that lets an agent call a tool, A2A is the protocol that lets one agent hand off a task to another agent, monitor its progress, stream back results, and handle authentication across organizational boundaries. According to our analysis of the 2026 agentic AI landscape, the combination of MCP for tool access and A2A for agent coordination provides the complete protocol stack for building production multi-agent systems that can span teams, vendors, and enterprise application estates.
What Is A2A and How It Works
A2A is an open HTTP-based protocol with a straightforward conceptual model. Every A2A deployment has two roles:
- A2A Client: The agent (or application) that delegates tasks — it discovers remote agents, sends task requests, and consumes results.
- A2A Server: The remote agent that receives task requests, processes them, and returns results. Every A2A server exposes an HTTP endpoint that’s compatible with the A2A specification.
Discovery happens through Agent Cards: JSON documents hosted at a standardized .well-known/agent.json endpoint on every A2A server. An Agent Card describes what the agent can do, what inputs it accepts, how to authenticate, and what response formats it supports. This is how an orchestrating agent finds a specialist agent to delegate work to — the same way a web browser uses DNS to discover a server and HTTP to communicate with it.
Once an A2A client finds a server via its Agent Card, interaction follows a task-based model: the client sends a task (a JSON object describing the work to do), the server processes it and streams back status updates and results, and the client consumes the stream. A2A supports single-turn request/response and multi-turn streaming conversations, making it suitable for both simple tool-like delegation and complex interactive workflows between agents.
What Changed in v0.3: Three Capabilities That Change Everything
1. gRPC Transport Support
A2A’s v0.2 specification supported only JSON-RPC over HTTP. v0.3 adds gRPC as a first-class transport binding, and this matters significantly for production deployments. gRPC uses Protocol Buffers (protobuf) for serialization, which delivers roughly 3–10x faster serialization and 2–3x smaller payload size compared to JSON for the same structured data. For high-throughput multi-agent systems — an orchestrator routing hundreds of sub-tasks per second to specialist agents, or real-time streaming of intermediate results across a pipeline — gRPC transport eliminates a latency and bandwidth bottleneck that JSON-over-HTTP cannot. The Python SDK provides protocol binding selection at initialization time, so switching a deployed agent from HTTP to gRPC requires a single configuration parameter change, not a code rewrite.
2. Signed Agent Cards
The most consequential security improvement in v0.3 is cryptographically signed Agent Cards. In v0.2, Agent Cards were plain JSON documents served at the .well-known/agent.json endpoint. Nothing prevented a network adversary from serving a malicious Agent Card that claimed to be a legitimate enterprise agent. In v0.3, Agent Cards can be digitally signed using JSON Web Signature (JWS), with card content canonicalized before signing using the JSON Canonicalization Scheme (JCS) defined in RFC 8785. An A2A client can verify a received Agent Card against a trusted public key, guaranteeing authenticity and integrity. For enterprise deployments where AI agents handle sensitive data, make financial transactions, or act with delegated authority, this is not optional — it’s the difference between an audit-compliant system and one that legal and security teams will not approve for production. The signing workflow integrates directly with existing enterprise PKI infrastructure, using the same key management systems that organizations already use for code signing and TLS certificates.
3. Expanded Python SDK
The v0.3 Python SDK ships via pip install a2a-sdk with a modular extras system that keeps the base package lightweight while making production capabilities available on demand:
# Install base SDK
pip install a2a-sdk
# Install with all production extras
pip install a2a-sdk[grpc,signing,telemetry,encryption]
# Or select specific capabilities
pip install a2a-sdk[grpc] # gRPC transport
pip install a2a-sdk[signing] # Agent card signing
pip install a2a-sdk[telemetry] # OpenTelemetry integration
pip install a2a-sdk[encryption] # Payload encryptionThe SDK provides server and client classes that abstract the protocol details, so developers can focus on agent logic rather than protocol implementation. A minimal A2A server exposing a text summarization agent:
from a2a.server import A2AServer, AgentCard, AgentSkill
from a2a.types import Task, TaskResult
skill = AgentSkill(
id="summarize",
name="Text Summarizer",
description="Summarizes long documents into concise paragraphs",
input_schema={"type": "object", "properties": {"text": {"type": "string"}}},
)
agent_card = AgentCard(
name="SummaryAgent",
description="Production text summarization agent",
url="https://agents.yourcompany.com/summary",
skills=[skill],
version="1.0.0",
)
server = A2AServer(agent_card=agent_card)
@server.skill_handler("summarize")
async def handle_summarize(task: Task) -> TaskResult:
text = task.params["text"]
summary = await your_summarization_logic(text)
return TaskResult(result={"summary": summary})
server.run(host="0.0.0.0", port=8080)The server automatically serves the Agent Card at /.well-known/agent.json, handles task routing, manages streaming connections, and exposes the OpenTelemetry spans that enterprise monitoring systems expect. On the client side, discovering and calling any A2A-compatible agent requires fewer than ten lines of code, regardless of which framework built the server.
Enterprise Deployment Paths
Google Cloud provides three official deployment options for A2A agents, each targeting a different operational requirement:
Agent Engine is the managed, fully-abstracted option. Deploy your A2A server as a managed agent and Google handles scaling, health monitoring, availability, and the A2A protocol infrastructure. No Kubernetes knowledge required, no container management, no autoscaling configuration. Agent Engine is the right choice for teams that want to expose agent capabilities to other systems without taking on infrastructure operations. It is also the path to listing agents in the AI Agent Marketplace (more on this below).
Cloud Run gives you serverless container infrastructure with full protocol flexibility. Your A2A server runs in a standard Docker container, scales to zero when idle, and scales horizontally under load. This is the right choice if you need control over the container environment, want to run custom middleware, or need to support both A2A and direct API access from the same service. Cloud Run’s per-request billing model makes it cost-efficient for agents that handle bursty or unpredictable traffic patterns.
Google Kubernetes Engine (GKE) provides maximum control for organizations with specific compliance, networking, or resource management requirements. Running A2A agents on GKE gives you cluster-level isolation, custom networking policies, the ability to co-locate agents with existing microservices, and full control over the runtime environment. This is the enterprise-in-house option for organizations that manage their own Kubernetes infrastructure and cannot use managed services for data residency or security reasons.
A2A vs. MCP: Complementary, Not Competing
A common misconception when developers first encounter A2A is that it competes with MCP. It does not. The protocols operate at different layers of the agent stack and are designed to be used together.
MCP (Model Context Protocol, the standard that reached 97 million monthly downloads in early 2026) governs how a single AI agent connects to external tools: databases, APIs, file systems, search engines, code execution environments. When Claude reads a file, queries a database, or runs a web search, it’s doing so via an MCP server. MCP is the agent-to-tool interface.
A2A governs how AI agents talk to each other: how an orchestrator delegates a task to a specialist agent, how that agent streams progress back, how authentication is handled across organizational boundaries, and how agents advertise their capabilities for discovery. A2A is the agent-to-agent interface.
In a realistic production multi-agent system, both protocols are active simultaneously. An orchestrating agent receives a task, uses A2A to delegate sub-tasks to three specialist agents (a research agent, a writing agent, a fact-checking agent), and each of those agents uses MCP to connect to the tools they need (web search, knowledge bases, structured data sources). MCP and A2A form the complete connectivity layer — the TCP/IP and HTTP of the AI agent internet.
The AI Agent Marketplace: Monetizing Specialist Agents
One of the most commercially significant features in the v0.3 ecosystem is the AI Agent Marketplace on Google Cloud, which allows organizations to sell access to their A2A-compliant agents directly to Google Cloud customers. An ISV that builds a specialist legal contract analysis agent, a pharmaceutical drug interaction checker, or a supply chain optimization agent can publish it to the marketplace and charge for access — with billing, discovery, and integration handled by Google Cloud’s infrastructure.
This creates a fundamentally new software distribution model. Instead of building a full SaaS application with its own frontend, authentication, and billing pipeline, a team with domain expertise can build a well-scoped A2A agent and monetize it as a service consumed by orchestrating agents. According to our analysis of the 2026 agentic AI landscape, the agent marketplace model is the most direct path to commercializing narrow, high-quality AI capabilities that do not require a standalone application interface — a category that includes most enterprise AI automation workflows.
Evaluating A2A Agents: The Missing Production Requirement
Building a functioning A2A agent is straightforward. Building one that performs reliably at production quality is harder, and the v0.3 ecosystem provides the tooling to close this gap. Google has extended the Vertex GenAI Evaluation Service to support A2A agent evaluation, enabling automated quality assessment of agents beyond simple unit tests.
A2A agent evaluation covers: task completion rate across a representative sample of real inputs, response latency distribution under load, accuracy of structured outputs against ground truth, graceful handling of malformed inputs or partial failures, and drift detection when upstream model versions change. These are the metrics that determine whether an agent is production-ready — and they are distinct from the model-level benchmarks that dominate AI discourse. Use our free token counter tool to estimate the context overhead of evaluation datasets when running A2A agent evaluations through the Vertex evaluation pipeline.
Who Should Build with A2A Today
A2A v0.3 is production-ready infrastructure for the following use cases:
- Enterprise orchestration systems: Any organization building a central AI orchestrator that needs to delegate tasks to specialist agents across business units, cloud regions, or external vendor systems.
- ISV agent products: Software vendors with deep domain expertise (legal, finance, healthcare, logistics) that want to expose agent capabilities to enterprise customers without building full SaaS applications.
- Internal platform teams: Engineering teams building internal AI platforms where business units can register and discover each other’s agents without direct API integration work.
- Framework builders: Teams extending LangChain, LlamaIndex, CrewAI, or Google’s Agent Development Kit (ADK) to support A2A-based agent communication natively.
If you are building multi-agent systems on a single framework, within a single codebase, and have no requirement for cross-vendor or cross-organizational agent communication, A2A may be more infrastructure than your current use case requires. But if you are building anything that needs to connect AI agents across team boundaries, vendor relationships, or cloud environments, A2A v0.3 is the right foundation to build on — before you find yourself maintaining a bespoke inter-agent RPC system that every protocol improvement will leave behind.
Getting Started in Under 30 Minutes
The fastest path from zero to a running A2A server:
- Install the SDK:
pip install a2a-sdk - Define your Agent Card with the capabilities your agent provides
- Implement skill handlers for each capability
- Run the server locally and verify the Agent Card at
http://localhost:8080/.well-known/agent.json - Test with the A2A client from the SDK: discover the agent via its card, send a task, receive the result
- Deploy to Cloud Run with a standard Dockerfile for production access
The official Python tutorial at the A2A project’s GitHub repository covers each step in detail, including multi-turn streaming conversations and authentication setup. For teams already using MCP with Claude or other frontier models, A2A integration is additive — the existing MCP tool connections remain unchanged, and A2A handles the agent-to-agent coordination layer above them.
The Bottom Line
A2A Protocol v0.3 is the release that turns agent interoperability from an interesting research direction into deployable enterprise infrastructure. The combination of gRPC transport for high-throughput performance, signed agent cards for security-compliant deployments, a mature Python SDK, and 50+ partner integrations creates a protocol ecosystem ready for production use. According to our analysis of the 2026 agentic AI stack, the development teams that invest in A2A-compatible agent architecture in Q2 2026 are building on the same foundation that will underpin the majority of enterprise AI deployments within 18 months — the same adoption curve that MCP followed from developer curiosity to 97 million monthly downloads in under 18 months.
Browse our developer tools collection for production-ready starter kits that include multi-agent orchestration patterns built for the 2026 AI stack. For a deeper look at how A2A fits alongside MCP in a complete agent architecture, read our MCP at 97M Downloads analysis and our Amazon Bedrock AgentCore guide for the cross-cloud perspective on production agent infrastructure in 2026.