[00:00] (0.16s)
Google has introduced a new protocol
[00:02] (2.00s)
called A2A which stands for agentto
[00:04] (4.48s)
agent protocol. It enables communication
[00:06] (6.96s)
between two agentic applications or
[00:09] (9.36s)
between two agents. And the crazy part
[00:11] (11.44s)
is that this protocol can connect AI
[00:13] (13.68s)
agents from any framework whether it's
[00:16] (16.08s)
Langchain, Crew AI, Google's ADK or even
[00:20] (20.24s)
customuilt systems. A2A runs on a shared
[00:23] (23.52s)
protocol built on HTTP. This protocol is
[00:26] (26.40s)
honestly wild and the kind of
[00:28] (28.08s)
applications and implementations we're
[00:29] (29.84s)
going to see from it are going to be
[00:31] (31.04s)
mind-blowing. Just like MCP, it's
[00:33] (33.36s)
expected to gain a lot of traction. As
[00:35] (35.36s)
more applications are developed with it,
[00:37] (37.28s)
the momentum will only grow. The best
[00:39] (39.28s)
part is that it doesn't replace MCP.
[00:41] (41.60s)
Instead, it works alongside it and both
[00:43] (43.68s)
can be used together to build powerful
[00:45] (45.60s)
systems. Let's jump into the video and
[00:47] (47.44s)
see how it all works. Right now, I'm
[00:49] (49.04s)
going to show you an official demo from
[00:50] (50.88s)
Google. This is exactly what they
[00:52] (52.72s)
demonstrated. It starts with the end
[00:54] (54.64s)
user which is you. The client in the
[00:56] (56.80s)
demo is Google agent space but it could
[00:59] (59.12s)
be any client. It begins with a single
[01:01] (61.28s)
agent called the remote agent. Based on
[01:03] (63.28s)
the task you give it, this agent looks
[01:05] (65.44s)
for other agents to hand the task over
[01:07] (67.36s)
to. This is where the A2A protocol steps
[01:09] (69.92s)
in. It enables smooth communication
[01:12] (72.00s)
between two AI agents. Now what exactly
[01:14] (74.80s)
is an AI agent? They're just large
[01:17] (77.04s)
language models with a set of tools.
[01:19] (79.04s)
These tools define what the agent can
[01:21] (81.20s)
do. In this new protocol, every agent
[01:23] (83.84s)
has an agent card that describes its
[01:25] (85.84s)
abilities. The remote agent reads the
[01:28] (88.00s)
agent cards of other agents and picks
[01:30] (90.40s)
the one best suited for the task. That
[01:32] (92.48s)
agent can then pass the task on to
[01:34] (94.32s)
another agent, creating a chain. This
[01:36] (96.40s)
forms a multi- aent workflow. This is
[01:38] (98.56s)
how A2A makes the process much easier.
[01:41] (101.36s)
I'll show you in the GitHub repo how it
[01:43] (103.44s)
works. You can connect agents from any
[01:45] (105.60s)
framework and once this protocol is in
[01:47] (107.84s)
place, those agents can talk to each
[01:49] (109.84s)
other without friction. Now this is the
[01:51] (111.68s)
demo that they have shared. The example
[01:53] (113.60s)
they've provided shows the agentto agent
[01:55] (115.84s)
protocol being used to hire a software
[01:58] (118.00s)
engineer along with a job description.
[02:00] (120.24s)
You can clearly observe how the A2A
[02:02] (122.40s)
protocol functions. First the protocol
[02:04] (124.96s)
is initiated which is visible in the
[02:07] (127.20s)
thinking
[02:08] (128.52s)
process. Then to discover different
[02:11] (131.04s)
agents suitable for the task, it
[02:13] (133.12s)
examines their agent cards. These are
[02:15] (135.12s)
the main source for understanding the
[02:16] (136.88s)
capabilities of each agent. There are
[02:18] (138.96s)
several ways to explore agent
[02:22] (142.12s)
cards. In this case, there's a tool
[02:24] (144.72s)
registry where it finds the sourcing
[02:26] (146.40s)
agent and initiates a call to it. You
[02:28] (148.80s)
can see here that additional constraints
[02:30] (150.64s)
have also been provided to the agent to
[02:32] (152.72s)
find the best possible candidate. Once
[02:35] (155.04s)
the sourcing agent completes its task,
[02:37] (157.36s)
we can see it identifies five candidates
[02:39] (159.60s)
for the job. If you continue watching
[02:41] (161.28s)
the video, you'll notice that two weeks
[02:43] (163.04s)
later, after the interviews are done,
[02:45] (165.20s)
the agent is used again to gather
[02:46] (166.96s)
updates and perform background checks.
[02:50] (170.00s)
Based on the candidate data, the system
[02:52] (172.24s)
is capable of running a background check
[02:54] (174.24s)
on a single candidate. This entire
[02:56] (176.40s)
process of hiring a software engineer
[02:58] (178.56s)
based on a job description is automated
[03:01] (181.12s)
using these agents. The most important
[03:03] (183.12s)
thing is that this wasn't handled by a
[03:05] (185.20s)
single agent. The A2A protocol allowed
[03:07] (187.84s)
multiple agents to work together. all
[03:10] (190.16s)
communicating through a single protocol.
[03:12] (192.24s)
Let's clear up another concept about the
[03:14] (194.08s)
agentto agent protocol and MCP. A2A is
[03:17] (197.52s)
meant to work with MCP and even Google
[03:20] (200.08s)
has confirmed this. To explain why,
[03:22] (202.48s)
think of it like this. MCP is an LLM
[03:24] (204.80s)
with tools or access to specific data.
[03:27] (207.04s)
Picture it as a repair man. He has a
[03:29] (209.12s)
screwdriver and the knowledge to fix
[03:30] (210.80s)
cars. That's the MCP part. But this
[03:33] (213.28s)
repair man also needs to talk to others.
[03:35] (215.76s)
Maybe he needs to speak with clients or
[03:37] (217.68s)
borrow parts from someone else. That's
[03:39] (219.60s)
where the agentto aagent protocol comes
[03:41] (221.60s)
in. It allows agents to communicate with
[03:43] (223.76s)
each other. These agents could even be
[03:45] (225.76s)
separate MCP servers acting as
[03:47] (227.68s)
independent agents. They can share tools
[03:50] (230.08s)
or request help when needed. The key to
[03:52] (232.16s)
all of this is the agent card. It
[03:54] (234.32s)
defines what each agent is capable of
[03:56] (236.24s)
and helps them interact in a structured
[03:58] (238.24s)
way. They've explained a clear
[03:59] (239.68s)
connection between A2A and
[04:07] (247.88s)
MCP. In their official documentation,
[04:10] (250.80s)
it's outlined that future Agentic
[04:12] (252.72s)
applications will need both of these
[04:14] (254.40s)
protocols to become more capable. As an
[04:16] (256.56s)
example, they use the same auto repair
[04:18] (258.32s)
shop analogy. There are multiple
[04:20] (260.24s)
handyman sub agents working in the shop.
[04:22] (262.40s)
They need tools, but to use those tools
[04:24] (264.64s)
properly, they also need extra context
[04:26] (266.72s)
from customers. Those customers could be
[04:29] (269.04s)
other agents. The interesting part is
[04:30] (270.80s)
how MCP fits into this setup. We know
[04:33] (273.20s)
that to identify an agent using the A2A
[04:35] (275.84s)
protocol, it must have an agent card.
[04:38] (278.24s)
These agent cards can be listed as
[04:40] (280.08s)
resources. The MCP server can then
[04:42] (282.48s)
provide access to them. The LLM fetches
[04:44] (284.72s)
these agent cards and passes them to sub
[04:46] (286.96s)
agents. Sub agents read the cards and
[04:49] (289.60s)
based on the information, the LLM
[04:51] (291.68s)
decides which external agent should be
[04:53] (293.76s)
used. It's a clever integration and
[04:55] (295.84s)
shows how both systems can work together
[04:57] (297.92s)
in a flexible way. The agent card
[05:00] (300.00s)
structure is clearly defined in their
[05:02] (302.20s)
documentation. It includes key
[05:04] (304.16s)
information about the AI agent such as
[05:06] (306.64s)
the version, name, description, and its
[05:09] (309.44s)
intended use. Then it lists the skills
[05:11] (311.84s)
which are the core capabilities the
[05:13] (313.68s)
agent can perform. It also shows the
[05:15] (315.68s)
default content type the agent supports
[05:18] (318.24s)
along with the parameters it accepts.
[05:20] (320.48s)
Basically, it shows the kind of input
[05:22] (322.40s)
the agent needs. For some agents,
[05:24] (324.32s)
authentication might be required. So,
[05:26] (326.48s)
specific document details would also be
[05:28] (328.56s)
part of the agent card. When an LLM or
[05:30] (330.88s)
another agent tries to access this
[05:32] (332.80s)
agent, it first reads the agent card.
[05:35] (335.44s)
Based on that, it decides whether to use
[05:37] (337.76s)
the agent and how to interact with it.
[05:40] (340.00s)
This makes the accuracy of the agent
[05:41] (341.92s)
card critical to how the entire agentto
[05:44] (344.40s)
agent system functions. Further in the
[05:46] (346.60s)
documentation, they've provided sample
[05:48] (348.96s)
agent cards and methods for sending
[05:50] (350.96s)
tasks and receiving responses. One of
[05:53] (353.60s)
the examples is a Google Maps agent. The
[05:55] (355.84s)
card includes a clear description of the
[05:57] (357.68s)
tasks it can perform along with the URL
[06:00] (360.32s)
and provider details. It also specifies
[06:02] (362.88s)
the type of authentication the agent
[06:04] (364.72s)
needs. Below that, there's a format
[06:06] (366.48s)
showing how a client can send a task to
[06:08] (368.40s)
a remote agent to start a new job. In
[06:10] (370.72s)
one simple example, the task is to tell
[06:12] (372.64s)
a joke. The response comes back as a
[06:14] (374.56s)
text output from a model which delivers
[06:16] (376.80s)
the joke. This is one way to send a task
[06:19] (379.04s)
and get a
[06:34] (394.36s)
result. Other methods are also
[06:36] (396.64s)
documented below. To get started,
[06:38] (398.88s)
there's no need to memorize the syntax.
[06:41] (401.12s)
You can feed the documentation into
[06:42] (402.88s)
cursor with the at docs feature which
[06:45] (405.36s)
will pick up the context and generate
[06:47] (407.28s)
code accordingly. In the GitHub repo,
[06:49] (409.60s)
they've included some sample agents that
[06:51] (411.76s)
show how A2A agents can be implemented.
[06:54] (414.80s)
One example is using Crew AI, a simple
[06:57] (417.92s)
image generation agent that uses the
[07:00] (420.48s)
Google Gemini API. You can install it
[07:02] (422.96s)
easily. It's a basic agent that just
[07:05] (425.20s)
runs on the A2A protocol. To get
[07:07] (427.36s)
started, you need to clone the full
[07:09] (429.28s)
GitHub repo because the commands depend
[07:11] (431.76s)
on that
[07:19] (439.16s)
structure. Once it's cloned, you can run
[07:21] (441.60s)
the setup using a few simple commands.
[07:23] (443.84s)
Just copy and paste them into your
[07:30] (450.52s)
terminal. They're very straightforward.
[07:32] (452.96s)
Once you run the command, it opens a
[07:35] (455.20s)
command line interface for the A2A
[07:37] (457.28s)
agent. The server passes it to the crew
[07:39] (459.36s)
agent which uses the Gemini API to
[07:41] (461.84s)
generate an image. That image is then
[07:43] (463.92s)
returned to the server and finally back
[07:46] (466.00s)
to the client. This is a simple
[07:47] (467.60s)
implementation and it's not widely used
[07:49] (469.60s)
just yet, but it will be. We saw the
[07:51] (471.84s)
same thing happen with MCP. Adoption
[07:53] (473.92s)
took some time, then it exploded. The
[07:55] (475.92s)
same will happen here. Once people start
[07:57] (477.92s)
building on top of it, AI agents will
[08:00] (480.08s)
become incredibly powerful. They'll
[08:01] (481.76s)
automate a huge amount of work and
[08:03] (483.60s)
change how we use AI everyday. If you
[08:06] (486.40s)
found this video helpful, subscribe to
[08:08] (488.88s)
the channel. And if you'd like to
[08:10] (490.48s)
support the content, check out the
[08:12] (492.32s)
donation link in the description. Thanks
[08:14] (494.48s)
for watching and I'll see you in the
[08:16] (496.40s)
next video.