The state of MCP: Insights from the Developers Summit

Learn about the latest developments in Model Context Protocol, based on key updates from the MCP Developers Summit - San Francisco, May 2025

Why everyone in AI is talking about MCP

Over the past year, Model Context Protocol (MCP) has gone from a curiosity on GitHub to the de facto way serious teams wire LLM agents into real-world data and tools. At the 2025 MCP Developers Summit in San Francisco, speakers from Anthropic, OpenAI, Bloomberg, GitHub, AWS, PayPal, Elastic, and others all echoed the same refrain:

We're shifting away from early experimentation and simple tool calling to building trustworthy, scalable and modular AI systems using MCP as the foundation.

That claim matters because integration, not model quality, is now the bottleneck. Organizations already have valuable data in GitHub, Snowflake, Elasticsearch, or Salesforce; what they lack is a safe, repeatable way for an agent to seesearch, and act on that data. MCP’s answer is a single, open, bidirectional interface - much like USB-C replaced drawers full of proprietary cables.

What exactly is MCP (Model Context Protocol)?

Model Context Protocol (MCP) is an open standard designed by Anthropic to bridge the gap between AI assistants and external data sources and make AI applications more relevant and context-aware.

Traditionally, AI models have struggled with integrating external data efficiently. The result has been fragmented implementations for each new data source.

MCP addresses this problem by providing a standardized framework for secure and scalable connections between AI tools and various systems. It supports secure two-way connections through MCP servers and clients:

  • MCP servers expose data from external sources such as Google Drive, Slack, GitHub, Postgres, and Puppeteer.
  • MCP clients (AI applications) connect to these servers to access structured information and enhance responses with relevant context.

Because the protocol is model-agnostic and server-agnostic, any Claude, GPT, Llama, or Mistral-based agent can talk to any compliant server. The result: the first real chance at plug-and-play interoperability across the AI stack.

MCP architecture
MCP architecture - source: Elastic Search Labs

A short history - and why MCP rose above competing specs

  1. The pain point. Early agent builders hard-coded REST calls for each data source, then rewrote everything when they switched clouds - or worse, when security asked for audit trails.
  2. The breakthrough. Anthropic’s early MCP prototype (late 2023) added typed messages and a lightweight stream transport so agents could pull just the context they needed.
  3. Network effects. By mid-2025, the spec had thousands of GitHub stars, dozens of SDKs, and backing from enterprise vendors who saw it as a neutral standard (unlike vendor-locked “agent frameworks”).

Competing proposals fall into two buckets:

  • Inter-agent protocols (Google A2A, Cisco/IBM ACP, Agora) focus on agents negotiating with each other.
  • Context-oriented protocols (MCP, to some extent LangGraph) focus on agents acquiring structured context - a more immediate production need.

MCP’s decision to “own context first, coordination later” gave it a pragmatic edge and the community momentum that React enjoyed over more academically pure rivals.

How MCP works today

MCP’s appeal is that it feels simple for the agent creator, yet hides a surprising amount of protocol engineering under the hood. Below is a plain-language tour of four features that were demoed at the San Francisco MCP Developers Summit, and why they matter when you move from “hello world” to production.

1. Streamable HTTP

Early MCP servers spoke over Server-Sent Events (SSE). That was fine on a laptop, but in production, SSE created headaches:

  • Enterprise firewalls often block or time out long-lived SSE connections.
  • Cloud functions such as AWS Lambda can’t keep an open socket alive for minutes.
  • Back-pressure was clunky: you either waited for the whole tool result or you polled.

The Streamable HTTP upgrade (released March 26, 2025) fixes all three issues by tunnelling a bidirectional byte-stream through a single, ordinary HTTP request that supports chunked transfer encoding. In practice, it means:

  • You can deploy an MCP server as a stateless Lambda/Fargate task; it spins up, emits streamed chunks, then shuts down.
  • Corporate proxies treat it as just another HTTPS call, so network teams stop saying “no”.
  • Agents start receiving partial results - say, the first 100 database rows - while the rest of the tool call is still running, which cuts perceived latency and lets the LLM answer sooner.

If you build on serverless or behind locked-down enterprise networks, this one feature may cut weeks from your rollout schedule.

📌
With MCP being open source, changes like Streamable HTTP are already available for developers to implement.

2. Typed schemas

  • Elicitation schema
    • A server can describe - in JSON - exactly what information it needs from the client (e.g. { "query": "string", "date_range": "YYYY-MM-DD…" }). Your UI can auto-render the right form controls. Your agent no longer has to guess the prompt that will satisfy the server, reducing error-handling glue code.
  • Tool output schema
    • A server declares the JSON shape it will return. Instead of pasting raw unstructured text into the LLM’s context window (token-hungry and brittle), the client can inject a concise, structured object - often 10–20× smaller. That leaves more budget for the actual reasoning tokens.
📌
These schemas are currently unreleased; they're in the draft stage and may evolve.

3. Sampling and Roots

Sampling and Roots are two primitives that are currently overlooked and underexplored. The keynote speech at the summit highlighted these. They're not new, but they can improve the security between agent interactions.

  • Sampling
    • Think of sampling as the server’s right to call back into your LLM for a small completion. It’s handy when the server wants, say, a short summary in the style the user already sees. Centralizing those completions in the client keeps API keys and model choice under your control.
📌
Caveat: Disable or gate sampling on third-party servers; the same mechanism could be abused for prompt-injection.
  • Roots
    • Before an agent starts browsing your Git repository, you can hand the server a roots object like ["/repos/acme-payments", "/repos/acme-docs"]. It’s an advisory scope marker - not a hard permission check - but servers that respect roots avoid needless crawling and keep responses focused. Pair it with normal ACLs (Access Control Lists) for real enforcement.

Together, these primitives facilitate richer, cooperative workflows without inventing extra protocol layers.

4. OAuth 2.1

Traditional OAuth flows assume the server redirects a human user to an auth page - a poor fit when the “user” is now an autonomous agent. Okta’s Aaron Parecki proposed (and the community adopted) an agent-first OAuth 2.1 pattern:

  1. The MCP client initiates the flow, directs the human user (or a service account) to sign in via SSO, and receives the access token.
  2. The MCP server only verifies the token and reads a tiny server.json that lists scopes and endpoints.
  • Why it matters
    • For enterprise IT: You preserve existing SSO, MFA, and audit logs; nothing new to approve.
    • For server authors: You skip building an OAuth UI, reduce your liability surface, and still accept any standards-compliant identity provider out of the box.

Put differently, OAuth 2.1 “done right” lets you keep security teams happy and ship features faster - a rare win-win in compliance-heavy environments.

Bottom line for agent builders

  1. Streamable HTTP gets your packets through corporate walls
  2. Typed schemas shrink token usage and UI glue code
  3. Sampling and Roots add power when you need it
  4. The revamped OAuth flow turns a traditional security blocker into a check-box exercise.

When you combine these four concepts, you’ll have crossed 80 percent of the gap between a weekend MCP demo and a production-ready, enterprise-approved deployment.

How to experiment with MCP right now

  • Try a turnkey client. Claude Desktop bundles a local MCP server; LibreChat and LangGraph have first-class clients.
  • Create your own server. Official SDKs exist in Python, TypeScript, and Go. Deploy with Streamable HTTP on Fly.io, Lambda, or a bare VM.
  • Try Apify's MCP Server.
    • Apify has built an MCP server that allows AI agents or frameworks that implement Model Context Protocol to access 5,000+ Apify Actors (microapps for web automation) as tools for data extraction, web searching, and other tasks.
    • This implementation lets agents collect data from websites (e.g. Facebook posts, Google search results pages, URLs), summarize web trends, and execute automated workflows without requiring user intervention.
    • The Actors MCP Server offers a bidirectional streaming endpoint at https://mcp.apify.com, a legacy SSE endpoint at https://mcp.apify.com/sse, and local stdin/stdout connections started with npx -y @apify/actors-mcp-server.
    • Users can interact with the server through clients like Claude Desktop, LibreChat, the VS Code MCP extension, and Apify's Tester MCP Client.
    • For more details, check out the docs.

Learn how to set up Apify MCP Server on Claude, add Actors, and get data in any format you want in this video

Why Model Context Protocol is worth the hype

MCP is still evolving, but its trajectory is clear. Just as HTTP unified the early web, Model Context Protocol is rapidly becoming the connective tissue of agentic AI. Getting familiar with it today is the surest way to future-proof your stack for whatever models and business demands arrive next.

Read more about MCP and AI agents

On this page

Build the scraper you want

No credit card required

Start building