Making Sense of the Agentic Web

April 2nd, 2026

AI agents are becoming a primary way people interact with the web. If you are a publisher, a product manager, or just someone trying to understand what is happening, this is where to start.


I have spent two decades building software that connects people and organizations to the web. Marketing automation, CRM systems, open source communities. In every case, the work came down to helping one type of consumer interact with web content more effectively.

Today, the consumer is changing. AI agents are browsing the web on behalf of humans. They read articles, fill forms, compare products, book appointments, and synthesize information across dozens of sources. They do this autonomously, at scale, and increasingly as the default way people access online content.

This shift is what I call the agentic web. It is not a product or a protocol. It is a structural change in who consumes web content and how. Understanding it does not require a computer science degree. It requires stepping back and seeing the pattern that is already emerging.

15x
traffic growth
AI crawling YoY (Cloudflare)
4x
token reduction
HTML to SOM compression
$2.7B
annual waste
industry token overhead

This post is my attempt to make sense of that pattern. If you are a publisher wondering why your server logs show unfamiliar traffic, a product manager evaluating how AI will affect your roadmap, or simply someone curious about where the web is heading, this is where I would start.

What we mean by "agentic"

The word "agent" gets used loosely in AI discussions. Let me be specific about what I mean.

An AI agent is software that acts on behalf of a human to accomplish a task. The human says "find me flights to Tokyo under $800" or "summarize what happened in Congress this week" or "schedule a dentist appointment for next Tuesday." The agent does the work: browsing websites, reading content, making decisions, and taking actions.

This is different from a chatbot, which waits for you to ask questions and responds with text. It is different from a search engine, which returns a list of links for you to click. An agent does the clicking, the reading, and the deciding. You get the result.

The "agentic web" is the web as seen through the eyes of AI agents. Same websites, same content, same infrastructure. But the consumer is fundamentally different from a human with a browser.

Why this matters now

You might reasonably ask: agents have existed for years. Why does this matter now?

The answer is scale.

Consider what happens when someone asks Claude, ChatGPT, or Gemini a question that requires current information. The model cannot answer from its training data alone. It needs to browse the web, find relevant pages, read them, and synthesize an answer. This happens millions of times per day. Each of those browsing sessions is an agent consuming web content.

The same pattern applies to more complex tasks. When an agent books a flight, it visits airline websites, navigates search forms, compares options, and selects the best result. When an agent researches a company for due diligence, it reads news articles, SEC filings, LinkedIn profiles, and review sites. When an agent monitors competitors, it checks pricing pages, product announcements, and job postings across dozens of sites daily.

The problem with how agents browse today

Here is where things get interesting, and where I have spent the last two years building.

AI agents today browse the web by fetching HTML pages, just like a regular browser. But unlike a browser, agents do not render pixels. They do not see the page. They process the text.

The problem is that HTML was designed for visual rendering. It contains enormous amounts of information about how a page should look: CSS class names, style attributes, layout containers, decorative elements, tracking scripts, advertising code. None of this matters to an agent trying to understand what the page says or what actions are available.

What agents receive (HTML)
<div class="sc-1x2y3z bg-gradient-to-r"> <header class="flex items-center justify-between px-4 py-2"> <nav class="hidden md:flex space-x-4"> <a href="/about" class="text-gray-600 hover:text-gray-900">About</a> </nav> </header> <main class="container mx-auto px-4 py-8"> <article class="prose lg:prose-xl"> <h1 class="text-4xl font-bold">Article Title</h1> <p class="text-gray-700 leading-relaxed">Content...</p> </article> </main> </div>
What agents need (SOM)
{ "regions": { "navigation": { "links": [{ "text": "About", "href": "/about" }] }, "main": { "type": "article", "title": "Article Title", "content": "Content..." } }, "affordances": ["navigate"] }

I wrote about this in detail in The Billion Dollar Tax on AI Agents. The short version: a typical web page contains 30,000 to 60,000 tokens of HTML markup. The actual content, the words and actions that matter, might be 5,000 tokens. The rest is presentation noise.

Average tokens per page
Raw HTML0
SOM0
Markdown0
SOM saves 0% of tokens vs HTML

Agents pay to process all of it. Language models charge by the token. Every CSS class name, every tracking pixel, every decorative div consumes tokens that contribute nothing to the agent's understanding. At scale, this waste adds up to billions of dollars per year across the industry.

When an agent's context window is filled with irrelevant markup, it has less room for actual content. It makes more mistakes. It misses important information. It hallucinates details that were obscured by the noise.

But cost is not the only problem. The noise also degrades quality. I explored this failure mode in When AI Reads the Web Wrong, looking at cases like the Air Canada chatbot that invented a refund policy. These failures are not random. They follow predictable patterns, and many trace back to the fundamental mismatch between how web content is formatted and how agents need to consume it.

What publishers are facing

If you run a website, you are already experiencing the agentic web whether you realize it or not.

Your server logs contain requests from AI crawlers. Some identify themselves clearly. Others do not. The traffic consumes your bandwidth and server resources. Unlike search engine crawlers, which visit periodically to update an index, agent traffic can be continuous and unpredictable. An AI assistant might fetch your pages every time a user asks a question about your content.

Publisher Response Strategies
Block (invisible to AI)25
Tolerate (bear the load)55
Cooperate (optimize for agents)20

Publishers are responding in different ways. Some block agent traffic entirely, treating it like unwanted scraping. Others tolerate it passively, accepting the load without optimizing for it. A few are starting to cooperate actively, recognizing agents as a legitimate audience worth serving well.

I explored these three responses in The Publisher's Dilemma. The choice matters because it affects how your content appears (or fails to appear) when people use AI assistants to find information. Publishers who block agents become invisible to that channel. Publishers who serve agents well become preferred sources.

The fourth state of the web

To understand where this is heading, it helps to see the pattern across web history.

In The Web's Fourth State, I argued that the web evolves in distinct states, each defined by the arrival of a new consumer class that existing infrastructure cannot serve well.

1991
State 1: The Human Web
HTML, CSS, JavaScript. Browsers render pixels for humans to read and interact with.
1998
State 2: The Index Web
Search engines crawl pages. Sitemaps, robots.txt, and structured data emerge.
2006
State 3: The Application Web
REST APIs, GraphQL, webhooks. Software consumes data directly.
2025
State 4: The Agent Web
AI agents need structured content, token efficiency, and semantic understanding.

The first state was the human web: browsers rendering HTML as pixels for people to read. Every feature of early web development, from HTML to CSS to JavaScript, was designed around this consumer.

The second state was the index web: search engines crawling pages to build rankings. This consumer could not render pixels. It needed structured signals about content. The web responded with sitemaps, robots.txt, and structured data (Schema.org, JSON-LD, meta tags). Publishers who adopted these formats were rewarded with better search visibility.

The third state was the application web: software consuming data through APIs. This consumer did not want HTML at all. The web responded with REST APIs, GraphQL, and webhooks. Data that had been locked in rendered pages became available as structured endpoints.

The fourth state is the agent web. AI agents are a new consumer class with distinct needs: structured content (not pixels), token efficiency (not visual fidelity), semantic understanding (not DOM selectors), and interactive affordances (what can I click, type, or select?).

The fourth consumer has arrived. But the infrastructure designed for it is only beginning to emerge.

Building for the fourth state

This is where my work comes in. After two decades of building infrastructure for web consumers, I recognized the pattern: agents need their own infrastructure layer, just as search engines and applications did before them.

That recognition led me to build Plasmate. I described the technical architecture in Why I'm Building a Browser No Human Will Ever Use. The short version: Plasmate is an open source headless browser that converts web pages into a structured format called SOM. It skips everything agents do not need (pixel rendering, visual layout) and preserves everything they do (content, structure, interactive elements, semantic regions).

But Plasmate is not just a tool. It is part of an ecosystem designed to serve the agentic web. If you are evaluating how to engage with this shift, here is how the pieces fit together.

The ecosystem: what exists and where to find it

SOM (Semantic Object Model) is a structured format for representing web pages to AI agents. It preserves content, semantic regions (navigation, main content, sidebar, footer), element types (links, buttons, form fields, headings), and interactive affordances (what can be clicked, typed, or selected).

A typical HTML page uses 30,000 or more tokens. The same page in SOM uses around 8,000 tokens. That is a 4x reduction in what agents need to process.

The specification is open and versioned. You can find it at the documentation site along with the JSON Schema that defines valid SOM documents.

Read the SOM Specification

Plasmate is the tool that converts HTML to SOM. It is open source (Apache 2.0), written in Rust, and available through multiple installation methods:

  • macOS: brew install plasmate
  • Node.js: npm install -g plasmate
  • Rust: cargo install plasmate

The compiler can run locally on your machine, in CI/CD pipelines, or as part of a server-side build process. Publishers can use it to pre-generate SOM representations of their pages and serve them alongside HTML.

View on GitHub

For agents that need to browse live web pages, fetching HTML and compiling to SOM on every request is inefficient. The SOM Cache solves this by providing a shared semantic CDN for agent consumption.

When an agent requests a page through cache.plasmate.app, the service checks if a recent SOM representation exists. If so, it returns the cached version. If not, it fetches the page, compiles it to SOM, caches the result, and returns it. Subsequent requests for the same page are served from cache.

This reduces redundant fetching (good for publishers) and reduces latency for agents (good for developers). The economics work because many agents requesting the same popular pages can share a single compilation.

If you are building with agent frameworks, Plasmate integrates directly:

  • LangChain: Drop-in SOM loader for your chains
  • LlamaIndex: SOM document reader for your indices

These integrations let you drop SOM into existing agent pipelines with minimal code changes. Instead of feeding raw HTML to your language model, you feed structured SOM.

LangChain IntegrationLlamaIndex Integration

For Claude and other AI assistants that support the Model Context Protocol (MCP), we provide an MCP server that exposes Plasmate's capabilities directly. This allows Claude to fetch and parse web pages as SOM without any integration code on your part.

MCP Server Documentation

For those who want to understand the theoretical foundations, I have published five research papers:

  1. The SOM format specification
  2. The agentic web infrastructure vision
  3. The Agent Web Protocol (AWP)
  4. Cooperative content negotiation via robots.txt
  5. A task-completion benchmark comparing web representations
Read the Research

Lessons from building Mautic

I founded Mautic in 2014. It became the world's first open source marketing automation platform before Acquia acquired it in 2019. Building Plasmate, I keep seeing the same patterns repeat.

I wrote about this in What I Learned Building Mautic That Applies to the Agentic Web. The core lesson: when a new consumer class emerges that existing tools cannot serve well, the opportunity is building purpose-built infrastructure for that consumer.

Open source matters in both cases. Mautic grew to millions of installations because organizations could try it without a sales call, deploy it without procurement, and extend it without permission. Plasmate follows the same distribution model: open source core, commercial services on top for those who need operational convenience.

Standards matter too. Mautic contributed patterns that other marketing tools adopted. With Plasmate, I am investing in specifications (SOM, AWP, robots.txt extensions) that can become web standards, not just Plasmate features. I participate in the W3C Web Content for Browser and AI Community Group because standards bodies matter even when they move slowly.

What to do from here

If you have read this far, you are probably wondering what concrete steps you can take. That depends on who you are.

Start by understanding your agent traffic. Check your server logs for requests from AI crawlers. Look at the User-Agent strings. Estimate what fraction of your traffic comes from agents versus humans.

Then consider your response strategy. Blocking agents is an option, but it makes your content invisible to AI assistants. Tolerating agents passively means you bear the server load without optimizing the experience. Cooperating actively, by serving structured representations or optimizing your HTML for agent consumption, positions you as a preferred source.

The robots.txt extension proposal (described in the research papers) offers a way to declare your preferences to agents. A SOM-Endpoint directive tells agents where to find structured representations of your pages. This is the agent equivalent of a sitemap: a machine-readable declaration that helps agents consume your content efficiently.

Think about how agents will interact with your product. Will users ask AI assistants to perform tasks on your platform? Will agents need to read your documentation, browse your product catalog, or fill your forms?

If so, consider whether your current web presence is optimized for agent consumption. Heavy JavaScript, complex navigation, and markup-heavy pages all create friction for agents. Structured data, clean semantics, and accessible forms reduce that friction.

The choice is not binary. You do not need to redesign everything for agents. But small investments in semantic markup and structured content pay dividends as agent traffic grows.

Try the tools. Install the Plasmate CLI and run it against a few pages. See the difference between raw HTML and SOM output. If you are building agent systems, integrate the LangChain or LlamaIndex packages and measure the impact on token usage and task accuracy.

The MCP server lets you expose Plasmate to Claude directly. The SOM Cache provides a shared infrastructure so you do not need to run your own compilation pipeline.

All of these are open source. You can inspect the code, file issues, contribute improvements, or fork for your own needs.

Read the other posts in this series:

Together, they form a comprehensive picture of where the web is heading and why it matters.

The transition is happening

Every transition between web states has been driven by economics, not ideology. Publishers adopted sitemaps because search visibility drove traffic. They built APIs because application integrations drove revenue. They will adopt agent-oriented infrastructure for the same reason: agents are becoming a primary channel for content discovery.

The publishers who prepare for this now will have structural advantages. Those who wait will find themselves in the same scramble that characterized early SEO, rushing to catch up after the window of easy gains has closed.

I am not saying this to sell you something. Plasmate is open source. The specifications are open. The research is public. My interest is in seeing the agentic web develop well, with infrastructure that serves both agents and publishers, rather than a chaotic land grab that benefits neither.

Interactive: Calculate your token savings
10,000
$2.5/M input tokens
HTML cost
$830
SOM cost
$208
You save
$622/mo
(75%)

If that interests you, the resources are available:

DocumentationGitHubResearch PapersPlasmate Labs

The agentic web is not coming. It is here. The question is whether you engage with it deliberately or let it happen to you.

I know which I prefer.


David Hurley is the founder of Plasmate Labs and the creator of Mautic, the world's first open source marketing automation platform. He writes about web infrastructure, AI agents, and the agentic web at dbhurley.com and publishes research at dbhurley.com/papers.