The evolution of machine-to-machine communication has always followed a predictable pattern. In my younger days of enterprise computing, email—designed for human communication with subjects, greetings, and signatures—was repurposed for system integration. Organizations would parse email subjects with specific keywords to trigger automated workflows, a crude but necessary workaround when systems sat behind firewalls with limited connectivity options.
This inefficiency drove innovation. The industry evolved through message queuing systems (MQ), then SOAP with its XML verbosity, and eventually to REST APIs with clean JSON payloads. Each transition represented a shift toward more efficient, purpose-built protocols for machine communication.
Today, we face a similar inflection point. AI agents are forced to consume HTML—a format designed for human visual consumption—to extract structured data. Just as email was never meant for system integration, HTML was never designed for AI consumption.
It's time for an AI Content Standard.
Every day, millions of AI agents and LLMs waste computational resources parsing HTML designed for human eyes, not machine consumption. When an AI agent visits a news website, e-commerce store, or social media platform, it must:
Download 100KB+ of HTML, CSS, and JavaScript
Parse through presentation markup (`<div>`, `<span>`, styling)
Extract semantic meaning from visual layouts
Handle inconsistent structures across sites
Deal with dynamic content and anti-scraping measures
The result? A single query that should cost pennies ends up costing dollars in compute and API calls.
Real Cost Example
Current State (HTML Scraping):
Average webpage: 2,000+ tokens of HTML
LLM processing: $0.003 per 1K tokens (Claude)
Cost per page: ~$0.006
1 million pages: $6,000
With Structured JSON:
Structured data: 200 tokens
LLM processing: $0.003 per 1K tokens
Cost per page: ~$0.0006
1 million pages: $600
With GraphQL this would be even less and equally proportional to HTML - JSON , as most of REST APIS return lot of unwanted data for specific processing , GRAPHQL solves that by letting Agents select exact data needed for processing a task.
The Benefits: Why This Matters
Savings: 90% reduction in cost and 10x faster processing
10x faster data extraction, thus lesser compute and Energy
Lower bandwidth consumption
Accuracy Improvement
Universal parsing code
- Easier maintenance
- Faster AI responses
- More reliable information
- Richer integrations
- Already improves search rankings
- Better visibility in AI results
- Future-proof for AI agents
What We Currently Have
1. HTML for Humans
<div class="article">
<h1 class="title">Breaking News</h1>
<span class="author">John Doe</span>
<time>2024-01-15</time>
</div>
Rich visual presentation
Inconsistent structure across sites
10-100x more tokens than needed
2. APIs for Developers
{
"title": "Breaking News",
"author": "Amin Asif",
"date": "2026-04-22"
}
- Efficient and structured
We already have systems and machines to deal with varaitions of schema in every api
3. Schema.org Embedded in HTML
<script type="application/ld+json">
{
"@context": "https://schema.org",
"@type": "NewsArticle",
"headline": "Breaking News",
"author": {"@type": "Person", "name": "John Doe"},
"datePublished": "2024-01-15"
}
</script>
- Standardized vocabulary (800+ types)
- Machine-readable
- Problems:
- Still embedded in HTML (must download full page)
- Voluntary adoption (most sites don't use it)
- Incomplete coverage
4. RSS/Atom Feeds
<item>
<title>Breaking News</title>
<author>John Doe</author>
<pubDate>2024-01-15</pubDate>
</item>
- Separate machine-readable endpoint
- Widely adopted for blogs/news
What We're Missing: The Universal Machine-Readable Web
The Vision: Dual-Interface Web
Just as websites provide both HTML (for humans) and RSS (for feed readers), every meaningful webpage should offer:
For Humans:
https://example.com/article/123
→ Returns HTML with styling, images, ads
For Machines:
https://example.com/article/123.json
→ Returns structured JSON with Schema.org vocabulary
Why Schema.org is the Nearest Answer
Schema.org already provides:
800+ standardized types: Article, Product, Person, Event, Recipe, JobPosting, etc.
Industry backing: Created by Google, Microsoft, Yahoo, Yandex
Proven adoption: Used by millions of sites for SEO
Extensible: Can add new types as needed
Language-agnostic: Works with JSON-LD, RDFa, Microdata
Three-Layer Solution :
Layer 1: Transport MCP (connectivity) -We have already standardized the Transport
How to discover available data sources
How to authenticate and connect
How to call functions/tools
Layer 2: GraphQL (precise querying)
The over-fetching problem with REST
How GraphQL lets agents request only needed fields
Real cost comparison: REST ($6,000) vs GraphQL ($150) for 1M queries
97.5% cost reduction when combining JSON + GraphQL
Concrete examples comparing:
REST returning 500 tokens (everything)
GraphQL returning 50 tokens (only what's needed)
GraphQL is essential for AI agent efficiency not a nice to have
Layer 3: Schema.org (common vocabulary)or any clever mechanism that solves this problem
What vocabulary to use
How to structure data
What properties are available
ALL layers are essential. MCP without Schema.org is like having a phone but no common language to speak.
Implementation:
With AI as Coding assistant , i dont think this is a difficult problem to solve
This document advocates for a fundamental shift in how we think about web content—not just for humans, but for the AI agents that increasingly mediate our interactions with information.