The spec Agent Readability, defined.
An open set of practices that make any website discoverable,
parseable, and comprehensible to AI agents. The scorecard
operationalizes the spec into pass/fail checks; the tools run
those checks against real sites.
What it is
Agent readability is the property of a website that lets AI
agents — assistants like ChatGPT, Claude, Copilot, and Cursor, as well as
autonomous agents built on top of them — find its pages, fetch their content
without wasting tokens on layout chrome, and extract the facts, code, and
context a user actually asked for. The goal isn't to turn every page into
markdown or build a separate "AI version" of your site. It's to expose the
signals agents need alongside the page you already serve to humans.
Why it matters
Discovery is shifting. For a growing share of users, the first answer about a
product, a help-center question, or a technical topic comes through an AI
agent, not a search engine — and often not the linked site at all. When an
agent is the reader, three things change. Tokens cost money, so scraping a
5 MB HTML document to surface two paragraphs is wasteful. Layout and
navigation chrome dominates raw HTML by byte count, so agents can miss the
content signal altogether and hallucinate from training data. And because
agents rarely cite their sources by URL, a site that is opaque to them
quietly disappears from conversation.
Agent readability is to AI-mediated discovery what SEO was to search:
cheap to implement, compounding in return, and increasingly table stakes.
It applies to any website that wants to be read by an agent — documentation
portals and developer references are a common and high-value target, but
product pages, marketing sites, help centers, and long-form content all
benefit from the same practices.
The three layers
Every check in the scorecard maps to one of these layers. They build on
each other: a page agents cannot find never gets parsed, and a page
parsed into garbage never gets comprehended.
Layer 1
Discoverability
Can agents find every page you want them to read? Covers
llms.txt, robots.txt (with AI bots allowed),
XML and markdown sitemaps, and AGENTS.md skill files.
Orphaned pages, absent sitemaps, and AI-excluding robots rules all
make a site invisible to agents by default.
Layer 2
Parsing
Once an agent fetches a page, can it extract the content efficiently?
Covers HTTP basics (no redirect chains, correct content types), HTML
metadata (canonical link, lang, OG, meta description), markdown
mirrors (.md at the same URL with frontmatter), and
content-to-chrome ratio. The cheapest win is usually a mirror — one
.md file per page, served with the right content type.
Layer 3
Comprehension
Once parsed, does the content self-describe? Covers structured data
(JSON-LD with dateModified and BreadcrumbList),
language-tagged code blocks, API schema links, and semantic heading
hierarchy. These are the signals an agent uses to generate
accurate citations, working code examples, and cross-references
that don't hallucinate.
How the scorecard operationalizes it
The a14y scorecard is a
versioned, frozen manifest of pass/fail checks that maps directly to the
three layers above. Each check has a stable id, a documented detection
rule, pass/fail examples, and fix guidance — so a failing result is an
immediately actionable work item, not a vague "improve your SEO" nudge.
Scorecards are frozen once shipped: v0.2.0 always evaluates the same
way, so historical scores compare cleanly even as the engine evolves.
Pinning a scorecard version lets teams track trend over time and gate
merges in CI (--fail-under 80) without every engine
improvement resetting the baseline.
Status
The spec is open and evolving. v0.2.0 is the current scorecard — see
the check list for exactly
what it measures today. New scorecard versions publish additively: older
versions stay supported so your existing audits keep reproducing.
Contributions, issues, and proposed new checks live on
GitHub.