The State of AI Readiness: Early Findings From 4,500 Domains
We scanned 4,500 domains across 12 industries and the data is clear: the average AI readiness score is 41 out of 100. Here are the key findings — which industries lead, which lag, and where the biggest opportunities are.
Founder & CEO at AgentReady
The Dataset: 4,500 Domains Across 12 Industries
Between January and March 2026, we scanned 4,500 domains using the AgentReady scoring framework. The dataset spans 12 industries: Technology, SaaS, E-Commerce, Healthcare, Finance, Education, Media & Publishing, Legal, Real Estate, Travel, Manufacturing, and Food & Beverage.
For each domain we evaluated the homepage, up to 50 interior pages, robots.txt, sitemap.xml, and checked for the presence of AI protocol files (llms.txt, NLWeb, MCP). Schema markup was validated against Schema.org specifications. Content quality was assessed through heading structure, depth, and citation density.
The goal was not to cherry-pick results. We selected domains from public directories, industry lists, and existing AgentReady user scans to build the most representative sample possible.
Headline Numbers
The average AI readiness score across all 4,500 domains is 41 out of 100 — a solid D grade. The median is even lower at 38. Only 6.8% of sites scored above 70 (B grade or higher), and a mere 2.1% scored above 85 (A grade).
At the other end of the spectrum, 29% of sites scored below 30, placing them firmly in F territory. These sites are functionally invisible to AI agents — they block crawlers, lack structured data, have no protocol files, and often serve content primarily through client-side JavaScript.
The distribution is heavily skewed toward the bottom. The web is not AI-ready. But within that reality, there are massive opportunities for sites willing to invest early.
Average Score by Industry
The gap between the highest and lowest scoring industries is 28 points — an enormous spread that reflects fundamental differences in technical sophistication, content strategy, and awareness of AI-driven discovery.
Technology & SaaS leads at 58, driven by developer-oriented teams, strong schema adoption, and early protocol experimentation. These sites are most likely to have llms.txt files and typically have clean, well-structured HTML.
Media & Publishing follows at 54, benefiting from content depth, author attribution, and established structured data practices from years of Google News optimization.
At the bottom, Food & Beverage averages 30, dragged down by template-based sites with minimal structured data and zero AI protocol awareness. Real Estate averages 32, hampered by IDX integration issues that make content inaccessible to crawlers. Healthcare averages 35, where regulatory concerns lead many sites to aggressively block all non-Google bots.
Average AI Readiness Score by Industry (N=4,500)
Protocol Adoption Rates
The AI protocol landscape is still in its infancy. Adoption rates across our dataset paint a picture of a web that has barely begun to prepare for AI interaction.
llms.txt is the most adopted protocol at 8.2% of sites. This is the simplest to implement — a plain text file at the domain root — yet more than 90% of sites have not created one. Among technology companies the rate jumps to 18%, while in food and beverage it sits below 1%.
NLWeb adoption is at 1.4%. As a more complex protocol that requires a query endpoint, lower adoption is expected. But among the sites that have implemented it, AI readiness scores average 22 points higher than non-adopters.
MCP (Model Context Protocol) is at 0.6%, confined almost entirely to developer tool companies and API-first platforms. This protocol represents the future of AI-website interaction but is still in early experimental stages for most industries.
The aggregate picture: 91% of sites have zero AI protocol adoption. This is the single largest untapped opportunity in AI readiness.
- llms.txt adoption: 8.2% — Simple text file, highest ROI, 15 minutes to implement
- NLWeb adoption: 1.4% — Conversational endpoint, higher complexity, 22-point average score lift
- MCP adoption: 0.6% — Action protocol, developer-focused, still experimental for most industries
- Zero protocols: 91% — The vast majority of the web has not begun preparing for AI interaction
What Top-Performing Sites Have in Common
We isolated the 95 sites scoring 80 or above and analyzed their shared characteristics. The patterns are remarkably consistent.
100% have comprehensive Schema.org markup spanning at least 3 distinct types (Organization plus content-specific types). 89% have an llms.txt file. 92% allow all major AI crawlers without restriction. 96% have clear author bylines with dedicated author pages. And 78% publish content that includes original data, studies, or unique research.
But the most revealing finding is about intentionality. Top-scoring sites treat AI readiness as a deliberate strategy, not a byproduct of good SEO. They have dedicated protocol files. They monitor AI crawler traffic in their analytics. They test how their content appears in AI-generated responses. They iterate.
This intentional approach is what separates a site that happens to score well from a site that is genuinely AI-ready.
The Biggest Gaps and Opportunities
Five issues account for the majority of lost points across our dataset. Fixing just these five would move the average score from 41 to an estimated 62 — a jump from D grade to C+.
Missing structured data (82% of sites): The most common and most impactful gap. Most sites either have no Schema.org markup or only basic Organization schema. Adding Article, Product, FAQ, or HowTo schemas to relevant pages is the highest-leverage single action.
No llms.txt file (92% of sites): A 15-minute task that immediately improves how AI systems understand your site. There is no reason not to have one.
Blocking AI crawlers (34% of sites): Many sites block GPTBot, ClaudeBot, or PerplexityBot in robots.txt, often unintentionally via inherited configuration files or overzealous security plugins.
Thin content on key pages (61% of sites): Pages with fewer than 300 words, no headings, or no supporting citations score poorly. AI systems prefer depth and structure.
No author attribution (58% of sites): Pages without clear authorship lose trust signals. AI systems weigh E-E-A-T factors heavily when deciding which sources to cite.
- Add Schema.org markup to all content pages — estimated 8-12 point score improvement
- Create an llms.txt file at your domain root — estimated 4-6 point score improvement
- Audit robots.txt for AI bot blocking — estimated 3-5 point score improvement if bots were blocked
- Expand thin content to 500+ words with clear headings — estimated 3-5 point score improvement
- Add author bylines and author pages — estimated 2-4 point score improvement
Frequently Asked Questions
How were the 4,500 domains selected?
Domains were selected from public business directories, industry association member lists, and existing AgentReady user scans. We aimed for representative coverage across company sizes (small business to enterprise) and geographic regions (primarily US, UK, and EU).
Will this study be updated?
Yes. We plan to re-run this analysis quarterly to track how AI readiness evolves over time. The next update is scheduled for June 2026 with an expanded dataset of 10,000+ domains.
Can I see my industry's detailed breakdown?
Detailed industry breakdowns with factor-level scores are available on our benchmarks page. You can also scan your own site and compare against industry averages directly in the AgentReady dashboard.
Check Your AI Readiness Score
Free scan. No signup required. See how AI engines like ChatGPT, Perplexity, and Google AI view your website.
Scan Your Site FreeRelated Articles
Introducing AgentReady: The AI Readiness Score for Every Website
We built AgentReady because the web is entering its AI era and nobody had a way to measure whether their website was ready. Today we are launching a free tool that scans any site and scores it across 8 factors that determine AI visibility.
Data & ResearchWhat Is AI Readiness and Why It Matters in 2026
ChatGPT, Perplexity, Claude, and Google AI Overviews are changing how people discover information. AI readiness is the measure of whether your website is visible to these systems — and in 2026, it matters more than ever.
MethodologyInside the Algorithm: How We Score AI Readiness
Full transparency on how AgentReady calculates your AI readiness score. The 8 factors, their weights, the grade scale from A to F, bonus signals, and how it all connects to real AI citation outcomes.