News

68.9 Million AI Crawler Visits Analyzed — OpenAI Commands 81% of All AI Crawl Traffic

A study of 858K sites and 68.9M AI crawler visits reveals OpenAI sends 81% of AI crawl traffic, content depth drives a 33x visibility multiplier, and Stanford's 2026 AI Index shows 53% global adoption in 3 years.

Updated April 20, 2026 Francisco Leon de Vivero
68.9 Million AI Crawler Visits Analyzed — OpenAI Commands 81% of All AI Crawl Traffic

Two datasets this week changed how I think about AI crawl traffic and its relationship to real business outcomes. The first: a 68.9 million AI crawler visit study across 858,457 sites that reveals exactly who is crawling the web, how often, and what makes sites visible to AI systems. The second: Stanford's 2026 AI Index, which documents the fastest technology adoption curve in history — 53% global enterprise adoption in just three years. Together, they paint a clear picture of what's working, what's not, and what to prioritize right now.


1. Inside the 68.9M AI Crawler Visit Study

The scale of this dataset is what makes it credible. Researchers analyzed 68.9 million AI crawler visits across 858,457 websites, covering every major AI crawling system active between mid-2025 and early 2026. Of those sites, 59% received at least one AI crawler visit — meaning over 506,000 sites are already being scanned by AI systems whether they know it or not.

68.9M Total AI crawler visits analyzed
858K Sites in the study
59% Sites received at least one visit
81% OpenAI's share of AI crawl traffic
AI crawler market share showing OpenAI at 81%
AI crawler market share — OpenAI dominates with 81% of all AI crawl traffic.

The market concentration is striking. OpenAI accounts for 81% of all AI crawl traffic — 55.8 million of the 68.9 million total visits. Anthropic sits at a distant second with 16.6% (11.5 million visits), followed by Perplexity at 1.8% (1.3 million) and Gemini at just 0.6% (380,000). This isn't a competitive market for crawl access — it's a near-monopoly.

AI CrawlerVisitsMarket Share
OpenAI (GPTBot + ChatGPT-User)55.8M81.0%
Anthropic (ClaudeBot)11.5M16.6%
Perplexity (PerplexityBot)1.3M1.8%
Gemini (Google-Extended)380K0.6%

But the type of crawling matters more than the volume. The study identified three distinct crawl categories:

  • User-fetch crawls (56.9%) — triggered when a user asks ChatGPT, Claude, or Perplexity a question that requires real-time web data. These represent actual demand for your content from end users.
  • Training crawls (28.8%) — systematic crawling to build or update foundational LLM training datasets. This is the crawl type that robots.txt directives target.
  • Discovery crawls (14.3%) — exploratory crawling to map site structure, assess content freshness, and build retrieval indexes.

ChatGPT alone generated 39.8 million user-fetch visits — by far the largest single source of demand-driven AI crawling. If your site is being crawled by AI, there's an 81% chance it's OpenAI, and a 57% chance it's because a real user asked a question that led to your content. The implications for AI agent readiness are significant — the crawlers are already at the door, and most sites aren't prepared.

Key takeaway

OpenAI sends more AI crawl traffic than all other AI systems combined. The 81% market share means that optimizing for GPTBot and ChatGPT-User crawlers should be the priority — not a balanced multi-crawler strategy. And with 57% of crawls being user-fetch (real user demand), AI crawler visits are increasingly a proxy for actual referral potential.


2. LLM Referral Traffic Is Growing Fast — but Unevenly

Crawling is one side of the equation. The other is whether those crawls translate into actual traffic. The referral data from the same study period shows total LLM referral traffic grew 72.7% year-over-year, from 93,484 to 161,469 measured referral sessions. But the growth is wildly uneven across platforms.

SourcePrevious YearCurrent YearGrowth
ChatGPT81,652136,095+66.7%
Copilot229,560+43,354%
Claude1062,488+2,247%
Perplexity10,50811,991+14.1%
Gemini1,1961,335+11.6%

ChatGPT dominates absolute volume with 136,095 referral sessions — 84.3% of all LLM referral traffic. But the growth stories are elsewhere. Microsoft Copilot exploded from near-zero (22 sessions) to 9,560 — a function of integration into Windows, Edge, and Office. Claude grew 23x from 106 to 2,488 sessions, reflecting Anthropic's expanding user base and the introduction of web search in Claude.

Perplexity's 14.1% growth is notably modest given its positioning as the "answer engine" that always cites sources. Despite being purpose-built for search with citations, Perplexity sends less referral traffic than ChatGPT by a factor of 11x. This aligns with the ChatGPT citation mechanics data — citation frequency doesn't automatically translate to click-through behavior.

Watch out: The absolute numbers are still small compared to traditional search. 161,469 total LLM referrals across all platforms is a rounding error for most sites getting millions of Google sessions. The signal isn't the volume — it's the 72.7% growth rate. At that trajectory, LLM referrals become material within 2-3 years for content-heavy sites.


3. What Actually Drives AI Crawler Visibility

This is where the study gets actionable. The researchers correlated dozens of site-level attributes with AI crawler visit frequency and identified a clear hierarchy of factors that predict whether AI systems will crawl your site — and how often.

Ranked factors driving AI crawler visibility
Visibility factors ranked by impact on AI crawler visit frequency.

Content depth is the single strongest predictor. Sites with 50+ published pages or posts averaged 1,373.7 AI crawler visits, compared to just 41.6 visits for sites with fewer than 10 pages. That's a 33x multiplier. The AI crawlers behave like search engines in this regard — they reward sites that produce substantial, indexable content. This resonates with the 2MB crawl cutoff findings: the more structured, crawlable content you publish, the more budget AI systems allocate to your domain.

FactorMetricImpact
Content depth (50+ posts)1,373.7 vs 41.6 visits33x multiplier
Yext listing sync97.1% crawl rateHighest platform correlation
Google Business Profile sync92.8% vs 58.9% crawl rate+33.9 percentage points
Local schema (10-11 fields)82% vs 55.2% crawl rate+26.8 percentage points
E-commerce vertical-5.0pp crawl rateNegative correlation

The local SEO signals are surprisingly strong for AI crawlers. Sites synced to Yext had a 97.1% AI crawl rate — nearly guaranteed visibility. Google Business Profile sync showed a 92.8% crawl rate versus 58.9% for sites without it, a 33.9 percentage point advantage. Local schema markup with 10-11 fields correlated with an 82% crawl rate, versus 55.2% for sites with incomplete or missing local schema.

But only 22.3% of sites in the study had any local schema at all. This is a massive gap between what works and what people are actually doing.

Surprise finding: E-commerce sites showed a -5.0 percentage point correlation with AI crawler visibility. Being an e-commerce site actually makes you less likely to be crawled by AI systems. The researchers suggest this may relate to product catalog structures (thin pages, parameter-heavy URLs, JavaScript-rendered content) that AI crawlers deprioritize compared to content-rich informational sites.

Key takeaway

The data is unambiguous: publish deep content, sync your Google Business Profile, and complete your local schema. These three actions account for the largest AI visibility gaps in the study. The 33x multiplier for content depth alone should change how you allocate content budgets.


4. The Correlation Paradox

The study's most provocative data point: sites that receive AI crawler visits also perform dramatically better on traditional web metrics. AI-crawled sites averaged 3.2x more human sessions (527.7 vs 164.9), 2.7x more form completions, and 2.5x more click-to-call actions compared to non-crawled sites.

3.2x More human sessions on AI-crawled sites
2.7x More form completions
2.5x More click-to-call actions
527.7 Avg sessions (AI-crawled sites)

On the surface, this looks like AI crawling causes better business outcomes. It doesn't — at least not directly. The researchers are explicit that correlation is not causation here. The more likely explanation is a shared root cause: sites that are well-built, content-rich, properly structured, and actively maintained tend to both (a) attract more AI crawlers and (b) perform better with human visitors. The same attributes that make a site interesting to GPTBot — depth, freshness, structured data, technical quality — also make it rank well in traditional search and convert visitors effectively.

The practical implication: You don't need to optimize "for AI crawlers" as a separate discipline. The actions that make your site visible to AI systems are the same ones that improve traditional SEO and conversion rates. There is no AI SEO vs regular SEO tradeoff — it's the same playbook, reinforced. A solid technical SEO foundation serves both audiences simultaneously.

The one exception is blocking decisions. If you block AI crawlers via robots.txt, you lose the AI visibility without gaining anything in traditional performance. The study found no evidence that blocking AI crawlers improves traditional search rankings or site performance. Unless you have specific IP or content licensing concerns, the default should be to allow crawling.


5. Stanford 2026 AI Index — Adoption at Unprecedented Speed

Stanford's annual AI Index dropped this week and the headline number is staggering: 53% of organizations globally have adopted AI within 3 years of generative AI becoming widely available. For context, personal computers took approximately 15 years to reach 50% enterprise adoption. The internet took about 7 years. AI did it in 3.

Stanford AI Index 2026 adoption timeline
Stanford AI Index 2026 — Global AI adoption reached 53% in 3 years, faster than any previous technology wave.
53% Global enterprise AI adoption
$581B Total AI investment (+130% YoY)
75M Daily AI Mode users
1.5B Monthly AI Overview users

The investment numbers match the adoption velocity. Total AI investment reached $581 billion, up 130% year-over-year. This isn't speculative venture funding — it's operational deployment budget from enterprises that have passed the pilot stage.

The AI agent success rate is perhaps the most under-discussed metric. Agent task completion jumped from 20% to 77% in 12 months. A year ago, AI agents failed 4 out of 5 tasks. Today they complete more than 3 out of 4. This connects directly to Google's agentic search reaching 75 million daily users — the agents are getting good enough to deploy at scale, and Google is doing exactly that.

Stanford AI Index MetricValueContext
Global enterprise adoption53%3 years from ChatGPT launch
AI agent success rate20% → 77%12-month improvement
Total AI investment$581B+130% year-over-year
AI transparency index58 → 40Declining (worse)
AI Mode daily users75MGoogle's agentic search
AI Overview monthly users1.5BGoogle's summarized results
AI Mode vs AI Overview citation overlap13%Different sources cited

One finding the Stanford report buries in the methodology section deserves front-page attention: AI Mode and AI Overviews share only 13% citation overlap. These are two Google products, built on the same index, answering similar queries — and they cite different sources 87% of the time. This means optimizing for AI Overviews does not automatically optimize for AI Mode. They are functionally separate channels that require separate strategies, separate monitoring, and separate content approaches. The 61% CTR collapse from AI Overviews in competitive verticals only tells half the story — the AI Mode impact is a separate calculation entirely.

Transparency is declining, not improving. The Stanford AI Transparency Index dropped from 58 to 40 (out of 100). Major AI companies are disclosing less about their training data, model architecture, and decision-making processes — even as adoption scales. This matters for SEO because it means we have less visibility into why AI systems cite certain sources and not others.


6. The Employment Signal Nobody Wants to Discuss

Labor market disruption

The Stanford report includes a data point that most AI coverage has sidestepped: junior developer job postings (ages 22-25) have declined approximately 20% since 2024. At the same time, demand for experienced developers has held steady or grown. The data covers US-based software development roles across major job boards.

This isn't about AI replacing developers. It's about AI changing the shape of the labor market. Entry-level tasks — boilerplate code, simple CRUD operations, basic testing — are increasingly handled by AI coding assistants. Companies still need senior developers to architect systems, review AI-generated code, and make judgment calls. But they need fewer junior developers to do the rote work that used to train those seniors.

The SEO parallel is obvious: routine technical audits, basic keyword research, and templated content creation are the "junior developer tasks" of our field. The practitioners who build strategic value — who understand the why behind the data — will grow in demand. The ones running checklists will face the same compression.


7. Actionable Synthesis — What to Do This Week

Combining the 68.9M crawler study with Stanford's adoption data, here's a prioritized hierarchy of actions ranked by data-supported impact.

Optimization hierarchy for AI search visibility
Action hierarchy — prioritize high-impact moves backed by the crawler study data.

Highest impact — do this week

  • Publish deep content consistently. The 33x visibility multiplier for sites with 50+ pages is the strongest signal in the study. Not thin pages — substantive, structured content that AI crawlers and users both find valuable. Aim for topical depth, not volume for volume's sake.
  • Sync your Google Business Profile. The 33.9 percentage point gap between synced and unsynced sites is the easiest win. If your GBP is stale or inconsistent with your website NAP data, fix it immediately.
  • Complete your local schema. Only 22.3% of sites have any local schema, and sites with 10-11 fields hit 82% AI crawl rate. Add LocalBusiness, PostalAddress, openingHours, geo coordinates, and all available fields — not just name and address.
  • Don't block AI crawlers. Unless you have specific licensing or IP concerns, the data shows no benefit to blocking AI crawlers and a clear cost. The 59% of sites receiving visits are building retrieval momentum. The 41% that aren't are invisible to an increasingly important discovery channel.

Medium impact — build into Q2 roadmap

  • Optimize for AI Mode AND AI Overviews separately. The 13% citation overlap means these are different ranking surfaces. Track which pages appear in AI Overviews vs AI Mode. Build content formats that serve each — AI Overviews favor concise, structured answers; AI Mode favors deep, navigable content.
  • Monitor LLM referrals by source. Set up UTM-level tracking or server log analysis to separate ChatGPT, Claude, Copilot, and Perplexity referral traffic. The 72.7% aggregate growth masks huge variance. Your site may over-index on one platform and miss another entirely.
  • Audit your crawlable content structure. AI crawlers allocate more budget to sites with clean, deep content hierarchies. Review your internal linking, sitemap coverage, and whether your key content is accessible without JavaScript rendering.

Skip for now — low data support

  • E-commerce-specific AI optimization. The -5.0pp correlation for e-commerce sites suggests AI crawlers currently deprioritize product catalog structures. Focus e-commerce AI efforts on informational content (buying guides, comparison pages) rather than product pages.
  • Perplexity-specific optimization. At 1.8% of AI crawl traffic and only 14.1% referral growth, Perplexity isn't generating enough volume or momentum to justify a dedicated optimization strategy. Monitor, but don't invest.
  • Gemini-specific optimization. At 0.6% of AI crawl traffic, Gemini's crawler activity is negligible. Google's AI search products (AI Overviews, AI Mode) pull from the traditional index, not from Gemini's crawl data, so Gemini crawler optimization has no practical impact on Google AI visibility.

The bottom line

The convergence of the 68.9M crawler study and Stanford's adoption data tells a clear story: AI is not a future channel — it's a current one, growing at 72.7% annually, with 1.5 billion monthly users already seeing AI-generated results. The sites that win are the ones that do the fundamentals well — deep content, complete structured data, clean architecture — because those are the same attributes that drive both AI crawler visibility and traditional search performance. There is no separate "AI SEO" playbook. There's just good SEO, applied consistently.


Related Articles

Frequently Asked Questions

Which AI crawler sends the most traffic to websites?

OpenAI dominates AI crawl traffic with an 81% market share — 55.8 million out of 68.9 million total AI crawler visits in the study. This includes both GPTBot (used for training and discovery) and ChatGPT-User (triggered by real-time user queries). Anthropic's ClaudeBot is a distant second at 16.6%, followed by Perplexity at 1.8% and Gemini at 0.6%.

How fast is LLM referral traffic growing?

Total LLM referral traffic grew 72.7% year-over-year across all platforms, from 93,484 to 161,469 measured sessions. ChatGPT leads in absolute volume with 136,095 sessions (+66.7%). The fastest-growing platforms by percentage are Copilot (from 22 to 9,560 sessions) and Claude (from 106 to 2,488 sessions, a 23x increase). Perplexity grew a more modest 14.1% despite its search-first positioning.

What is the biggest factor driving AI crawler visibility?

Content depth is the single strongest predictor. Sites with 50 or more published pages averaged 1,373.7 AI crawler visits, compared to just 41.6 for sites with fewer than 10 pages — a 33x multiplier. After content depth, the next strongest factors are local SEO signals: Yext sync (97.1% crawl rate), Google Business Profile sync (92.8% vs 58.9%), and local schema with 10-11 fields (82% vs 55.2%).

Does AI crawling directly improve human traffic and conversions?

The study found that AI-crawled sites average 3.2x more human sessions, 2.7x more form completions, and 2.5x more click-to-call actions. However, the researchers explicitly note this is correlation, not causation. The likely explanation is a shared root cause: well-built, content-rich, properly structured sites attract both AI crawlers and human visitors. Optimizing for one inherently optimizes for the other.

How fast is global AI adoption happening compared to previous technologies?

According to the Stanford 2026 AI Index, 53% of organizations globally adopted AI within 3 years of generative AI becoming widely available. For comparison, personal computers took approximately 15 years to reach 50% enterprise adoption, and the internet took about 7 years. AI adoption is occurring at roughly 2x the speed of internet adoption and 5x the speed of PC adoption.

What is the AI agent success rate and why does it matter?

AI agent task completion jumped from 20% to 77% in 12 months. This matters because it crosses the usability threshold — from failing 4 out of 5 tasks to completing more than 3 out of 4. This improvement is driving Google's deployment of AI Mode (75 million daily users) and enterprise adoption of agentic workflows. For SEO, it means AI agents are increasingly reliable enough to mediate real commercial transactions, not just answer informational queries.

Why does AI Mode vs AI Overviews having only 13% citation overlap matter?

Because it means optimizing for one Google AI product does not optimize for the other. AI Overviews (1.5 billion monthly users) and AI Mode (75 million daily users) cite different sources 87% of the time despite being built on the same index. This effectively doubles the optimization surface area — you need separate tracking, separate content strategies, and separate measurement for each. A page ranking well in AI Overviews has no guarantee of appearing in AI Mode results, and vice versa.

About the Author

Francisco Leon de Vivero

Francisco Leon de Vivero

VP of Growth at Growing Search

15+ years in enterprise, ecommerce, and international SEO. Former Head of Global SEO Framework at Shopify. Speaker at UnGagged and SEonthebeach. Now leading growth strategy at Growing Search.

LinkedIn · YouTube · Book a Consultation

Next step

Turn this background reading into a more current SEO plan.

Use the most relevant current page below if this topic is still on your roadmap, then review the proof and contact paths if you want direct support.

Current service page

Technical SEO Advisory

The goal is not audit sprawl. It is translating complex technical issues into prioritized actions that development and marketing teams can actually execute.

Explore this service