Is your CDN the key to growing in AI Search? A study on how AI agents crawl and cite your content
Most digital marketing teams track where their human visitors come from. They monitor organic sessions, referral sources, bounce rates. What far fewer teams track is an entirely separate category of traffic that's been growing aggressively since 2024 — and that directly determines whether AI models will recommend their brand in a conversation.
That traffic comes from AI crawlers. And your CDN logs are where most of the evidence lives.
Your content can't be cited if it's never been read
Generative Engine Optimization (GEO) is built on a simple premise: AI search visibility depends on whether AI models can access, parse, and index your content in the first place. Before ChatGPT or Perplexity can cite a page, a crawler has to reach it, download it, and process it.
According to Cloudflare's July 2025 analysis, total crawler traffic rose 18% year-over-year between May 2024 and May 2025, with GPTBot alone growing 305% in that same window. By April 2026, a separate log-file study published by PPC Land covering 7 billion OpenAI bot events found a 3.5x surge in OAI-SearchBot activity following the GPT-5 release. These crawlers are not small visitors. They are among the most active agents on the web — and standard analytics tools like GA4 don't register them at all.
If you're only watching visitor analytics, you're looking at half the picture.
Not all AI crawlers are doing the same thing
This is where most teams get tripped up. They hear "AI crawler" and assume it's a monolithic process. It isn't. Each major AI model sends distinct crawlers for distinct purposes, and understanding that distinction changes what you optimize for.
Take OpenAI as an example. GPTBot is the training crawler — it scrapes content at high frequency to build and refine the underlying model. ChatGPT-User, by contrast, is activated only when a real person triggers a web-browsing query inside ChatGPT. It retrieves live information to answer that specific question. Anthropic, Perplexity, Google, and others each follow similar two-tier patterns: an indexing bot that caches your content for model training, and a retrieval bot that fetches it in real time during a user conversation.
This creates a measurable gap. Cloudflare's August 2025 "crawl-to-click" research found that training crawling now accounts for nearly 80% of all AI bot activity. Your pages may be getting crawled thousands of times with zero corresponding user visits — unless the model has also decided to cite you when answering user questions.
For GEO performance metrics, this means you need two data streams: server-side crawler logs that show indexing behavior, and visitor analytics that show when citations translate into actual human traffic.
Why your CDN is the most reliable data source
Server-side CDN logs are the only place where AI crawler activity is fully visible. Tools like Cloudflare, Akamai, AWS CloudFront, and Google Cloud CDN capture every HTTP request, including user-agent strings like GPTBot/1.1, ClaudeBot, PerplexityBot, and anthropic-ai. These requests never show up in JavaScript-based analytics because crawlers don't execute the tracking scripts that fire those events.
The CDN log, then, becomes your source of truth for three things:
- Which pages AI crawlers are actually visiting — not which ones you think they should be visiting
- How frequently each crawler returns — crawl cadence is a strong signal of how much a model values that content
- Which crawlers are active vs. absent — a page not visited by PerplexityBot will not appear in Perplexity responses, regardless of how well-written it is
Crawl frequency matters beyond just presence. The more active a crawler is on your site, the stronger the signal that the model is pulling from your content regularly. Pages with higher crawl rates from citation-oriented bots like ChatGPT-User and anthropic-ai correlate with higher citation rates in live AI responses. That relationship is worth measuring directly.
Combining crawler logs and visitor analytics closes the loop
Crawler data alone tells you what AI agents are reading. Visitor analytics tells you what's converting into human traffic. The signal that matters is the relationship between the two.
A page that gets crawled 400 times by GPTBot but generates no chatgpt.com referral traffic might have an indexing-to-citation problem. The model has accessed the content but isn't deploying it in responses. A page with low crawl frequency but high AI referral traffic is likely getting cited from cached training data rather than fresh retrieval — which means its content could be outdated in model responses without you knowing.
Bridging these two data streams is where platforms with combined infrastructure make the biggest difference. Promptwatch's Agent Analytics connects CDN-sourced crawler logs directly to citation monitoring and visitor traffic data, so marketing teams can identify which pages are crawled but never cited, which are cited but never visited, and which are driving the full loop from model indexing to user click. You can connect logs from Cloudflare Workers, AWS CloudFront, Google Cloud CDN, and Akamai directly, with the platform filtering out non-AI traffic automatically.
Compared to platforms that only surface AI-referred referral traffic from tools like GA4, or only parse log files without connecting them to citation data, this unified view substantially reduces the time from "our content exists" to "our content is influencing AI responses."
What the data patterns actually tell you
When you have both signals together, certain patterns become immediately actionable:
High crawl frequency, low citation rate: The model is reading your pages but not recommending them. This points to a content quality or authority issue — the crawler indexes you, but the model's ranking layer deprioritizes the content in generated responses. A content gap analysis will often reveal the missing context.
Low crawl frequency, moderate citation rate: You're being cited from stale training data. The model has older versions of your content in its weights. New pages or updates aren't being retrieved because the crawler hasn't revisited. Improving crawlability (ensuring no robots.txt blocks, correct structured data, fast CDN response times) directly addresses this.
High citation rate, low visitor volume: The model is citing you, but users aren't clicking through. This typically indicates you're appearing in conversational contexts where the AI summarizes your content without linking out — common in instructional or factual queries. Optimizing for AI-referred traffic means writing content that benefits from a click to fully answer the question.
The crawlers have preferences — and they're measurable
Different AI models treat different types of content differently. According to the LLM crawler user agent directory, there are now 30+ distinct AI crawlers actively indexing the web. Each has crawl patterns that reflect the retrieval architecture of the model it serves.
PerplexityBot, for example, is designed around real-time retrieval-augmented generation. It tends to revisit pages frequently and favors content with clear factual structures, direct answers, and up-to-date information. GPTBot, used primarily for training, has broader coverage but is less temporally sensitive. anthropic-ai shows strong affinity for well-structured long-form content with clear semantic organization.
These preferences are legible in log data. Segments of your site that attract disproportionate attention from citation-oriented bots are content worth doubling down on. Segments that are only attracting training crawlers but no retrieval bots have a structural visibility problem that no amount of prompt testing will fix on its own.
Closing the gap between indexing and citation
The fundamental insight here is that AI search analytics requires a server-side layer that most teams haven't built yet. Your CDN is already collecting the data. The question is whether you're connecting it to anything that makes it interpretable alongside your citation and traffic data.
For content and SEO teams starting this process, the most practical first step is to verify that the major AI crawlers — GPTBot, ClaudeBot, PerplexityBot, Google's AI crawlers — are actually reaching your key pages. If they aren't, no amount of content optimization changes what AI models know about you. Once crawl access is confirmed, correlating crawl frequency with citation rate and referral traffic gives you the feedback loop needed to prioritize which content to update, expand, or create.
Results will vary by site, CDN provider, and content type. Connect your site to Promptwatch to see your own data.




