Agntbase
AI bot analytics

GPTBot is not a lead

AI crawlers may already be reading your website, but ordinary analytics often shows nothing. A visit from GPTBot, ClaudeBot, PerplexityBot, or another crawler is not the same thing as a customer visit, a search referral, or a recommendation.

Short version: AI visibility work needs two layers: machine-readable entrypoints for bots, and server-side analytics that proves which bots actually received a clean response.

The crawl-to-referral gap

The uncomfortable part of AI discovery is that many crawlers take far more pages than they send back as traffic. SEOmator's 2026 analysis of Cloudflare Radar data describes this as a crawl-to-refer ratio: pages crawled divided by referrals returned to the website.

The exact ratios differ by data source and time window, but the business lesson is stable: a crawler reading your site is not proof that buyers are arriving. You need to separate training crawls, search indexing, and user-triggered fetches.

Different AI bots mean different things

GPTBot

OpenAI crawler used for content that may help improve foundation models. It is not the main signal for ChatGPT Search traffic.

OAI-SearchBot

OpenAI search crawler used to surface sites in ChatGPT search features. This is the bot to watch for ChatGPT Search discoverability.

ChatGPT-User

User-triggered access. It may appear when a real ChatGPT user asks a question that requires a live page.

ClaudeBot

Anthropic crawler for public web content that may contribute to model development.

Claude-SearchBot

Anthropic crawler for search-related indexing inside Claude. Blocking it can reduce visibility in Claude search scenarios.

PerplexityBot

Perplexity's crawler for search-style indexing. Its traffic should be measured separately from referral visits.

Why GA4 and browser scripts miss the picture

Most AI crawlers do not behave like human visitors. They usually do not run your analytics JavaScript, click buttons, open forms, or behave like a normal browser session. If you only look at client-side analytics, you may conclude that nothing happened while your server logs show thousands of bot reads.

That is why Agntbase treats AI traffic as a server-side event stream. A useful dashboard should show the bot family, requested path, status code, response class, and whether the request hit an agent-ready entrypoint.

What to measure

Access

Did the bot get 200?

Track status classes: 2xx, 3xx, 4xx, 5xx, and edge cases like 499 or reset connections.

Intent

Which bot was it?

Separate training bots from search bots and user-triggered fetchers.

Entrypoints

What did it read?

Watch reads of /llms.txt, /agenthub.json, /profile.json, and canonical profile JSON.

Outcome

Did people arrive?

Track referrals and UTM data separately. Bot reads are infrastructure signals, not revenue by themselves.

The practical Agntbase position

The goal is not to trick AI systems. The goal is to make the business readable, measurable, and diagnosable.

Sources and context

This article is an English adaptation and product-positioning version of a Russian vc.ru note, rewritten for Agntbase.com and expanded with direct crawler documentation.

  1. Original Russian note on vc.ru
  2. SEOmator analysis of Cloudflare Radar crawl-to-refer data
  3. Cloudflare AI Crawl Control / AI Audit
  4. OpenAI crawler documentation
  5. Anthropic crawler documentation
  6. Perplexity robots.txt documentation