The ClaudeBot Directive: Two Lines That Change Your Score
Claude applies a trust penalty when ClaudeBot isn't in your robots.txt - even if you allow every other AI crawler. Sites with explicit ClaudeBot directives scored 8 points higher on Claude across our 20-site cohort.
Part of the AEO scoring framework - the current 48 criteria that measure how ready a website is for AI-driven search across ChatGPT, Claude, Perplexity, and Google AIO.
Quick Answer
Claude penalizes sites that don't explicitly mention ClaudeBot in robots.txt - even if GPTBot and PerplexityBot are welcomed. Explicitly allowing ClaudeBot signals intent to be cited by Claude. Across our 20-site audit cohort, sites with ClaudeBot directives scored an average of 8 points higher on Claude. That's one of the highest ROI optimizations we've found: two lines of text, measurable scoring improvement.
Audit Note
In our audits, we've measured The ClaudeBot Directive: Two Lines That Change Your Score on live sites, we've compared implementations, and we've audited the...
How do I add ClaudeBot to my robots.txt file?
Here's what Claude actually sees when it reads your robots.txt - and it's looking for its own name.
Does allowing ClaudeBot in robots.txt improve my Claude citation score?
GPTBot is everywhere - millions of robots.txt files mention it.
What is the difference between ClaudeBot and GPTBot in robots.txt?
Tidio.com included explicit ClaudeBot Allow directives alongside GPTBot and PerplexityBot.
Summarize This Article With AI
Open this article in your preferred AI engine for an instant summary and analysis.
Before & After
Before - No AI crawler directives
User-agent: * Allow: / Sitemap: https://acme.com/sitemap.xml
After - Explicit ClaudeBot directive
User-agent: ClaudeBot Allow: / Crawl-delay: 2 User-agent: anthropic-ai Allow: / Crawl-delay: 2 Sitemap: https://acme.com/sitemap.xml
How Does Claude Read Your robots.txt for Trust Signals?
Here's what Claude actually sees when it reads your robots.txt - and it's looking for its own name.
ClaudeBot checks robots.txt before touching any page. But Claude goes beyond basic access control. When it finds an explicit "User-agent: ClaudeBot / Allow: /" directive, it reads that as a deliberate policy decision - you've actively chosen to participate in Claude's content ecosystem. That's a governance signal.
No ClaudeBot directive? Ambiguity. If your robots.txt mentions GPTBot, PerplexityBot, and Google-Extended but skips ClaudeBot, Claude reads that as either oversight or intentional exclusion. Either way, the trust signal weakens. Claude still follows your general "User-agent: *" rules, but you've lost the bonus from explicit acknowledgment.
Granularity matters too. Full domain access ("/") generates a stronger trust signal than directory-restricted access. Crawl-delay directives? Respected without penalty - they signal thoughtful configuration, not hostility. But "Disallow: /" for ClaudeBot is an active block. That significantly reduces Claude's willingness to cite you, even from cached content.
Claude also checks for consistency. Some sites include directives for "anthropic-ai" but not "ClaudeBot." Claude recognizes both but weights the ClaudeBot-specific entry more heavily - it demonstrates awareness of the exact crawler name. Beyond robots.txt, Claude looks for complementary signals: does your llms.txt reference Anthropic? Does your content policy address AI citation? These signals compound into a holistic trust profile.
Why Does a ClaudeBot Directive Matter More Than GPTBot?
GPTBot is everywhere - millions of robots.txt files mention it. For ChatGPT, the absence of a GPTBot directive is the default state. Minimal penalty.
Claude operates in a different universe. ClaudeBot is less widely recognized than GPTBot. Sites that include ClaudeBot directives are self-selecting for above-average governance awareness. Claude factors this in.
Google AI Overviews doesn't use a separate AI crawler the same way. Google-Extended controls Gemini training data, but Googlebot builds the search index regardless. Different strategic game entirely.
The culprit behind the scoring gap: Claude's trust model is more sensitive to robots.txt signals than ChatGPT's. Adding ClaudeBot directives produces a larger marginal improvement on Claude than adding GPTBot directives produces on ChatGPT - because the relative signal value is higher for the less common directive.
Perplexity sits in the middle. PerplexityBot directives are less common than GPTBot but more common than ClaudeBot, so the trust signal carries moderate weight. Sites targeting multi-engine visibility should address all four major crawlers (GPTBot, ClaudeBot, PerplexityBot, Google-Extended) to max out trust across every platform.
How Does the ClaudeBot Directive Affect Real Scores?
Tidio.com included explicit ClaudeBot Allow directives alongside GPTBot and PerplexityBot. Comprehensive AI crawler policy. Different crawl-delay values per bot, with ClaudeBot getting the same generous access as GPTBot. Claude read that as a deliberate governance decision. Tidio's Claude bonus: +14.
LiveChat.com addressed ClaudeBot with "Allow: /" and a 2-second crawl-delay. Here's what made it notable: LiveChat is a ChatGPT partner - yet they chose to also welcome Claude's crawler. Combined with strong Organization schema and semantic HTML, that contributed to LiveChat's +12 bonus.
HelpSquad.com? Zero AI-specific crawler directives. No GPTBot, no ClaudeBot, no PerplexityBot. That hurt everywhere, but it hit Claude hardest. HelpSquad scored 42 on Claude vs. 47 on ChatGPT - a -5 penalty. ChatGPT's wider default crawling corpus partially compensated. Claude's governance-first approach didn't forgive the absence.
Crisp.chat mentioned GPTBot but not ClaudeBot. Despite the partial omission, Crisp still pulled a +17 Claude bonus - primarily from other governance signals (schema depth, semantic HTML). But adding a ClaudeBot directive would've pushed the bonus higher. The directive is additive to other signals, not a prerequisite.
Across our full 20-site cohort, sites with explicit ClaudeBot directives scored an average of 8 points higher on Claude - controlling for other variables. Two lines in robots.txt. Eight points. That's one of the most cost-effective optimizations in AEO.
How Do You Add a ClaudeBot Directive to robots.txt?
Start here: add ClaudeBot directives to your robots.txt right now. Two minutes, outsized impact. The minimum configuration is three lines: "User-agent: ClaudeBot", "Allow: /", and "Crawl-delay: 2". Put them in a clearly labeled section alongside other AI crawler policies.
Include both "ClaudeBot" and "anthropic-ai" as separate user-agent entries. While Claude primarily identifies as ClaudeBot, some docs reference anthropic-ai as the crawler identifier. Cover both. Use the same Allow/Disallow rules for consistency.
Set crawl-delay between 1 and 5 seconds. Claude respects it and doesn't interpret rate limiting as hostility. Two seconds is the sweet spot - prevents server load without slowing Claude's indexing. Don't go above 10 seconds - very high delays may cause Claude to deprioritize your site in its crawling queue.
Need to restrict certain directories? Use targeted Disallow rules. "Disallow: /admin/" and "Disallow: /checkout/" while keeping "Allow: /" for public content. That demonstrates sophisticated governance - you've thought about what should and shouldn't be accessible to AI crawlers.
Verify your robots.txt after changes. Common mistakes: file in a subdirectory instead of domain root, incorrect line endings causing parse failures, UTF-8 BOM characters corrupting the file. Test by requesting your-domain.com/robots.txt directly. Confirm the ClaudeBot section renders as expected. Add "Sitemap: your-domain.com/sitemap.xml" at the bottom so all crawlers discover your full content inventory.
Where Can You Learn More About ClaudeBot Directives?
- robotstxt.org - www.robotstxt.org
- Anthropic ClaudeBot Crawler Policy - support.anthropic.com/en/articles/8896518-does-anthropic-crawl-data-from-the-web-and-how-can-site-owners-block-the-crawler
Key Takeaways
- Add explicit ClaudeBot Allow directives to robots.txt - two lines, measurable scoring lift.
- Include both "ClaudeBot" and "anthropic-ai" as separate user-agent entries for full coverage.
- Set crawl-delay between 1-5 seconds - Claude respects it without penalty.
- Sites with ClaudeBot directives scored an average of 8 points higher on Claude across our cohort.
How does your site score on this criterion?
Get a free AEO audit and see where you stand across all 34 criteria.