Tool: Crawler Health

Tests how well 5 major AI bots can access your site — checking robots.txt, WAF blocks, and JavaScript rendering.

What it does

Crawler Health simulates access from 5 AI bot user agents and produces a per-bot status report plus an overall health score.

Bots tested

BotUser Agent StringOperator
GPTBotGPTBot/1.0OpenAI
ClaudeBotClaudeBot/1.0Anthropic
PerplexityBotPerplexityBot/1.0Perplexity
GoogleBotGooglebot/2.1Google
ApplebotApplebot/0.1Apple

How it works

Step 1: Robots.txt check

Fetches and parses /robots.txt to check if each bot's user agent is disallowed. This uses standard robots.txt parsing rules.

Step 2: HTTP fetch with UA spoofing (Pro)

On Pro plans, the tool makes an actual HTTP request using each bot's user agent string. This detects WAF/CAPTCHA blocks that only appear for specific user agents but allow browsers.

Step 3: JavaScript rendering heuristic

Checks if the page uses heavy JavaScript frameworks that would require a rendering engine that most AI bots don't have.

Scoring

ComponentScore Impact
Base score100
Per blocked bot−15
WAF/CAPTCHA detected−10
Heavy JS (client rendering)−20

This feeds into the Bot Access Score of the AI Readiness radar.

JSON output fields

API endpoint

GET /api/crawler-health?url=https://example.com