Medium — Crawl Access

Blocked by robots.txt

Your robots.txt file is preventing Google (and other crawlers) from accessing pages you want indexed. This is one of the simplest GSC errors to fix — but one of the most costly when misconfigured, because blocked pages can never rank.

Tests robots.txt, meta robots, and bot access for 10+ crawlers. No account required.

Common robots.txt Mistakes

How to Fix It

  1. Audit your robots.txt. Use GSC's robots.txt tester to verify which URLs are blocked.
  2. Use SEODiff's Crawl Access Checker to test access for Googlebot, GPTBot, ClaudeBot, and other crawlers simultaneously.
  3. Be specific. Instead of blocking entire directories, use precise path patterns.
  4. Allow CSS/JS/images. Never block /static/, /assets/, or /wp-content/themes/.
  5. Test before deploying. Always test robots.txt changes against your URL list before pushing to production.

robots.txt vs meta robots vs noindex

These are different mechanisms with different effects:

Important: If you block a page in robots.txt AND have a noindex meta tag, Google will never see the noindex tag (because it can't crawl the page). The noindex is effectively invisible.

Don't accidentally block your best pages

SEODiff tests crawl access for Google, AI bots, and LLM crawlers in one scan.

Check Crawl Access →

Related Diagnostics