robots.txt Validator

Audit crawl directives, AI bot coverage, and sitemap health

🔒 Runs 100% in your browser — your files never leave your device

🕐 Cron Builder & Visualiser Visualize job overlaps 🔒 SSL Checker Check cert expiry 🐳 Docker Auditor Audit compose files 🛡️ Firewall Auditor Audit UFW rules 🔀 Reverse Proxy Mapper Map routing & audit 🤖 robots.txt Audit crawl & AI bots

Enter a URL or paste your robots.txt, then run the audit.

Frequently Asked Questions

Does this fetch my robots.txt live?

Yes in URL mode. Nothing is stored — fetch happens in your browser.

Why does AI bot coverage matter?

GPTBot, ClaudeBot, and CCBot index content for AI training. If you haven't explicitly addressed them, they may crawl freely.

Does blocking GPTBot actually stop AI training?

It signals your preference to OpenAI's crawler. GPTBot has stated it respects robots.txt exclusions for training data. ClaudeBot and Google-Extended also honour explicit blocks. CCBot compliance is less consistent. Blocking in robots.txt is the standard opt-out mechanism but cannot be technically enforced.

What's the file size limit?

Google recommends robots.txt stay under 500KB. Larger files may be partially ignored.

Does Disallow: / block everything?

For User-agent: * yes. But individual bots can have their own rules that override the wildcard.

What are unknown directives?

Lines that aren't part of the robots.txt standard (RFC 9309). Some like Host: are common but not universally supported.