Skip to main content

Robots.txt Analyzer

Validate and optimize your robots.txt file for search engines. Get instant feedback on blocking rules, syntax errors, and SEO recommendations.

Analyzing your robots.txt file...

Choose Your Input Method

Enter Website URL

Just enter the domain (e.g., example.com) - we'll automatically fetch /robots.txt

Paste robots.txt Content

Paste your robots.txt file content

Health Score

robots.txt health score
rules ยท sitemap(s) ยท critical issue(s)
Total Lines
Total Rules
User Agents
Allow Rules
Disallow Rules
Sitemaps

About this tool

The Robots.txt Analyzer checks your robots.txt file for common problems and optimization opportunities. Here's what it looks for:

You'll get a health score from 0 to 100 based on how well your file is configured, along with categorized issues, detailed rule breakdowns, and tips for improvement. This helps you avoid accidentally blocking search engines while still keeping sensitive areas of your site protected.

Perfect robots.txt Example

A well-configured robots.txt file should be clear, purposeful, and avoid common pitfalls. Here's an example of a properly structured robots.txt that follows best practices:

# Allow all search engines to crawl the entire site
User-agent: *
Disallow:

# Block sensitive areas from all crawlers
Disallow: /admin/
Disallow: /api/
Disallow: /private/

# Allow access to CSS and JavaScript (important for SEO)
Allow: /css/
Allow: /js/
Allow: /assets/

# Specific rules for GPTBot (OpenAI's crawler)
User-agent: GPTBot
Disallow: /api/
Allow: /

# Sitemap location (helps search engines discover your content)
Sitemap: https://example.com/sitemap.xml

# Optional: Crawl delay for aggressive bots (use sparingly)
User-agent: *
Crawl-delay: 1

Key principles demonstrated:

Frequently Asked Questions

What is robots.txt?

robots.txt is a text file that tells search engine crawlers which pages they can and cannot access on your website. It's placed in your website's root directory (e.g., https://example.com/robots.txt) and follows the Robots Exclusion Protocol. This file helps you control how search engines interact with your content.

Why should I analyze my robots.txt file?

A misconfigured robots.txt can accidentally block search engines from indexing your entire site, preventing it from appearing in search results. Common issues include blocking CSS/JavaScript files (which affects SEO), overly restrictive rules, syntax errors, and missing sitemap declarations. Regular analysis helps catch these problems before they impact your visibility.

What does the health score mean?

The health score (0-100) reflects your robots.txt configuration quality:

  • 90-100 (Excellent): Well-configured with no critical issues
  • 70-89 (Good): Minor improvements possible but generally healthy
  • 50-69 (Needs Improvement): Several issues that should be addressed
  • 0-49 (Critical Issues): Serious problems blocking search engines
What are critical issues?

Critical issues prevent search engines from indexing your site properly. Examples include:

  • Disallow: / - Blocks your entire website from all search engines
  • Blocking homepage - Prevents indexing of your main page
  • No User-agent directives - Invalid robots.txt structure

These should be fixed immediately to maintain search visibility.

Should I block CSS and JavaScript files?

No. Google and other modern search engines need to render your pages to understand content and usability. Blocking CSS/JavaScript files can hurt your SEO by preventing proper page rendering. Remove rules like Disallow: /css/ or Disallow: /js/ from your robots.txt.

What is crawl-delay and should I use it?

Crawl-delay tells bots to wait a specified number of seconds between requests. While it can reduce server load, values above 10 seconds may significantly slow down indexing. Most modern search engines ignore this directive in favor of automatic rate limiting. Use with caution and keep values low (โ‰ค5 seconds) if needed.

Why am I seeing old results after updating my robots.txt?

Results are cached for 5 minutes to ensure fast response times. If you just made changes to your robots.txt file, wait a few minutes and analyze again to see the updated results.