Free Tool

Robots.txt Analyzer

Check which crawlers are allowed or blocked by your robots.txt file. See rules for Googlebot, GPTBot, ClaudeBot, and 20+ other bots.

Robots.txt Examples

Copy these templates to control how crawlers access your site.

Allow All Crawlers

Best for most websites wanting maximum visibility

User-agent: *
Allow: /

Sitemap: https://example.com/sitemap.xml

Block AI Training Only

Allow search engines but block AI training crawlers

User-agent: *
Allow: /

User-agent: GPTBot
Disallow: /

User-agent: Google-Extended
Disallow: /

User-agent: CCBot
Disallow: /

Sitemap: https://example.com/sitemap.xml

Allow Search + AI Citations

Be indexed by search engines and cited by AI assistants

User-agent: *
Allow: /

# Allow AI assistants to cite your content
User-agent: ChatGPT-User
Allow: /

User-agent: ClaudeBot
Allow: /

User-agent: PerplexityBot
Allow: /

Sitemap: https://example.com/sitemap.xml

Block Specific Paths

Allow most content but protect admin/private areas

User-agent: *
Allow: /
Disallow: /admin/
Disallow: /api/
Disallow: /private/
Disallow: /*.json$

Sitemap: https://example.com/sitemap.xml
For AI-built JavaScript Sites

Allowing crawlers is just step one.

If your site is built with React, Vue, or Angular, crawlers may see an empty page even with robots.txt configured correctly. LovableHTML pre-renders your JavaScript into static HTML that every crawler can read.

  • Get indexed by Google in hours, not weeks
  • Be cited by ChatGPT, Claude & Perplexity
  • No code changes required
See Live Demo

Robots.txt FAQ