Robots.txt Generator
DocsGenerate a custom robots.txt file to control how search engines crawl your website. Specify which pages to allow or disallow for different user agents.
Robots.txt Guide
Common Directives:
- User-agent: Specifies which crawler
- Disallow: Blocks access to paths
- Allow: Permits access (overrides Disallow)
- Sitemap: Points to XML sitemap
- Crawl-delay: Sets request frequency
Common Patterns:
/admin/ - Block admin areas/wp-admin/ - Block WordPress admin/*.pdf$ - Block PDF files/search? - Block search pages/cart/ - Block shopping cart
⚠️ Important: Robots.txt is publicly accessible. Don't use it to hide sensitive content - use proper authentication instead.
Validation
After uploading your robots.txt file, test it using: