Robots.txt Generator

Generate robots.txt files online. Configure user-agents, allow/disallow paths, sitemap URL, and crawl-delay. Presets for common configurations.

Common presets

User-agent

Allow / Disallow paths

Sitemap & Crawl-delay

Preview

User-agent: all
Disallow: /

What Is robots.txt?

robots.txt is a standard file used by websites to communicate with web crawlers. Placed in the root directory (e.g. https://yoursite.com/robots.txt), it tells search engines which parts of your site they can crawl and which to avoid.

Common Use Cases

Use robots.txt to block admin panels, staging areas, or API endpoints from being indexed. Allow all crawlers to access public content while blocking specific folders. Include your sitemap URL so search engines can discover all pages efficiently.

Best Practices

Always include a Sitemap directive. Use Disallow for paths you want to exclude; use Allow to override within a Disallow block. Test your robots.txt in Google Search Console. Remember that robots.txt does not hide pages from indexing if they are linked from other sites — use noindex for that.

Frequently Asked Questions

Related Tools

Explore More Tools

Find this tool useful? Buy us a coffee to keep DuskTools free and ad-light.