Common presets
User-agent
Allow / Disallow paths
Sitemap & Crawl-delay
Preview
User-agent: all Disallow: /
What Is robots.txt?
robots.txt is a standard file used by websites to communicate with web crawlers. Placed in the root directory (e.g. https://yoursite.com/robots.txt), it tells search engines which parts of your site they can crawl and which to avoid.
Common Use Cases
Use robots.txt to block admin panels, staging areas, or API endpoints from being indexed. Allow all crawlers to access public content while blocking specific folders. Include your sitemap URL so search engines can discover all pages efficiently.
Best Practices
Always include a Sitemap directive. Use Disallow for paths you want to exclude; use Allow to override within a Disallow block. Test your robots.txt in Google Search Console. Remember that robots.txt does not hide pages from indexing if they are linked from other sites — use noindex for that.
Frequently Asked Questions
Related Tools
Explore More Tools
Find this tool useful? Buy us a coffee to keep DuskTools free and ad-light.