robots.txt (Developer Guide)
robots.txt is the gatekeeper of crawling. Done right, it reduces crawl waste; done wrong, it can block your entire site.
1) robots.txt is not noindex
- Disallow blocks crawling
- noindex blocks indexing (meta robots / X-Robots-Tag)
2) A practical template
text
User-agent: *
Disallow: /admin
Disallow: /api
Allow: /assets/
Sitemap: https://example.com/sitemap.xmlAvoid blocking important CSS/JS—Google may fail to render content correctly.
3) Fast workflow with our generator
- Open robots.txt Generator and pick paths to block.
- Review for accidental blocks (assets, important pages).
- Publish it at `/robots.txt` on your domain.
More: Indexability checklist
Related tools
FAQ
Common robots.txt mistakes and misconceptions.