Create custom robots.txt files to control search engine crawling and indexing of your website.
Robots.txt is a text file in your site's root directory that instructs search engine crawlers which pages they should or shouldn't access. It's a critical SEO tool for controlling how search engines interact with your site.
User-agent: Specifies which crawler the rules apply to (* = all)
Disallow: Paths that should not be crawled
Allow: Overrides Disallow for specific paths
Sitemap: Location of your XML sitemap
Crawl-delay: Seconds between requests
Block duplicate content: Disallow: /print/
Block search results: Disallow: /*?s=
Block specific bots: User-agent: BadBot + Disallow: /
Allow everything: User-agent: * + Disallow: (empty)
No, but highly recommended for better crawl control.
Robots.txt is not security—use actual authentication for sensitive content.
User-agent: * and Disallow: /