Back to SEO Glossary
SEO Definition

What Is Robots.txt?

Robots.txt is a simple text file placed in the root directory of your website that gives instructions to search engine crawlers about which pages or sections they are allowed or not allowed to access.

It acts as a guideline for bots, helping you control how they interact with your website.

The file is always located at:

https://yourwebsite.com/robots.txt

For example, if someone visits:
https://example.com/robots.txt

They will see the instructions written inside the file.

Search engines such as Google and Bing check this file before crawling your website. It’s part of what’s known as the Robots Exclusion Protocol.

How Robots.txt Works

When a search engine crawler visits your site, it looks for the robots.txt file before crawling any pages. The file contains directives like "allow" or "disallow" that specify which URLs or directories the crawler can access.

If you disallow a URL in robots.txt, crawlers will not visit that page. However, this does not prevent the page from being indexed if other sites link to it. To block indexing, you need a noindex tag, not just robots.txt.

Why Robots.txt Is Important for SEO

Robots.txt plays a technical SEO role by:

  • Controlling crawl access – Prevents bots from crawling unnecessary pages.
  • Optimizing crawl budget – Helps search engines focus on important pages.
  • Blocking sensitive areas – Stops crawlers from accessing admin pages, test environments, or duplicate content.
  • Specifying sitemap location – You can include your sitemap URL inside robots.txt.

However, it’s important to understand:

Robots.txt controls crawling — not indexing.

If a page is blocked but linked elsewhere, it may still appear in search results (without content). For full removal, you’d use a noindex directive instead.

Common Directives Explained

1. User-agent

Specifies which crawler the rule applies to.

Example:

User-agent: Googlebot

Targets only Google’s crawler.

2. Disallow

Blocks access to specific pages or folders.

Example:

Disallow: /checkout/

3. Allow

Used to permit access to specific pages inside a blocked folder.

4. Sitemap

Helps search engines discover all important URLs on your site.

When Should You Use Robots.txt?

You should use robots.txt to:

  • Block duplicate
  • Prevent crawling of filter
  • Hide staging or development
  • Protect system directories

But you should not use it to hide confidential data — because anyone can still view the robots.txt file publicly.

Final Thoughts

Robots.txt is a small file with a big impact. When used correctly, it improves crawl efficiency, protects sensitive areas, and supports better SEO performance. But when misconfigured, it can seriously harm visibility.

Need help with SEO?

Understanding terms is the first step. If you're looking for help with actual execution that drives results, let's talk.

Get in touch

Recommended Reading

Next Term: What is Schema Markup?

Read Next