Your robots.txt file tells search engine crawlers which parts of your site to access and which to skip.
Use this free tool to build a clean, correctly formatted file in seconds: set crawl rules, block sensitive directories, specify a sitemap,
and control individual bots — all without touching a line of code.
The path is relative to root and must contain a trailing slash "/"
Manage which search engines can access your content
Set crawl delays and protect server resources
Block access to private directories and files
A robots.txt file is a plain-text file placed at the root of your domain (yourdomain.com/robots.txt). It uses the robots exclusion protocol to communicate crawl instructions to bots before they visit your site.
It's a foundational part of technical SEO because it directly affects the following:
Example location:
https://yourdomain.com/robots.txt
Important: robots.txt controls crawling, not indexing. A page blocked in robots.txt can still appear in search results if other sites link to it. Use a noindex meta tag or X-Robots-Tag header to prevent indexing.
The tool above handles the syntax automatically. Here's how to use it:
Just select your settings and click Generate — no coding required.
Marcitors Free Robots.txt Generator — Key Features
No coding needed. Use our Free robots.txt generator to produce a fully valid file with one click.
Whether you’re an SEO beginner or developer, our robots.txt builder and robots txt maker give you total control.
Create rules such as:
Optimized for Googlebot.
We also support rules for Bingbot, Yandex, AhrefsBot, SemrushBot, and custom crawlers.
Add your sitemap automatically using best SEO robots txt practices.
Use our tool to validate whether your robots.txt is correctly formatted.
Click Generate to build your robots.txt file.
Upload it to:
https://marcitors.com/free-tools/robots-txt-generator
In less than 5 seconds, you’ll have a clean file—whether you searched for Free robots.txt generator, create robots.txt generator, or robots disallow rules.
Perfect for users looking for robots txt example, robots txt file example, or sample robots txt:
User-agent: *
Disallow: /admin/
Disallow: /private/
Allow: /
Sitemap: https://yourwebsite.com/sitemap.xml
Create an optimized robots.txt file for Shopify stores to manage crawling and prevent duplicate content issues from filters and search pages. While Shopify provides a default file, customizing it improves SEO performance.
Example:
User-agent: *
Disallow: /search
Disallow: /cart
Disallow: /checkout
Disallow: /collections/*?*
Disallow: /*?variant=
Sitemap: https://yourstore.com/sitemap.xml
Use our free tool to generate a fully customized Shopify robots.txt file in seconds.
WooCommerce sites often create duplicate URLs due to filters, tags, and cart pages. A clean robots.txt file helps search engines focus on important pages.
Example:
User-agent: *
Disallow: /cart/
Disallow: /checkout/
Disallow: /my-account/
Disallow: /?s=
Disallow: /tag/
Sitemap: https://yourstore.com/sitemap.xml
Create a SEO-friendly robots.txt for WooCommerce with our free generator.
Magento stores can generate thousands of URLs, making crawl optimization essential. A proper robots.txt file prevents indexing of unnecessary pages.
Example:
User-agent: *
Disallow: /checkout/
Disallow: /catalogsearch/
Disallow: /customer/
Disallow: /review/
Disallow: /*?SID=
Sitemap: https://yourstore.com/sitemap.xml
Generate a powerful robots.txt file tailored for Magento stores.
Small business websites need a simple robots.txt file to guide search engines toward key pages like services and contact sections.
Example:
User-agent: *
Disallow: /admin/
Disallow: /login/
Disallow: /thank-you/
Sitemap: https://yourwebsite.com/sitemap.xml
Easily create a clean and effective robots.txt file for your website.
Blogs often have tags, categories, and archives that can create duplicate content. A robots.txt file helps manage crawl efficiency.
Example:
User-agent: *
Disallow: /wp-admin/
Allow: /wp-admin/admin-ajax.php
Disallow: /tag/
Disallow: /category/*?*
Sitemap: https://yourblog.com/sitemap.xml
Generate a robots.txt file that improves your blog’s SEO performance.
News websites publish content frequently, so controlling crawl behavior is critical for fast indexing and avoiding duplicate pages.
Example:
User-agent: *
Disallow: /search/
Disallow: /archive/
Disallow: /tag/
Sitemap: https://yournewswebsite.com/sitemap.xml
Create a fast and optimized robots.txt file for your news site.
robots.txt disallow all
User-agent: *
Disallow: /
robots.txt allows all
User-agent: *
Allow: /
robots.txt no index
Robots.txt doesn't support "noindex" anymore.
Use this HTTP header instead:
X-Robots-Tag: noindex
Crawl-delay
Asks bots to wait a set number of seconds between requests. Useful if your server has limited resources. Note: Google ignores this directive — use Google Search Console to manage Googlebot's crawl rate instead.
Crawl-delay: 10
Need different rules for different bots? The tool supports multiple User-agent declarations, so you can allow Googlebot full access while blocking SemrushBot or restricting Bingbot to specific sections. This is useful for:
AI companies train large language models by crawling the web. If you don't want your content used as training data, you can block AI crawlers in your robots.txt file.
The most common AI bots to block are:
User-agent: GPTBot
Disallow: /
User-agent: ClaudeBot
Disallow: /
User-agent: CCBot
Disallow: /
User-agent: Google-Extended
Disallow: /
Note: robots.txt is a voluntary standard. Most reputable AI companies respect it, but it isn't legally enforceable. For stronger protection, use Terms of Service language alongside robots.txt rules.
Google recommends:
Our tool applies these automatically.
Following Google's guidelines prevents accidental crawl blocks that can tank rankings:
Your robots.txt file should be reviewed whenever your site structure changes. Common triggers include:
Tip — validate after every change
After uploading your updated robots.txt, test it in Google Search Console under Settings → robots.txt. GSC will flag any pages that are blocked but currently indexed.
Solutions
Resources