RankPill Logo

Robots.txt Generator

Create a valid robots.txt file to control which pages Google and Bing can crawl and index on your site.

Crawl rules (User-agent: *)

Adds rules to block GPTBot, ClaudeBot, Google-Extended, PerplexityBot, Bytespider, CCBot from scraping your content for AI model training.

Generated robots.txt

User-agent: *
Allow: /

Common robots.txt Examples

Ready-to-use configurations for the most common website setups.

Allow everything (default)

Open to all crawlers. Standard for public websites.

User-agent: *
Allow: /

Block admin and private areas

Allow public crawling but protect backend paths.

User-agent: *
Disallow: /admin
Disallow: /wp-admin
Disallow: /private

Sitemap: https://example.com/sitemap.xml

Block AI training crawlers

Allow search engines but opt out of AI model training.

User-agent: *
Allow: /

User-agent: GPTBot
Disallow: /

User-agent: ClaudeBot
Disallow: /

User-agent: Google-Extended
Disallow: /

Block all crawlers

Completely private site — no bot crawling of any kind.

User-agent: *
Disallow: /

What is a robots.txt File?

A robots.txt file is a plain text file placed at the root of your website (yourdomain.com/robots.txt) that instructs search engine crawlers which pages they can and cannot access. It follows the Robots Exclusion Protocol, a standard recognized by Google, Bing, and all major search engines.

While robots.txt is not a security tool and cannot prevent determined bots from accessing your site, it is an important part of technical SEO. Properly configured, it prevents crawl budget from being wasted on admin panels, duplicate content, and internal search result pages, keeping Google focused on the content you actually want indexed.

Protect admin areas

Disallow paths like /admin, /wp-admin, and /dashboard to keep backend pages out of search results and reduce unnecessary crawling.

Add your sitemap

Including your sitemap URL in robots.txt helps crawlers discover your sitemap automatically without relying on Google Search Console submission.

Do not block CSS or JS

Google needs to render your pages to understand them. Blocking stylesheets or JavaScript files can prevent proper indexing and hurt your rankings.

Test before deploying

Use Google Search Console's robots.txt tester to verify your file before going live. A misconfigured file can accidentally block your entire site from Google.

Frequently Asked Questions

Common questions about robots.txt and search engine crawler directives.

Want these on autopilot?

RankPill automates everything these tools do. Meta descriptions, titles, content briefs, and full articles published to your site every day without lifting a finger.