Free tool · No signup
robots.txt builder for AI bots
Pick a stance, toggle individual bots, get a ready-to-paste file. Covers GPTBot, ClaudeBot, PerplexityBot, Google-Extended, CCBot, Bytespider, and 10+ more — with honest notes on what each one does. Runs in your browser, built by GEO Tracker AI.
Use this tool if
- • You don't have a
robots.txtyet, or yours hasn't been reviewed in 12+ months. - • You want a deliberate stance on whether your content trains AI models vs gets cited in live AI search.
- • You're on WordPress / Webflow / Framer / Next.js / static and want a copy-paste file.
Skip this tool if
- • You already have a carefully-tuned
robots.txtreviewed recently — keep yours. - • You need IP-level blocking or rate-limiting, not just a polite request — handle that at the CDN / WAF layer.
- • See the dynamic-route pattern in the guide for environment-conditional rules.
AI bots — pick or toggle(0 blocked / 18)
First time? Read this 30-second guide
What is robots.txt? A plain-text file you host at the root of your site that tells crawlers which paths they may or may not fetch. It is the main lever for controlling whether your content trains AI models, appears in AI citations, or both.
- Pick a stance — "Allow everything" (most B2B SaaS), "Block training, allow citation" (publishers, IP-sensitive), or "Block all AI" (paid content). Load preset to apply.
- Toggle individual bots if you want more nuanced rules. Hover the info hint on each bot to see what it does.
- Add your sitemap URL and any site-specific paths you want blocked from all crawlers (e.g.
/admin,/api/private). - Copy or download — drop the file at
yourdomain.com/robots.txt.
Need more detail? The full setup guide covers per-bot context, framework setup, common mistakes, and how to verify the rules take effect.
OpenAI
Anthropic
Perplexity
Google & Apple
Other major AI crawlers
Extras
Optional fields applied to the wildcard User-agent: * rule.
Recommended. Google, Bing, and several AI crawlers use it for discovery.
Applied to all crawlers under the default User-agent: *rule. Don't use this for secret URLs — they will be publicly listed.
Generated robots.txt
Drop in your site root as /robots.txt
Output is a standard robots.txt with one rule block per blocked AI bot plus your wildcard. No backend call; generated in your browser. Inputs auto-save to this browser's local storage so an accidental refresh won't lose your work — use Reset to clear.
The honest part — what this controls and what it does not
robots.txt is a 30-year-old convention. It is the primary lever for telling polite crawlers what they may fetch, but it is not enforcement and it is not a ranking signal. A few clarifications worth getting right before deploying:
Does blocking GPTBot stop me from being cited in ChatGPT search?
GPTBotis OpenAI's training crawler. Citations in ChatGPT search use a separate bot called OAI-SearchBot. To stop being cited in ChatGPT search you would also need to disallow OAI-SearchBot and ChatGPT-User. The split is intentional — OpenAI lets you opt out of training without losing citation eligibility.Does Google-Extended block Google AI Mode or AI Overviews?
Google-Extended is an opt-out token for Bard / Gemini / Vertex AI training. AI Mode and AI Overviews are driven by Googlebot, which is unaffected by the Google-Extended rule. You cannot cleanly opt out of AI Mode / AI Overviews without also opting out of all of Google Search.How long does it take for new rules to take effect?
robots.txt on their own schedule — typically daily, sometimes weekly. New rules generally take 24–72 hours to take effect across major engines. Some less polite crawlers (e.g. Bytespider) ignore robots.txt entirely; for those, rate-limit at the CDN / WAF layer.Is robots.txt security?
Disallow are more visible to attackers, not less, because they are explicitly listed in the file. Use real authentication for anything sensitive.How do I verify the rules actually work?
curl -I yoursite.com/robots.txt to confirm the file is served at HTTP 200 with Content-Type: text/plain. (2) Google Search Console → Settings → Crawling → robots.txt report shows Google's cached version and lets you re-fetch. (3) Your own server logs, filtered by User-Agent on the blocked bot names — polite bots should stop fetching disallowed paths within 24–72 hours. None of that tells you whether AI engines cite you more or less afterwards — for that, run a controlled audit on a buyer-question panel.Disclaimer & limits
This is a free helper that produces a standard robots.txt file at the time of our most recent review (2026-05-13). A few honest notes so expectations are right from day one:
- robots.txt is a polite request, not enforcement. Well-behaved crawlers (OpenAI, Anthropic, Google, Perplexity, Common Crawl) honour it. Less polite ones (some training scrapers, low-quality data brokers, Bytespider on a bad day) ignore it. For real blocking use IP-level rules at your CDN / WAF in addition to the file.
- Bot landscape changes faster than tools. Vendors add, rename, and split bots several times a year — OpenAI introduced
OAI-SearchBotseparately fromGPTBotin 2024, Anthropic addedClaude-UserandClaude-SearchBotin 2024–2025, Apple addedApplebot-Extended. We track these and update the tool when they happen, but cannot promise the tool always reflects the absolute latest practice — verify vendor docs for anything mission-critical. - Output is not validated. We don't check the sitemap URL or extra paths you enter. Test the resulting file with Google's robots.txt Tester before deploying, and re-curl
yoursite.com/robots.txtafter deploy to confirm it is reachable. - Blocking AI bots does not improve AI search visibility. In most cases it reduces it — fewer bots means less content available for citations. Reasons to block are about IP, licensing, and brand control, not ranking. If your goal is more AI citations, blocking is usually the wrong lever; the right one is content quality, entity clarity, and a citation footprint in third-party sources.
- Provided as-is. The tool is provided free of charge without warranty, express or implied. We accept no liability for outcomes related to AI search visibility, organic ranking, training inclusion, or business results that follow from using or not using the generated file.
The reliable way to know whether your robots.txt changes moved AI citation behaviour is controlled measurement — before vs after, on a fixed buyer-question panel.
Next step
You shipped robots.txt. Now what does AI see?
Blocking is half the story — the other half is whether the bots you did let through are actually citing your brand on the buyer-questions that matter. Free 60-second audit, no card required: