Robots Ai

v1.0.0

Analyze and generate robots.txt files with AI crawler awareness. Detect which AI bots (GPTBot, ClaudeBot, PerplexityBot, Google-Extended, etc.) are blocked o...

0· 135·0 current·0 all-time

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for sharozdawa/robotstxt-ai.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "Robots Ai" (sharozdawa/robotstxt-ai) from ClawHub.
Skill page: https://clawhub.ai/sharozdawa/robotstxt-ai
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install robotstxt-ai

ClawHub CLI

Package manager switcher

npx clawhub@latest install robotstxt-ai
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description (robots.txt analysis/generation with AI-crawler awareness) match the SKILL.md instructions. The skill requires no binaries, credentials, or config paths — appropriate for an instruction-only robots.txt helper.
Instruction Scope
SKILL.md explicitly tells the agent to fetch /robots.txt from arbitrary websites, parse User-agent/Allow/Disallow rules, and produce suggested robots.txt content. That behavior is expected for this purpose, but it implies the agent will perform network fetches of target sites (and may reveal provided site URLs to any runtime network logs). The instructions do not request unrelated files, credentials, or external endpoints beyond the target site's robots.txt.
Install Mechanism
No install spec and no code files are present — lowest-risk instruction-only skill. Nothing is downloaded or written to disk by an installer.
Credentials
The skill declares no required environment variables or credentials. No disproportionate access to secrets or unrelated services is requested.
Persistence & Privilege
always is false and the skill does not request special persistent privileges or modifications to other skills/configs. Autonomous invocation is allowed by default but not excessive here.
Assessment
This skill appears coherent and safe in intent. Before installing, note that: (1) it will fetch robots.txt from arbitrary URLs you provide — ensure you are comfortable with the agent making outbound requests to those sites and with any logs that may record those URLs; (2) generated recommendations (e.g., 'block all AI' or blocking Googlebot) can have major consequences like removing your site from search/indexing — review generated robots.txt carefully before applying it to production; (3) verify the bot list if you rely on it for compliance, since crawler names and behaviors can change. If you plan to use it on private/internal sites, ensure the agent has appropriate network access and that no sensitive URLs are shared accidentally.

Like a lobster shell, security has layers — review code before you run it.

latestvk977prd20ntzhskwnfn8nwa42983a9e4
135downloads
0stars
1versions
Updated 1mo ago
v1.0.0
MIT-0

robots-ai

Analyze, audit, and generate robots.txt files with full awareness of 20+ AI crawlers.

Capabilities

  • Analyze any website's robots.txt to see which AI bots are blocked/allowed
  • Generate a robots.txt with toggleable AI bot blocking
  • Audit existing robots.txt for completeness and issues
  • List all known AI crawlers with their user-agents, companies, and documentation links

AI Bots Database

You know about these AI crawlers and their user-agents:

BotUser-AgentCompanyType
GPTBotGPTBotOpenAIAI Crawler
ChatGPT-UserChatGPT-UserOpenAIAI Search
OAI-SearchBotOAI-SearchBotOpenAIAI Search
ClaudeBotClaudeBotAnthropicAI Crawler
anthropic-aianthropic-aiAnthropicAI Crawler
Google-ExtendedGoogle-ExtendedGoogleAI Crawler
PerplexityBotPerplexityBotPerplexityAI Search
CCBotCCBotCommon CrawlAI Crawler
BytespiderBytespiderByteDanceAI Crawler
DiffbotDiffbotDiffbotAI Crawler
cohere-aicohere-aiCohereAI Crawler
AmazonbotAmazonbotAmazonAI Crawler
Meta-ExternalAgentMeta-ExternalAgentMetaAI Crawler
Meta-ExternalFetcherMeta-ExternalFetcherMetaAI Crawler
Applebot-ExtendedApplebot-ExtendedAppleAI Crawler
YouBotYouBotYou.comAI Search
TimpibotTimpibotTimpiAI Crawler
img2datasetimg2datasetOpen SourceAI Crawler

Important Notes

  • Google-Extended controls Gemini training access but does NOT affect Google Search indexing
  • Blocking Googlebot removes the site from Google Search entirely — never do this unless explicitly asked
  • CCBot feeds Common Crawl, which is used by many AI companies for training data
  • Bytespider (ByteDance) and Timpibot are commonly blocked by default due to aggressive crawling

How to Analyze

When asked to analyze a robots.txt:

  1. Fetch the robots.txt from the URL (append /robots.txt if not included)
  2. Parse all User-agent directives and their Allow/Disallow rules
  3. Check each AI bot against the rules
  4. Report: which bots are blocked, which are allowed, and any issues found
  5. Suggest improvements if relevant

How to Generate

When asked to generate a robots.txt:

  1. Ask which AI bots to block (or accept "block all AI" / "allow all AI")
  2. Ask for sitemap URL(s)
  3. Ask for any custom rules (e.g., Disallow: /admin/)
  4. Generate clean robots.txt with comments explaining each section
  5. Always include User-agent: * with Allow: / as the default
  6. Group blocked AI bots together with comments
  7. Add sitemap directives at the end

Output Format

Always format the generated robots.txt in a code block with syntax highlighting. Add comments explaining what each section does. Example:

# Allow all crawlers by default
User-agent: *
Allow: /

# Block AI training crawlers
User-agent: GPTBot
Disallow: /

User-agent: ClaudeBot
Disallow: /

# Sitemap
Sitemap: https://example.com/sitemap.xml

Comments

Loading comments...