Free tools for AI agent readiness
Small focused utilities to help your site become readable to autonomous AI agents and search crawlers. No accounts, copy-paste output, every tool runs in your browser.
robots.txt Generator
GeneratorsBuild a robots.txt with explicit rules for AI crawlers like GPTBot, ClaudeBot, PerplexityBot, and Google-Extended. Presets, sitemap line, copy-ready output.
robots.txt Tester
ValidatorsPaste a robots.txt, a URL path, and a user agent. See whether it is allowed using the Google matching algorithm: longest match wins, Allow beats Disallow on tie.
Structured Data Validator
ValidatorsPaste JSON-LD or a script tag and validate schema.org structure, required fields per type, and rich result eligibility for Article, FAQPage, BreadcrumbList, and more.
sitemap.xml Validator
ValidatorsPaste a sitemap.xml and validate the structure, URL count, oversize warnings, lastmod date format, changefreq enum, priority range, and per-URL field issues.
Feed Validator
ValidatorsPaste an RSS or Atom feed and validate the structure, required elements, dates, and uniqueness. Links out to the W3C Feed Validation Service for an authoritative check.
llms.txt Generator
GeneratorsCreate an llms.txt that gives autonomous agents a structured map of your most important pages and APIs. Sections, links, and short summaries in markdown.
llms-full.txt Generator
GeneratorsBundle markdown blocks into a single llms-full.txt for deep AI ingestion. Add titles, source URLs, see a token estimate, then copy or download the file.
JSON-LD Generator
GeneratorsGenerate JSON-LD structured data for Organization, WebSite, FAQPage, BreadcrumbList, and Article without writing JSON by hand. Validates against Google rich results.
Meta Tag Generator
GeneratorsGenerate Open Graph and Twitter Card meta tags with a live preview that shows how your URL renders when shared on social platforms. Copy a paste-ready block.
Content-Signal Builder
GeneratorsCompose a Content-Signal directive that declares your AI training, search, and AI-input preferences. Choose allow or disallow per signal and copy a robots.txt line.
Link Headers Builder
GeneratorsCompose RFC 8288 Link headers for agent-friendly discovery of api-catalog, sitemap, llms.txt, and feeds. Outputs a single header value plus a Next.js proxy snippet.
sitemap.xml Generator
GeneratorsPaste URLs to get a valid sitemap.xml with optional lastmod, changefreq, and priority. Validates each URL and produces a downloadable sitemaps.org-conformant file.
AI Bot Directives
ReferencesReference of AI crawler user agents with quick toggles to generate allow or disallow rules for each. Includes owner, purpose, and notes for 15 major bots.
User-Agent Lookup
ReferencesPaste a User-Agent string and identify whether it belongs to a known AI crawler, search bot, social previewer, or browser. Detects 45+ common agents instantly.
Agent Card Builder
GeneratorsGenerate a /.well-known/agent-card.json so A2A clients can discover your agent's identity, capabilities, and skills. Includes a Next.js route handler snippet.
API Catalog Builder
GeneratorsGenerate a Linkset for /.well-known/api-catalog so agents can discover every API on your site. Includes a drop-in Next.js route handler snippet (RFC 9727).
Markdown Negotiation
SnippetsCopy-paste Next.js proxy.ts and route handler snippets that return text/markdown when an agent sends Accept: text/markdown. Browsers still receive HTML.
Skip the tools, ship the baseline
Get one prompt that covers all six core checks. Paste it into your AI coding agent and it will apply the entire AgentScan baseline to your repo.
Why these tools exist
Most SEO tooling treats AI agents as an afterthought. These tools focus on the machine-readable surface that determines whether GPTBot, ClaudeBot, PerplexityBot, Googlebot, and similar crawlers can find, parse, and use your content correctly. Each tool is single-purpose, runs entirely in your browser, and produces output you can drop straight into your project.
Looking for something we do not cover yet? Open the contact form and tell us. We add tools based on what teams actually use during deploys, not based on traffic.