AI Visibility Checklist
25 signals that determine whether ChatGPT, Claude, Perplexity, and Gemini can find, understand, and recommend your product. Check each one manually — or run a free AI visibility audit to check all 25 at once.
AI Crawl Access
Before an AI assistant can recommend your product, it needs permission to crawl your site and a structured way to understand it. These 5 signals ensure AI crawlers can index your content.
sitemap.xml is present at /sitemap.xml
Search engines and AI crawlers use sitemaps to discover all your pages. Without one, AI systems may miss important product pages.
Most frameworks (Next.js, Nuxt, Astro) can generate sitemaps automatically with a plugin or built-in route.
sitemap.xml contains 3 or more URLs
A sitemap with only 1 URL suggests your content architecture is shallow, which reduces crawl depth.
Include your homepage, features, pricing, about, and blog pages at minimum.
robots.txt allows GPTBot (ChatGPT's crawler)
GPTBot is the official OpenAI/ChatGPT crawler. Blocking it prevents your site from being indexed in ChatGPT's knowledge base.
Add 'User-agent: GPTBot' with 'Allow: /' to your robots.txt. Do not block it unless you have a specific reason.
robots.txt allows ClaudeBot (Anthropic's crawler)
ClaudeBot is Anthropic's official web crawler. Blocking it prevents Claude from knowing about your product.
Add 'User-agent: ClaudeBot' with 'Allow: /' to your robots.txt.
robots.txt allows PerplexityBot
PerplexityBot powers Perplexity AI's web index. If it's blocked, Perplexity cannot cite your product in answers.
Add 'User-agent: PerplexityBot' with 'Allow: /' to your robots.txt.
Content Quality
AI systems analyze your site's content to understand what you do. JavaScript-heavy sites with minimal crawlable text score poorly — LLMs read HTML, not rendered JavaScript.
Homepage has 500+ words of visible, crawlable text
LLMs need sufficient text to extract product information, positioning, and value propositions. Sub-500 word pages give AI little to work with.
Ensure your key marketing content is server-rendered and appears in the HTML source, not just after JavaScript executes.
Text-to-HTML ratio is above 15%
A low text-to-HTML ratio indicates a JavaScript-heavy page with minimal real content — a pattern common in SPA frameworks where text is injected dynamically.
Use server-side rendering (SSR) or static generation for your marketing pages. Check your homepage's raw HTML source to verify.
3+ pages are reachable from internal links on the homepage
Internal links help AI crawlers discover more of your site. A single-page app with no deep links limits what LLMs know about your product.
Add navigation links to your features, pricing, about, and blog pages from your homepage.
Navigation links are present and crawlable
Well-structured navigation tells AI systems what sections of your site exist and how your content is organized.
Use standard HTML anchor tags for navigation. JavaScript-only routing may be invisible to crawlers.
Product Clarity
AI systems recommend products when they can confidently describe what the product does. Vague or missing H1s, no features documentation, and no pricing reduce citation confidence.
A clear H1 headline is present on the homepage
The H1 is the most authoritative statement about what your product does. AI systems weight it heavily when generating product descriptions.
Your H1 should describe what the product does and who it's for in one clear sentence. Avoid 'Welcome to [Company]' — be explicit.
Features page or feature keywords are present
When someone asks an AI 'what does X do?', it needs feature information to give a complete answer. Missing features = incomplete AI responses.
Add a /features page or include a features section on your homepage with clear feature titles and descriptions.
Pricing page or pricing information is present
Pricing is frequently asked about in AI queries ('how much does X cost?'). A missing pricing page means AI cannot answer this question accurately.
Add a /pricing page and link to it from your navigation. Include your free tier if you have one.
Structured Data & Meta
Structured data is the highest-fidelity machine-readable format for product information. JSON-LD schema and proper meta tags allow AI systems to extract accurate, structured facts about your product.
OpenGraph og:title tag is set
og:title is used by AI systems when generating summaries and recommendations. It should match your product name and core value proposition.
Set og:title to '[Product Name] — [One-sentence value proposition]'. Most meta tag libraries handle this.
OpenGraph og:description tag is set
og:description is frequently extracted by AI systems when generating product summaries. It should clearly state what the product does.
Write a 1-2 sentence og:description that explains the product's core function and primary benefit.
OpenGraph og:image tag is set
Visual context helps AI systems build richer product profiles. A missing og:image reduces visual citations in AI-generated responses.
Create a 1200x630px OG image with your product name and key value proposition.
Page title tag is present and descriptive
The <title> tag is one of the most heavily-weighted page signals. Vague titles like 'Home' give AI systems minimal product information.
Use the format '[Product Name] — [Value Proposition] | [Brand]' for your homepage title.
Meta description is present (150-160 characters)
Meta descriptions are frequently used as summaries by AI systems. They appear in citations and AI-generated comparisons.
Write a descriptive meta description that includes your product name, primary function, and one key benefit.
JSON-LD SoftwareApplication, Product, or Organization schema is present
This is the single highest-impact structured data signal. JSON-LD gives AI systems precise, machine-readable facts: product name, category, description, pricing URL, and more.
Add a <script type='application/ld+json'> block with SoftwareApplication schema. AIExposureTool generates this for you.
Canonical URL is set
Canonical URLs tell AI crawlers which version of a URL is authoritative, preventing duplicate content penalties.
Set rel='canonical' in your page head to your preferred URL. Most Next.js and meta tag setups handle this automatically.
Agent Readiness
llms.txt files are a direct, unambiguous brief for AI agents. When a Claude or ChatGPT agent browses your site, llms.txt is the first file it should read — designed specifically for LLMs.
llms.txt file exists at /llms.txt
llms.txt is the AI equivalent of robots.txt — a plain-text brief that tells LLMs exactly what your product does, who it's for, and where to find key information. AI agents will look for it first.
Create a plain text file at /llms.txt with: product name, tagline, description, target audience, key use cases, pricing summary, and links to important pages. AIExposureTool generates this automatically.
llms-full.txt file exists at /llms-full.txt
llms-full.txt is an extended version of llms.txt with complete product documentation. It gives AI systems enough context to answer detailed product questions accurately.
Create /llms-full.txt with full feature descriptions, methodology, all pricing tiers, and comprehensive documentation. AIExposureTool generates this for you.
Check all 25 signals automatically
Instead of checking each signal manually, paste your URL into AIExposureTool and get your AI Exposure Score (0-100) in 30 seconds — with exactly which signals you're passing and failing.