← All posts

April 17, 2026

Will ChatGPT cite your site? How to test and fix it

Whether ChatGPT will cite your site depends on three things: whether its crawlers can access your content, whether your content contains extractable and quotable passages, and whether your site has enough authority signals for the model to trust the citation.

Most B2B sites fail all three. Not because the content is bad — because the content isn't structured for extraction. There's a difference between writing that's good to read and writing that's good to quote.

Why do some sites get cited and others don't?

ChatGPT cites content that is:

  • Self-contained at the passage level. A cited paragraph makes sense without reading the 10 paragraphs around it.
  • Specific and verifiable. Numbers, named tools, dates, defined terms. Generic claims ("leading solution," "best-in-class") contribute nothing citable.
  • Structurally signaled. H2 headers framed as questions, definition-first section openings, numbered steps.
  • Schema-attributed. Article, Organization, and FAQPage JSON-LD tell models who wrote the content and what type of content it is.
  • Crawlable. OpenAI's GPTBot, Anthropic's ClaudeBot, and Perplexity's PerplexityBot all respect robots.txt.

How to test whether ChatGPT cites you right now

Method 1: Direct query test. In ChatGPT with web browsing enabled, ask a question your site is supposed to answer. Try 5 different phrasings. Note whether your domain appears in any citations.

Method 2: Perplexity test. Perplexity does live web retrieval, so results are more current than ChatGPT's training-based responses. Ask the same 5 questions. Check the source panel.

Method 3: CiteReady free audit. Paste your URL at citeready.io/free-audit. The free preview checks your AI crawler accessibility, llms.txt presence, and content freshness in about 5 seconds. The full audit adds schema completeness, passage-level citability scored by Claude, AI Overviews positioning, and structured pattern detection — plus a Claude Opus executive summary and prioritized fix plan.

The manual tests tell you the outcome. The audit tells you why.

The 3 most common reasons B2B sites fail

Reason 1: Homepage marketing prose

The most cited page on most websites is the homepage. The homepage of most B2B SaaS companies looks like this:

"We help modern teams move faster, collaborate better, and unlock their full potential through intelligent automation."

This sentence contains zero citable information. There is no definition, no number, no named capability, no specific claim.

Compare that to an opening that reads:

"Acme is a supply chain automation platform that reduces procurement cycle time by an average of 34%, based on data from 120 enterprise customers."

That second version has a definition, a specific claim, a scope, and a sample size. A model can quote that.

The fix: Rewrite the first paragraph of your homepage to lead with a definition of what you do and one specific, verifiable outcome. Avoid superlatives. Include numbers.

Reason 2: Missing schema

FAQPage schema is the single highest-leverage schema type for AI citation. FAQ content is already structured as a question/answer pair — exactly the format models use to answer queries.

Here is a minimal FAQPage JSON-LD block:

{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [{
    "@type": "Question",
    "name": "What is an AI search visibility audit?",
    "acceptedAnswer": {
      "@type": "Answer",
      "text": "An AI search visibility audit evaluates whether a site's content, structure, and technical setup allow AI tools like ChatGPT and Perplexity to find, parse, and cite its pages."
    }
  }]
}

The fix: Identify your 5 highest-traffic pages. Add FAQPage schema to any that contain Q&A content. Add Article schema with dateModified to all blog posts.

Reason 3: No comparison tables or structured patterns

B2B buyers use AI assistants to compare options. "Best CRM for small teams." "Datadog vs. New Relic for APM." These are high-intent queries, and the sites that get cited in responses almost always have structured comparison content.

Unstructured prose comparisons ("Some tools handle this better than others, but it depends on your use case...") are not extractable. A markdown table with named alternatives and specific attributes is.

The fix: Add one comparison table to your category page or primary product page. Name at least 3 alternatives (including yourself) and compare on 4-6 specific, objective attributes.

What score should you aim for?

In CiteReady's 7-dimension scoring system, each dimension is scored 0-100:

  • Sites scoring below 40 overall are invisible to AI citation in all but the most specific branded queries
  • Sites scoring 40-65 appear occasionally but inconsistently
  • Sites scoring 65+ appear in AI citations for their core topics regularly

Most fresh B2B sites land in the 30-45 range. The three fixes above — homepage rewrite, schema, comparison tables — routinely move a site 15-25 points.

Action checklist

  1. Run the direct query test: ask 5 questions in ChatGPT (with browsing) and Perplexity that your site should answer, and record whether your domain appears in citations.
  2. Add FAQPage JSON-LD to your top 3 pages with Q&A content, and rewrite the homepage opening paragraph to lead with a definition and one specific number.
  3. Run a free CiteReady audit to get a scored baseline and a prioritized fix plan across all 7 dimensions.

FAQ

Does ChatGPT cite sites based on SEO ranking or something else? Both, depending on the mode. When ChatGPT uses its training data (no browsing), it reflects the content it saw during training. When using web browsing, live retrieval favors crawlable, structured, schema-rich content.

My site has great backlinks and high DA. Why doesn't ChatGPT cite me? Domain authority is necessary but not sufficient. High DA tells AI tools that your site is trustworthy — it doesn't tell them which passage to quote or why it's relevant to a specific query. A site with DA 30 and excellent passage-level structure will out-cite a DA 70 site full of generic marketing prose.

How is CiteReady's passage-level citability score calculated? CiteReady crawls your homepage and top pages via Firecrawl, then passes the content to Claude, which evaluates each major section against a rubric: definition presence, specificity, structural signaling, and self-containment. Claude returns a score per section and a set of rewrite suggestions.

Run a free CiteReady audit on your site

Get your score across all 7 dimensions — with a Claude-Opus executive summary and prioritized fix plan. First full audit is free.