The AI Visibility Stack is the conceptual model AIVZ uses to organize all 93 AEO factors into three dependency layers — Access, Understanding, Extractability. It's the most important methodology choice in the product. It tells you what to fix first, and why fixing it second is wasted work.
The layers stack vertically, and the dependency runs upward: each layer depends on the one beneath it. A failure at any layer collapses the layers above it.
When this layer fails, AI reads you but quotes someone else. The most common citation-failure mode for content-heavy sites.
When this layer fails, AI reaches your content but can't reliably interpret it. Citations don't fire because the system isn't confident about the source identity.
Without this layer working, no other layer matters. AI can't optimize for content it can't reach.
If your robots.txt blocks GPTBot, your perfectly-structured FAQ schema does literally nothing for ChatGPT visibility. The schema is correct. The content is excellent. ChatGPT will never see it.
The Stack tells you what to fix first. Bottom-up. Always.
Most AEO tools treat factors as independent contributors. The Stack treats them as dependent. The difference changes how you sequence work.
"Your page scored 47. To improve, fix any of these 23 issues — they each contribute to the total."
User picks the easiest fix. Adds FAQ schema. Score moves 47 → 51. Marginal gain.
"Your page scored 47. The score is capped at 65 because Layer 1 is failing — your robots.txt is blocking GPTBot. Fix that first. Projected score after Layer 1 fix: 78."
User fixes Access. Score moves 47 → 78. Step-function gain — because the upstream work was already done; it just wasn't counting.
The dependency model captures something real about how AI systems behave: they fail gracefully at the boundaries you expect (no access → no citation), and they compound success at the boundaries where multiple signals reinforce each other (clean schema + structured answer + named author = high citation rate). A flat scoring model can't represent this. The Stack can.
The question this layer answers: Can AI crawlers physically reach your content?
Layer 1 measures whether AI crawl agents — GPTBot, ClaudeBot, PerplexityBot, GoogleOther, Bingbot — can access, discover, and render your pages. Without this layer working, no other layer matters.
| Sub-score | What it measures |
|---|---|
| Crawl Access | Whether AI bots can fetch your pages — server response codes, redirect chains, JavaScript rendering requirements, page render time. |
| Feed & Discovery | Whether AI bots can find your content — XML sitemaps, RSS/Atom feeds, internal linking, llms.txt declarations. |
| Bot Blocking | Whether bot-protection infrastructure (robots.txt, WAF rules, CDN configurations) is permitting or blocking AI crawlers. |
robots.txt blocks AI crawlers (often unintentionally)llms.txt publishedrobots.txt for AI bot disallows; explicitly allow if discovery is desiredllms.txt at site root with crawl preferencesConfidence: Most Layer 1 factors are Established — well-supported by web standards. A few (e.g., specific AI-bot user-agent identification, llms.txt adoption) are Emerging / Experimental because the standards are still settling.
The question this layer answers: Can AI parse the structure and meaning of your content?
Layer 2 measures whether AI systems can identify what your page is about, who wrote it, what entities it discusses, and whether the metadata accurately represents the content. This is where structured data does its work.
| Sub-score | What it measures |
|---|---|
| Schema | Presence and correctness of JSON-LD structured data — Organization, Person, Article, FAQPage, and other Schema.org types. |
| Structure | Document structure — heading hierarchy, semantic HTML, content organization. |
| Entity | Entity grounding — sameAs linking, Wikidata presence, knowledge graph alignment, cross-page consistency. |
| Schema Accuracy | Whether structured data matches the actual content — no hallucinated fields, no stale metadata, no schema-content drift. |
| Author E-E-A-T | Author identification — named authors, credentials, bio pages, expertise signals, publication history. |
sameAs linking to authoritative profilessameAs to Wikidata, LinkedIn, CrunchbaseConfidence: JSON-LD and core Schema.org types are Established. Entity grounding via sameAs is Strongly Inferred. E-E-A-T author signals are Strongly Inferred for most platforms, Established for Google-derived AI surfaces.
The question this layer answers: Can AI cleanly isolate and reuse the answer?
Layer 3 measures whether AI systems can extract a clean, citable answer block — the kind of structured, concise, fact-dense passage that AI includes in generated responses. When Layer 3 fails, AI reads you but quotes someone else.
| Sub-score | What it measures |
|---|---|
| FAQ | Presence and quality of FAQ structures — proper <dl>, FAQPage schema, scannable Q&A blocks. |
| Summary | Definitional and summary density — explicit definitions, summary paragraphs, key-takeaway blocks at the top of long content. |
| Content Richness | Density of citation-friendly elements — lists, tables, statistics with sources, structured comparisons, numbered procedures. |
| Content Quality | Readability, factual accuracy, originality, depth on topic, citation formatting. |
| Speakable | Voice-extractability — speakable schema, voice-friendly answer length, conversational phrasing. |
<table> markupConfidence: Front-loaded answers, concise blocks, and HTML structure factors are Strongly Inferred. FAQPage schema is Established. Speakable schema is Emerging / Experimental — the standard is settling and platform support is uneven.
Each layer produces its own score (0–100). The three aggregate into a composite AI Visibility Score. The aggregation is not a simple average — lower-layer weakness caps the contribution of upper layers.
| Score | Tier | What it means |
|---|---|---|
90–100 |
Healthy across all three layers. AI systems can find you, understand you, and cite you reliably. Maintenance mode. | |
70–89 |
AI Extractable | Strong foundation. Layer 3 needs refinement — you're being read but not always cited. Most fixes are content-structural. |
40–69 |
AI Readable | AI can access and partly understand you, but answers aren't extractable cleanly. Most sites land here on first scan. |
0–39 |
Invisible to AI | Failing in early layers. AI can't reliably reach or interpret your content. High-leverage fix opportunity — large gains available with focused work. |
The score is a leading indicator. Citation outcomes — whether AI systems actually cite you — are the lagging indicator. Improvement in the score should be followed by improvement in observed citation rate, on a lag of weeks-to-months as AI systems re-crawl and re-evaluate.
Different AI systems weight the three layers differently. A single composite score under-represents what's actually happening — the platforms surface different visibility patterns from the same underlying content.
| Platform | Tends to weight |
|---|---|
| ChatGPT | Training-data exposure (older, well-cited content); structured answer extractability; entity grounding |
| Google AI Overviews | E-E-A-T signals; existing search ranking; Schema.org compliance; Google-aligned freshness |
| Perplexity | Live crawl access (Layer 1 disproportionately); citation-friendly formatting; recency |
| Microsoft Copilot | Bing-derived ranking; Office-integration content surfaces; entity disambiguation |
| Gemini | Google-derived signals + Gemini's specific extraction model; multimodal content readiness |
| Voice assistants | Speakable schema; concise answer blocks; conversational phrasing |
A site can be in the AI Extractable tier overall but Invisible to AI on Voice — typical for content-heavy sites that haven't added speakable schema. The composite hides the gap; the per-platform breakdown surfaces it.
Per-platform factors and readiness scoringThe AI Visibility Score by itself is a number. The interesting part is what produced it. Here's how a real (anonymized) domain scored on first scan.
Verdict: Layer 1 was not the bottleneck. Most of the score was inherited from this layer being healthy. One small WAF issue with GoogleOther — fixed in 30 minutes.
Verdict: The primary bottleneck. Article schema present but inconsistent. Author identification on ~60% of posts. No sameAs. No Wikidata entity. The score capped because Layer 2 weakness propagated upward.
Verdict: Structurally underdeveloped. Answers buried in paragraph 4–6. Few question-based headings. No FAQ structures. No tables. No speakable schema. Capped by Layer 2 — until that improved, fixing this wouldn't show.
sameAs to LinkedIn. Added Wikidata entity for the company (manual edit; took two weeks to verify). Fixed three pages with stale metadata.Step-function gain. The Layer 3 work hadn't been counting until Layer 2 gave it something to anchor to. A flat scoring system would have surfaced 23 issues at 47 and let the team pick whichever was easiest. The Stack told them: Layer 2 first, Layer 3 second, ignore Layer 1.
The Stack is a methodology framework, but it also serves as a communication framework. It changes how AEO work gets sold internally — to executives, to clients, to budget owners.
"Your composite score is 42 and the limiter is Layer 1. Specifically, your robots.txt is blocking GPTBot. This is a 30-minute fix. After we ship it, your score will likely jump 15–25 points without changing anything else, because all the existing Layer 2 and Layer 3 work suddenly counts."
"Last quarter: 56 average across the client portfolio. This quarter: 73. Most of the gain came from a focused Layer 2 schema standardization push across 14 client sites. The Layer 3 work is in flight; we project another 10–12 points next quarter."
"We measure visibility across three dependency layers — Access, Understanding, Extractability. Each depends on the one before. We can show you exactly where in this stack your visibility is breaking, what the highest-leverage fix is, and what the projected score gain looks like. Want to run the scan?"
Get the composite score, the three-layer breakdown, the failing factors per layer, and the prioritized fix path — in 60 seconds.
Run a free scanThe 93 factors that compose the three layers — organized into 9 categories with confidence labels.
Read the 93-factor taxonomyThe eleven factors with disproportionate impact on citation outcomes — drawn from across all three layers.
Read the Citation Core 11Or — return to the methodology hub: Read the AEO methodology →
Run a free scan. Get the three-layer breakdown plus per-platform readiness across the 6 AI surfaces — in under 60 seconds.