Here’s a problem most marketing teams aren’t diagnosing correctly: you have solid content — thorough, well-structured, genuinely useful — and it still isn’t showing up in AI search answers.

The usual response is to go back and add more statistics. Tighten the schema. Shorten the intro. These fixes work sometimes. Most of the time, they miss the actual problem.

A study published this month by researchers at Virginia Tech and Zhejiang University (arXiv:2603.09296) ran the first systematic analysis of why topically relevant webpages fail to be cited by generative AI engines. They analyzed 949 contrastive pairs — cases where two pages were both retrieved for the same query, but only one got cited — and built a taxonomy of what separated the cited page from the invisible one.

The finding that should change how your team audits this: 62.2% of citation failures are semantic alignment failures, not content quality failures. Your page isn’t being skipped because it’s thin. It’s being skipped because the AI engine judged it as a poor match for the query’s intent — even when the topic overlap was obvious.

And it gets more specific than that.


The four failure modes, ranked by frequency

The research team analyzed 949 contrastive pairs from GEO-Bench — the Princeton/Georgia Tech citation research dataset — mapping each non-citation event to a specific pipeline stage. Here’s the breakdown:

Failure mode Share of failures Pipeline stage
Semantic Alignment 62.2% Response generation
Content Quality 27.1% Response generation
Technical Integrity 10.1% Fetching / parsing
Systemic Exclusion 0.6% Domain authority

Semantic Alignment — 62.2%. The dominant problem. Intent divergence is the most common sub-failure: your page is informational, but the query was transactional. Or your page walks through an evaluation process when the searcher wanted setup steps. Contextual gaps are next — the page is relevant to the general topic but missing the specific entity, terminology, or sub-question the query is actually asking about. Then outdated signals: accurate content with 2024 benchmark stats loses to a page with a 2026 number for time-sensitive queries. Last is localization mismatch — UK regulation details on a US compliance page, or enterprise-scale guidance when the query is from an SMB context.

Content Quality — 27.1%. This is where most teams focus all their energy, and it’s the minority problem. The failures here are information scarcity (too shallow to extract a citable claim), content fragmented across sections in ways the engine can’t synthesize, excessive verbosity, and unstructured prose where a table would let the engine extract specific thresholds.

Technical Integrity — 10.1%. JavaScript-heavy rendering the crawler can’t process, excessive boilerplate drowning the signal, unparseable content structure. This category genuinely requires a technical fix — it’s just not where most of your volume is.

Systemic Exclusion — 0.6%. A high-authority source (Wikipedia, a major industry publication) covers the same facts with stronger domain signals. Content optimization alone won’t beat this — you need third-party authority in the citation chain itself.

This taxonomy matters because the right fix depends entirely on which failure mode you’re dealing with. Adding statistics to a page with an intent alignment failure does nothing. Restructuring a page with a domain authority problem does nothing. Most teams are applying the content quality fix to a semantic alignment problem — which is why their citation rates don’t move.


What the evidence says about the scale of this

The AgentGEO study isn’t the only recent data pointing at the severity of the citation gap.

Research from Moz’s 2026 analysis of 40,000 Google queries found that 88% of AI Mode citations don’t overlap with the organic top 10 — meaning ranking well in traditional search gives you almost no predictive signal for whether you’ll be cited in AI answers. The two populations are largely independent.

A September 2025 study from Berkeley and Leanid Palkhouski — the GEO-16 framework, which audited 1,702 citations across Brave, Google AIO, and Perplexity — found that pages meeting the GEO-16 threshold (overall score ≥0.70 and ≥12 pillar hits) achieved a 78% cross-engine citation rate, compared to far lower rates for pages below it. The pillars most strongly associated with citation: Metadata & Freshness (r=0.68), Semantic HTML (r=0.65), and Structured Data (r=0.63). But critically — the same study confirmed what AgentGEO proved independently: “generative engines heavily weight earned media and often exclude brand-owned and social content.” Even technically perfect pages don’t get cited if they sit on a vendor blog with no third-party authority.

Forrester’s 2026 B2B buying research found that 94% of B2B buyers now use AI during their buying process — and those who use generative AI search tools cite them as the most meaningful or important information source in the purchase journey, ahead of vendor websites and product experts. That’s the audience you’re trying to reach, and they’re seeing whatever the AI engine decides to cite.

Muck Rack’s Generative Pulse data, drawn from over a million AI prompts, found that 85%+ of AI citations come from earned media sources. The practical implication: for the 62% of citation failures caused by semantic alignment, fixing the content helps. For the 0.6% caused by systemic exclusion, only earned media in trusted publications closes the gap. Most B2B brands are somewhere in between — and the correct sequence matters.


What to actually audit

Most AI citation audits default to checking technical SEO signals — schema markup, metadata freshness, page speed. Those matter, but they’re addressing the 10% problem while ignoring the 62% one.

Run this audit against your top 10 pages that should be driving AI citations but aren’t:

Step 1: Map query intent to page intent. Pull the specific AI queries where you’d expect to appear. Are they transactional (“how to implement X”) or informational (“what is X”)? Compare that against what your page delivers. A page written to educate will not get cited for a query that needs a setup guide — regardless of how thorough the education is.

Step 2: Check entity coverage. Run the query yourself in Perplexity or Google AI Overviews. Note what specific companies, frameworks, tools, or metrics the cited sources mention that your page doesn’t. Contextual gaps usually show up as missing proper nouns. If every cited page mentions “HubSpot’s attribution model” in the context of marketing attribution, and yours doesn’t, that’s the gap.

Step 3: Verify data freshness. For any stat in your top pages, check the date. The GEO-16 research found Metadata & Freshness has the strongest correlation with citation likelihood across engines (r=0.68). A 2023 benchmark stat loses to a March 2026 one for time-sensitive queries. Virginia Tech’s 2026 guidance is specific: surface human-visible timestamps, populate machine-readable dates, and note substantive revisions.

Step 4: Measure information density against the cited competitors. Not word count — density. The AgentGEO research found information scarcity is the core content quality failure: a page that can’t be extracted into a citable claim doesn’t get cited. A 2,000-word page with three data points is worse than a 600-word page with six. Tables, specific thresholds, named tools with specific use cases — these produce extractable density.


The fix priority order

The AgentGEO system achieved a 40% relative improvement in citation rates by modifying only 5% of page content — because it diagnosed the specific failure mode first, then applied the targeted repair. Generic optimization that rewrote 25% of a page produced less improvement because it changed the wrong things.

The researchers also found that generic optimization can actively harm long-tail content — applying broad rewrites to specialized pages can make them less citable, not more, because it dilutes the domain-specific precision that made them relevant in the first place.

For your workflow: fix semantic alignment issues before touching anything else. If the page’s intent doesn’t match the query intent, rewriting the intro for clarity accomplishes nothing. You need either a new page targeting the correct intent, or a genuine restructure of the existing page’s angle.

Once intent is aligned, close the entity gaps. Add the specific terminology, frameworks, or proper nouns that the cited competitors include. These are usually surgical edits — a paragraph, a section subhead, a table row.

Then check data freshness. Update outdated stats before you touch structure or schema. Then technical fixes — and only if you’ve confirmed the page isn’t being properly rendered or parsed.


The failure mode that optimization can’t fix

One honest finding: some citation failures aren’t recoverable through content improvement alone.

When a Wikipedia page, a Forbes deep-dive, or a university research paper covers the same facts with higher domain authority, no amount of structural optimization on your brand-owned page beats that citation. The AI engine treats the higher-authority source as the primary reference, and your vendor blog as a secondary option.

This is the structural limit of content-side optimization. The only path around it is getting your claims, data, and analysis cited in those higher-authority sources — which is how earned media becomes the prerequisite for AI citation, not a nice-to-have.

The mechanism is direct: your original data point appears in a Forbes article or industry research publication, that source gets cited by the AI engine, and your brand gets attributed through the citation chain. Machine Relations describes this at the infrastructure level — earned media in trusted publications isn’t a branding exercise, it’s the citation substrate. The AI engines already trust those publications. Getting your analysis into them is how you move from content-side optimization (which has real limits) to position in the citation sources the engines are already drawing from.

Most B2B operators are trying to make their vendor blog competitive with sources that have decades of trust signals. The faster path is contributing original analysis to the sources that already have that trust — and structuring your own pages to capture the queries those sources don’t address directly. Authoritytech.io’s research on earned vs. owned citation rates shows the gap: earned media distribution produces 325% more AI citations than owned content distribution for the same underlying content.


Takeaways

If you want to understand where your brand currently sits in AI search citation chains before auditing individual pages, the AT visibility audit gives you a baseline.