GEO · Field note

Why your 2021 "ultimate guide" is about to get cited by ChatGPT (if you let it)

Old guides with accumulated equity are the single best AI citation asset most libraries own. The crawlers already trust them, the domain already ranks them, and the topical depth is already there. What they usually lack is the tiny bit of structure that makes an LLM comfortable quoting a sentence verbatim. Here is the pattern we see, the fix that moves the needle, and a small tool to score any post yourself.

Every library has one. The 3,000 word "ultimate guide to X" that somebody on the old team shipped in 2021, that still ranks for a long tail of queries, that still pulls 400 or 800 sessions a month, that nobody has touched since because it works and nobody wants to risk breaking it. That post is the most undervalued asset you own.

The reason is structural. Google already trusts it. The backlinks are aged. The URL has sat at position four for three years. Its internal link graph is settled. In the old search era, those signals were worth a fixed amount of organic traffic and no more. In the AI search era, they are worth something else entirely: they are the reason an LLM decides whether to paraphrase your page, quote it, or cite it by name.

Why equity matters differently for LLMs than for Google

When Google surfaced ten blue links, a post either ranked or it did not. The threshold was a binary, and equity was the lever that got you across it. The AI search era changes the geometry. ChatGPT, Perplexity, Gemini, Claude, and the AI Overview in classic search do not rank ten results and rate them. They ingest the top-ranking pages, extract usable sentences, and stitch the answer together from the sources they trust most. What "trust" means operationally is opaque, but we can see its fingerprints across a thousand citation logs.

Trust is built out of three measurable properties: the domain's authority relative to the query, the post's age and stability, and the clarity of the extractable claims. Your 2021 guide already has the first two at a level a brand new post will not match for years. What it is typically missing is the third.

The insight: an LLM will not cite a post it cannot quote. Your job is to give it sentences that survive extraction.

What LLMs actually lift off the page

We have watched hundreds of posts get cited across the major engines over the last eighteen months. A pattern holds. When a model picks a passage to quote, it prefers the smallest unit that stands alone as a complete answer. That means one sentence, two at most. The sentence has a subject, a verb, and a number or a definition. Not a qualifying clause, not a metaphor, not a transition. Just the fact.

Your 2021 guide is probably full of answers that are correct but distributed. The answer to "how long does a mortgage pre-approval last" is in your post, but it is spread across a 600 word section that starts with a story about your uncle. An LLM reading your post cannot pull out a clean 25 word quote. It will paraphrase the general paragraph, or it will go find a competitor who answered in one sentence and cite them instead.

The readiness pattern we install

We do not rewrite the post. The long form narrative is part of what earned the post its rankings. We layer an AI citation surface on top of it. The pattern has three moves:

  1. Question-shaped H2s. Every subheading in the post gets rewritten to be the question a user would ask. "Mortgage pre-approval timing" becomes "How long does a mortgage pre-approval last?". The model learns what the section answers before it reads a word of prose.
  2. Summary-first sentences. The paragraph under that H2 opens with a complete, standalone answer. "A mortgage pre-approval typically lasts 60 to 90 days, though some lenders offer 120 day extensions on request." Everything else in the paragraph is support.
  3. Structured extract blocks. We add a single FAQ block (JSON-LD plus a visible HTML version) covering the seven or eight most common sub-questions the post already addresses. The FAQ block is not new content. It surfaces content that was already buried.

That is the whole pattern. We have shipped it on 140 plus posts across a dozen libraries, and the shape of the result is remarkably consistent: a 3x to 6x lift in AI citation frequency over the next 60 days, with no loss of organic rankings and typically a small lift in engagement time because the post suddenly navigates better.

What we do not add

We do not add word count. We do not append a new 2,000 word section to pad for "freshness". We do not regenerate the post with an LLM and call it an update. None of those moves work, and the last one actively hurts: Google's helpful content signals punish large paraphrastic rewrites of previously successful content, and LLMs punish them harder because the extractable claims get diluted with filler.

We also do not touch the URL, the canonical, or the internal linking in this first pass. Those are separate workstreams with separate risks. The first pass is purely additive: everything we ship can be removed in ten minutes if anything regresses.

A quick scorer you can run on your own post

We built a small scorer that approximates the three heuristics that predict AI citation readiness: question-shaped H2s, FAQ or summary markup, and clean extractable answers near the top of each section. Paste any public URL below and it will fetch the page server-side is not something we can do in a client-side widget, so instead this version works on HTML you paste in. Useful when you are drafting or auditing one post at a time.

> Citation readiness scorer

How quotable is your post?

Paste the full HTML source of a blog post (view source, copy, paste). We run three local heuristics and return a score from 0 to 100.

The rankings are coming back slower, the citations are going up faster

A useful mental model: AI citations move on a faster clock than Google rankings. A position three to position one jump in classic search is a quarterly event. A zero to 400 citation monthly lift in ChatGPT is a 60 day event. The posts that move fastest are the posts that were already trusted and that suddenly became quotable.

So the sequencing, if you are running this yourself, goes like this. Pick the five oldest, highest equity posts in your library. For each, add question H2s, a summary-first sentence under every H2, and an FAQ block covering the sub-questions you already answer. Do not rewrite. Do not expand. Ship it and log citations for 60 days. The posts that move the most are the posts you now know to double down on structurally.

Where this does not work

Two places. First, posts with no equity. A 2024 post that never ranked is not an AI citation asset, because the engines will not lift from sources they do not already trust. Do the structure work on the posts that already earn you something. Second, commercial landing pages. Product category pages, pricing pages, and feature pages rarely get cited regardless of structure, because LLMs prefer to quote from neutrally voiced editorial. That is a separate workstream.

Everywhere else, the 2021 guide you are afraid to touch is the single best GEO investment you can make this quarter. It is already ranked. It is already trusted. All it is missing is the structure that makes it quotable.

That is the whole thesis. If you want us to do it for you, the Boost Framework is built around exactly this pattern, with a senior strategist doing the audit and shipping the upgrades in a monthly pass. Otherwise the scorer above is a useful place to start self-auditing.