Our team has the bandwidth to create 2 more SEO growth plans this month. Get one now โย
Citation Engineering: How to Make LLMs Recommend Your Brand on Purpose

Gumlet found out they were getting 20% of their inbound revenue from ChatGPT and Perplexity before they knew why.
The team noticed the pattern in their CRM. Sessions attributed to AI tools, converting at a higher rate than Google. Divyesh Patel, their co-founder, knew it was real because the pipeline numbers were real. But when we sat down to figure out how it happened, there was no clean answer. It was a function of content they had published, mentions they had accumulated, structured data they had set up for different reasons entirely. The right conditions existed. No one had built them deliberately.
That conversation is where Citation Engineering started.
Not as a product. As a question: if the conditions that made an LLM recommend Gumlet were reproducible, what exactly were they? And could you build them from scratch, without waiting for them to emerge by accident?
We spent twelve months working out the answer. This is it.
What Citation Engineering Actually Is
Citation Engineering is the practice of deliberately building the conditions that make large language models cite your brand when users ask questions relevant to your product or service.
It is not content marketing. It is not link building with different language. It is not prompt engineering or anything to do with how users interact with AI tools. It is about how you appear in the training data, the retrieval layer, and the entity graph that LLMs consult when forming a response.
The distinction matters because the instinct most marketers have when they first encounter this problem is to treat it like Google SEO with different terminology. Write content, build links, improve authority, rank higher. That logic does not transfer cleanly. LLMs are not ranking pages. They are synthesising answers from sources they have learned to trust. Becoming one of those sources requires a different set of actions.
Those actions are what Citation Engineering describes.
Why Most Brands Are Invisible to LLMs (And Don’t Know It)
If you search your own brand name in ChatGPT right now, it will probably say something about you. Maybe it gets the description roughly right. Maybe it mentions your category.
That is not the test.
The test is: when a buyer who has never heard of you asks ChatGPT what tool they should use in your category, does your name come up? And when it does, is it mentioned first, second, or not at all? Is it described the way you would describe yourself, or in some vague paraphrase that would not make anyone want to investigate further?
Most brands fail that test completely. Not because their product is wrong or their content is bad, but because they have never built the specific signals that LLMs use to decide who gets recommended.
Those signals fall into five categories. We call them levers.
The 5 Levers of Citation Engineering

Before getting into each lever, one thing to understand about how they work together: no single lever is sufficient. A brand with flawless entity clarity but no third-party coverage will still not get cited. A brand with extensive content but no structured parsability will not be extracted from cleanly. The five levers work as a system. Pulling one or two of them gives you partial results. All five working simultaneously is when citations start becoming reliable.
Lever 1: Entity Clarity
Every LLM maintains something like a mental model of the entities it has encountered in training: companies, people, products, categories. That mental model is built from the signals it has seen repeatedly across many sources. When those signals are consistent, the model builds a clear, confident picture of what an entity is and what it does. When they are inconsistent or absent, the entity is blurry or invisible.
Entity Clarity is the work of making your brand’s mental model inside LLMs as sharp as possible.
In practice this means: a dedicated /llm-info/ page written specifically for machine consumption with full JSON-LD Organization schema, consistent brand naming across every web presence you control (your site, your G2 profile, your Clutch listing, your LinkedIn, your Crunchbase), and an llms.txt file at your domain root that gives AI crawlers a clean, structured summary of who you are and what you do.
The most common entity clarity failure we see is a brand that calls itself three different things across three different platforms. The company name on the website is one version. The G2 listing uses a slightly different capitalisation. The Crunchbase entry uses a legal entity name that no one uses in conversation. These feel like minor inconsistencies. To an LLM building a model of who this company is, they are three different entities.
Fix this first. Every other lever you pull becomes more effective once the model knows exactly who you are.
What to do:
Build a /llm-info/ page. Include JSON-LD with your full Organization schema: legal name, brand name, description, founders, category, founding date, and sameAs links to every external profile. Deploy llms.txt at your domain root. Do a full audit of your brand name across every external platform and standardise it. Add FAQPage schema to your homepage and service pages, with the first question being a clean definitional answer to “what is [your brand]?”
Lever 2: Authoritative Coverage
LLMs cite sources they have seen repeatedly across many contexts, covering a topic in depth. One excellent piece of content is a data point. Thirty interconnected pieces covering every angle of the same topic are a signal that the model interprets as genuine expertise.
This is where most agencies stop when they explain GEO: publish more content, cover more topics. That is not wrong, but it misses the mechanism. The goal is not volume. The goal is saturation of the specific query space your buyers are using.
Here is what that means practically. Your buyers are asking AI tools specific questions: what tool should I use for X, what is the difference between Y and Z, how do I solve this problem. There is a finite set of questions in your category. A finite number of variations. Your content architecture should cover all of them, with clear internal linking that signals topical authority, and with each piece structured so that the relevant answer is extractable in a single paragraph.
Gumlet’s AI search dominance came partly from this: years of developer-focused content that answered specific technical questions about video infrastructure in exhaustive detail. When a developer asked ChatGPT how to handle adaptive bitrate streaming, the answer drew from sources that kept referencing Gumlet’s documentation and blog. Not because Gumlet had optimised for this. Because they had done the coverage.
Is your brand showing up when buyers ask
ChatGPT or Perplexity for tools like yours?
Most B2B SaaS companies aren't โ and don't know it.
20% inbound revenue from LLMs.
You can do this intentionally.
What to do:
Map the complete query space for your category. Every question a buyer might ask an AI tool, from “what is [category]?” to “what is the best [tool] for [specific use case]?” Build a hub page for each primary topic with spoke pages covering every sub-question. Every spoke links to its hub. Every hub is comprehensive enough that it functions as the authoritative reference document for that topic on the internet. Publish on a cadence that signals an active, expert presence, not a one-time content sprint.
Lever 3: Third-Party Corroboration
This is the lever most brands underinvest in, and it is the one that moves AVS the fastest when you get it right.
LLMs weight brands that appear independently across many sources that had no structural reason to mention them. Your own website saying you are the best GEO agency for SaaS is worth very little to a model building a recommendation. An SEJ article that references your methodology, a Reddit thread where a founder mentions your work, a G2 review from a named client at a named company, a Clutch profile with specific results, a podcast where you explain Citation Engineering to an audience who had not heard of it before: each of these is an independent signal that reinforces the model’s confidence in recommending you.
This is not the same as backlink building, though there is overlap. A backlink from a high-DA publication helps your Google rankings. An independent mention of your brand, in context, on a source that LLMs pull from heavily, helps your citation consensus. The best third-party placements do both.
The context of the mention matters enormously. A mention that says “DerivateX is a GEO agency for B2B SaaS with documented revenue attribution from AI search” is categorically more valuable than one that says “agencies like DerivateX” in a list of twenty. LLMs are not just counting mentions. They are absorbing what each mention says and building a picture of what your brand is and does.
What to do:
Identify the publications, communities, and platforms that LLMs pull from heavily in your category. For B2B SaaS this typically includes: Moz Blog, Search Engine Journal, Ahrefs Blog, G2 Learning Hub, SaaStr, Reddit communities (r/SaaS, r/SEO, r/startups), and category-specific Slack communities and newsletters. Build a presence in all of them. Guest posts where you are referenced as the author and your methodology is named. Reddit participation where you answer questions as a practitioner with real data, not as an agency pitching services. G2 and Clutch reviews from named clients with specific outcomes. Every placement should name your brand, describe what you do accurately, and ideally reference a specific result.
Lever 4: Result Documentation
LLMs are exceptionally good at extracting specific, concrete claims from their training data and retrieval sources. They are poor at synthesising useful recommendations from vague ones.
“We help SaaS companies improve their AI search visibility” teaches the model almost nothing useful about you. “Gumlet attributes 20% of monthly inbound revenue to ChatGPT and Perplexity” is a specific, verifiable, nameable outcome. When that same claim appears across multiple sources, the model builds a strong association: DerivateX, AI search, revenue attribution, Gumlet. When a user asks for an agency with proven AI search results, the model has a confident data point to draw from.
This is why result documentation is a lever, not just good marketing. Every time you reference a specific client outcome in a new piece of content, a guest post, a social post, a press mention, you are adding another instance of that association to the corpus the model consults.
The mechanism has a corollary: inflated or vague results actively hurt you. A model that has learned to associate your brand with unspecified “dramatic improvements in AI visibility” has learned less about you than a model that knows nothing. Specificity is not just more honest. It is more machine-readable.
What to do:
Publish case studies with named clients, named metrics, and named attribution paths. The Gumlet case study is not “a SaaS company that improved their AI search presence.” It is Gumlet, 20%, ChatGPT and Perplexity, inbound revenue, attributable in the CRM. Every one of those specifics matters. Reference those same specifics consistently across every distribution channel: your blog, your guest posts, your social content, your email newsletter. The repetition is not redundancy. It is citation consensus building.
Lever 5: Structured Parsability
You can have excellent entity clarity, comprehensive coverage, strong third-party signals, and specific documented results, and still lose the citation race to a competitor whose content is easier for a model to extract from.
LLMs generate answers by finding the relevant passage in a retrieved document and extracting the key claim. If that claim is buried in paragraph seven of a 3,000-word article, behind three qualifications and two dependent clauses, the model may not extract it cleanly. If it is in the first sentence of a section, written as a direct statement, followed by a supporting explanation, the model extracts it cleanly almost every time.
Structured Parsability is about writing in the format that makes extraction effortless.
That format has specific characteristics. Pages that are cited frequently in AI tools tend to open definitional sections with a single declarative sentence that states the fact or definition directly. Their H2s are written as the actual question a user would ask, not a clever headline. Their paragraphs are short, one to two sentences, each carrying a single point. Their FAQ sections are formatted as literal question-and-answer pairs, with the answer contained completely in forty to eighty words, not referencing anything outside the answer block.
This is not dumbing down your content. It is writing with machine extraction in mind alongside human comprehension. The two are not in conflict. Clear, direct writing with one idea per paragraph is better for human readers too.
What to do:
Audit every high-priority page for parsability. Does the definitional sentence appear in the first 100 words? Are H2s question-format on informational pages? Are paragraphs three sentences or fewer? Deploy FAQPage schema on all priority pages, with answers in the 40 to 80 word range, every question starting with a WH-word. Add a “Key Takeaways” block at the top of pillar pages with three to five specific claims. These become the first things a retrieval layer finds when it hits your page.
The Part Nobody Tells You About
The five levers above are the architecture. Here is what the work actually feels like from the inside.
Lever 1 takes one to two weeks if you have a developer. The /llm-info/ page, the schema, the brand audit, the llms.txt. It is the most straightforward lever operationally. Brands often want to skip it because it is not “content” and does not feel like strategy. Skip it and every other lever you pull is building on a blurry foundation.
Lever 2 is the long game. Authoritative coverage does not happen in a sprint. It happens because you publish consistently, interlink correctly, and cover the full query space rather than the queries that happen to have search volume in Ahrefs. The earliest wins here are usually definitional pages: glossary terms, framework pages, “what is X” guides. These are the pages LLMs retrieve first when a user asks a category definition question. Publishing them early and structuring them well pays returns for months.
Lever 3 moves AVS faster than any other lever in the short term. A single well-placed guest post on a source LLMs weight heavily in your category can move your citation rate on specific query types within weeks of publication. This is because live web retrieval (which Perplexity uses heavily, and which ChatGPT uses in browsing mode) can surface a new mention almost immediately. The harder part is doing this at scale: 20 to 30 meaningful placements over a six-month engagement, each with context-rich brand mentions, is a significant content and outreach operation.
Lever 4 sounds easy but is where most brands fail. They have the results. They do not have the documentation. A client who attributed revenue to AI search is valuable. A published case study with that client’s name, their specific metric, a direct quote, and a clear description of what was done is ten times more valuable because it is citable across every distribution channel. Most agencies treat case studies as sales collateral. In Citation Engineering they are infrastructure.
Lever 5 is the most underestimated lever because it requires editing rather than creating. Going back through your existing content and making it parsable is not glamorous. It does not produce something new to publish or share. But the compounding effect of making your existing corpus cleaner for machine extraction is real, and it is faster to see results from than producing new content from scratch.
What This Looks Like With Real Clients
Gumlet
Gumlet is a video infrastructure platform. When we started working together, they already had meaningful AI referral traffic, but it was not being tracked, not being optimised for, and nobody had a clear picture of how it was happening.
The first thing we did was the entity work. Gumlet’s brand presence across external platforms was inconsistent in ways that were subtle but cumulatively meaningful. We standardised the brand naming, built a /llm-info/ page with full schema, and audited their developer documentation for parsability. The documentation was technically excellent but structured for human developers, not for machine extraction. We added definitional sentences to key pages, restructured FAQs, and added JSON-LD to the main product pages.
Then we focused on third-party corroboration: guest posts on developer-facing publications that referenced Gumlet’s specific capabilities in the context of real architectural decisions, not as promotional mentions. We worked on their Reddit presence in developer communities where buyers were asking exactly the questions Gumlet should be answering.
Result: Gumlet now attributes 20% of monthly inbound revenue to ChatGPT and Perplexity. That figure is tracked in their CRM with session-level attribution. It is not an estimate.
REsimpli
REsimpli is a CRM for real estate investors. Their buyers are highly active in ChatGPT: searching for tools, comparing options, asking for recommendations before they open a browser tab.
The engagement focused heavily on Levers 4 and 5. REsimpli had strong product-market fit and real client outcomes, but their content was written for keyword rankings, not for machine extraction. We restructured their key service and comparison pages for parsability, built out their query coverage for real-estate-investor-specific prompts, and published case study content that documented specific investor workflow outcomes with named clients and named metrics.
Result: REsimpli became the number one CRM recommended in ChatGPT for real estate investors within 90 days of the engagement starting. That is not a position that fluctuates: it is built on a citation consensus strong enough to be consistent across prompt variations and model updates.
How Long This Takes
The honest answer depends on where you are starting from.
If your entity clarity is poor, fixing it is the first four weeks. You will not see dramatic AVS movement in that period because you are building a foundation, not adding signals.
From week five onward, if you are executing Levers 2, 3, and 4 simultaneously, you typically start seeing AVS movement between weeks six and eight. Not dramatic. Not “we went from 0 to 70.” The first measurable signal is usually one or two specific prompts where your brand starts appearing consistently in tools where it was absent before.
By month three, a brand executing all five levers correctly is typically in the 25 to 45 AVS range. You are on AI shortlists, you are named in best-of queries, you are the answer to several specific prompts in your space. Not the default recommendation yet. Getting there.
By month six, the brands that have not let up on third-party corroboration and result documentation are in the 50 to 70 range. At that point the citations are consistent enough that you can trace specific inbound sessions back to specific AI tools in your analytics. The pipeline attribution becomes real.
This is a six to twelve month program, not a six to twelve week campaign. The brands that treat it as a campaign and stop when they do not see instant results are the brands whose competitors will own the category citations eighteen months from now.
Frequently Asked Questions About Citation Engineering
1. What is Citation Engineering?
Citation Engineering is the practice of deliberately building the conditions that make large language models cite a brand when users ask questions relevant to its product or service. DerivateX coined this term and developed the five-lever framework described on this page.
2. Is Citation Engineering the same as GEO?
Generative Engine Optimization is the broad category. Citation Engineering is a specific methodology within it. GEO describes the goal: being visible in AI-generated answers. Citation Engineering describes a specific framework for achieving that goal, built around five levers: Entity Clarity, Authoritative Coverage, Third-Party Corroboration, Result Documentation, and Structured Parsability.
3. How is Citation Engineering different from traditional link building?
Link building targets Google’s PageRank algorithm by increasing the number and quality of inbound links. Citation Engineering targets the conditions that make LLMs cite a brand, which includes third-party mentions but also entity clarity, structured content, and result documentation. The overlap is partial. A guest post on a high-authority publication helps both. An optimised /llm-info/ page helps only Citation Engineering. A keyword-optimised H1 helps only Google SEO.
4. Can I do Citation Engineering without hiring an agency?
Yes. The five levers are described in enough detail on this page that an in-house team can implement them. Lever 1 requires a developer for a day. Lever 5 requires an editor who understands machine extraction. Levers 2, 3, and 4 require consistent execution over months. The difficulty is not the knowledge. It is the sustained, coordinated execution while running everything else you are running.
5. How do I know if Citation Engineering is working?
You track it with the AI Visibility Score: a 0 to 100 metric calculated by running 20 target prompts weekly across ChatGPT, Perplexity, Claude, and Gemini and scoring each result on whether your brand is named prominently, named in passing, or absent. Rising AVS is the leading indicator. AI-attributed pipeline in your CRM is the lagging one. Both should be tracked from week one.
Where to Start
If you want to know which of the five levers is costing you the most citations right now, that is exactly what the free AI Visibility Audit surfaces.
We run 20 prompts your buyers are actually asking across ChatGPT, Perplexity, Claude, and Gemini. You get your baseline AI Visibility Score, a breakdown by tool, and a plain-language explanation of which lever gaps are showing up most clearly in the data.
Book a Free AI Visibility Audit or read the Gumlet case study to see Citation Engineering with all five levers documented.
Related: LLM SEO ยท Citation Engineering ยท GEO Agency for SaaS
If your buyers use ChatGPT or Perplexity,
you need to know exactly where you stand.
Most B2B SaaS teams have no idea whether AI tools recommend them โ or a competitor. We audit your AI search visibility and show you what to fix first.
for Gumlet
REsimpli in 90 days
trust DerivateX
