There is no universal pharma citation playbook — because different AI engines draw from structurally different source pools. In oncology, Perplexity's top three citation domains are all NCCN. In metabolic disease, FDA labels dominate. In inflammatory dermatology, society guidelines share the top positions with peer-reviewed literature and specialist hubs. The source architecture is TA-specific, engine-specific, and, as the brand-mention correlation data shows, strongly predictive of which brands get cited. This is the methodology that underpins the PharmaGEO measurement framework, and it is the primary reference for any GEO strategy built on the May 2026 PharmaGEO public index data.

The four source archetypes: not all authority is equal across engines

Before examining TA-specific citation data, it is worth establishing the taxonomy of source types that drives retrieval across pharma queries (insight E1, May 2026 PharmaGEO public index). Four archetypes account for the majority of pharma citations across Perplexity, ChatGPT, and Gemini:

Source archetype Example domains Citation strength Primary TAs
Society guideline aad.org, nccn.org, ESC, ESMO, EASD Highest in regulated TAs Oncology, inflammatory dermatology, cardiology
Regulatory primary accessdata.fda.gov, ema.europa.eu (EPAR), nice.org.uk Heavy in metabolic / safety-loaded TAs Obesity, GLP-1, rare disease, any REMS product
Peer-reviewed literature nejm.org, pmc.ncbi.nlm.nih.gov, thelancet.com Universal floor across all engines and TAs All therapeutic areas
Specialist hub / disease-state psoriasis-hub.com, dermnetnz.org, aafp.org Engine-dependent; Perplexity favours when peer-review coverage is thin Inflammatory dermatology, allergy, primary care TAs

The archetype framework reveals the first principle of pharma GEO: there is no generic content type that wins citations across all engines and therapeutic areas. A content strategy built on society guideline alignment will perform well in oncology and inflammatory skin disease but will underperform in metabolic TAs, where regulatory labels dominate. The strategy must be calibrated to the specific citation architecture of the target TA on the target engine. And because each engine draws from a structurally different source pool — only 11% domain overlap between ChatGPT and Perplexity, per the Digital Bloom 2025 AI Citation and LLM Visibility Report — the calibration must be engine-specific as well as TA-specific.

Perplexity source data by TA: the most transparent citation layer

Perplexity's explicit source link layer makes it the most auditable engine for citation analysis. The May 2026 PharmaGEO public index records actual citation use counts across all four TAs, and the patterns are sharply differentiated (insight A5):

Therapeutic area Top citation sources (Perplexity, May 2026) Dominant archetype GEO strategy implication
Lung Cancer nccn.org (136+64+18 uses); NEJM KEYNOTE-189 (16 uses); ~218 of ~258 total = nccn.org Society guideline monopoly (84% nccn.org) NCCN inclusion is the single citation lever; no content substitute
Atopic Dermatitis aad.org (168), pmc.ncbi.nlm.nih.gov (94), aafp.org (92) Society guideline + literature + specialist hub Multi-channel; all three archetype types contribute
Psoriasis pmc.ncbi.nlm.nih.gov (154), aad.org (114), psoriasis-hub.com (70), nice.org.uk (40), dermnetnz.org (40) All four archetypes represented Most levers available; single-channel strategy underperforms
Obesity accessdata.fda.gov Zepbound PI (36), Wegovy PI (32), novo-pi.com Wegovy PI (26), nejm.org (24) Regulatory label dominance; manufacturer PI in top 4 Crawlable FDA label and PI are direct citation assets

The brand-mention correlation: what predicts citation before the query fires

An Ahrefs study (December 2025) of 75,000 brands found a correlation coefficient of r = 0.664 between web brand mentions and AI Overviews citation frequency. YouTube mentions show the strongest single-factor correlation at r ≈ 0.737. The finding is directionally consistent with what the PharmaGEO source data shows: brands that are widely mentioned across the indexed web — in literature, in guideline documents, in news coverage, in disease-state hubs — accumulate citation share in proportion to their mention density across multiple independent sources.

The r = 0.664 correlation is not perfect, which means source quality moderates the relationship. A brand with 10,000 low-quality forum mentions will not generate the same citation share as a brand with 5,000 citations spread across society guidelines, PMC papers, and FDA labels. The productive interpretation of the correlation is that it establishes the direction and approximate magnitude of the relationship between indexed brand presence and AI citation — and that this relationship is strong enough to make brand mention strategy an input to GEO planning, not a separate activity. For pharma brands, the practical corollary is that investments that increase brand mentions in authoritative indexed channels — peer-reviewed journals, society meeting abstracts, medical education programmes published on indexed sites — accumulate as a retrievable asset in the citation graph over time, not just as marketing impressions.

The front-loading rule: where citations are drawn within a source

The source-level citation question (which domains get cited) is only half the picture. The other half is where within a source document citations are drawn. Analysis of 1.2 million ChatGPT responses by Kevin Indig, published in Search Engine Land, found that 44.2% of citations come from the first 30% of source content. The majority of what gets cited is what appears earliest in a document.

For pharma content strategy, this rule has a direct application: every page that is intended to contribute to GEO citation share — whether an owned brand site page, a PMC-indexed journal abstract, or a disease-state educational hub article — should front-load its primary claim. The FDA label front-loads boxed warnings and contraindications, which is partly why the label is such a high-citation source: the most safety-critical content appears first, in the most heavily sampled 30%. Brand content that buries its clinical conclusion in a fourth section is being sampled primarily for its introductory framing, not for its data.

The Reddit and Wikipedia problem: pharma's compliance gap in the citation graph

The Everything-PR AI Platform Citation Source Index 2026, covering 680 million total citations across AI engines, identifies a structural tension specific to pharma: Reddit accounts for approximately 40% of ChatGPT citations, and Wikipedia for 26–48%. These figures represent the general web, not pharma specifically — but they point to a compliance gap. Reddit and Wikipedia are not MLR-reviewed channels, and if ChatGPT's pharma answers draw on the same citation architecture as its general answers, the pipeline from forum content into AI-generated drug information is a compliance exposure. The operational implication: a brand with dense PMC coverage, guideline presence, and well-indexed owned content will have a smaller proportion of its citation share driven by forum sources than one whose authoritative content is thin or gated behind PDFs. Source quality strategy is also source dilution strategy.

EMA EPARs: the underused citation channel pharma already owns

One of the most consistently underused citation vectors in pharma GEO is the EMA European Public Assessment Report (EPAR). The May 2026 PharmaGEO public index shows EMA EPARs appearing in the top sources for five brands in the atopic dermatitis top 10 and two in the psoriasis top 10 — in English-language answers (insight E2).

EPARs are authoritative English-language regulatory documents hosted at ema.europa.eu, a high-trust domain. They are freely accessible and contain safety, efficacy, and indication data that aligns with the archetypes retrieval engines favour. The barrier is not competitive — it is that brands have not ensured EPAR pages are correctly indexed and that the EPAR URL appears as an inline citation in their own disease-state content. A brand that links to its own EPAR from its owned content creates a citation path from owned content through one of the most trusted regulatory documents in the European pharma ecosystem. Given that only 11% of domains overlap between ChatGPT and Perplexity citations, a brand that correctly surfaces its EPAR on Perplexity gains citation share in a pool that ChatGPT-focused competitors are not contesting — at zero incremental content creation cost.

The older-drug literature moat: why publication depth is a durable competitive advantage

The May 2026 Perplexity data for atopic dermatitis includes Protopic at SOV rank 4 (29.9%), Elidel at rank 6 (20.6%), and Eucrisa at rank 7 (13.4%) — drugs from the early 2000s (insight A6, May 2026 PharmaGEO public index). OpenAI's equivalent positions are occupied by newer JAK inhibitors. The reason these older brands maintain high Perplexity SOV is not their current clinical dominance but their accumulated literature depth: two decades of PMC-indexed publications, guideline mentions, and specialist hub references have created a citation density that newer brands with 3–5 years of evidence cannot match on a pure volume basis.

This is the literature moat — a structural citation advantage that compounds over time. For newer brands, it has two implications. First, the novelty penalty in retrieval is real: a brand launching without 5+ years of literature pays a retrieval cost relative to established alternatives in the same class, particularly on Perplexity, which follows the published literature most directly. Second, the remedy is time and publication quality, not a content sprint. The fastest path to eroding a competitor's literature moat is not trying to out-publish them in volume but publishing in the highest-indexed channels — NEJM, Lancet, and PMC-open-access journals — where a smaller number of citations carries disproportionate weight in the retrieval graph because these domains are the universal-floor sources that every engine trusts, regardless of TA. A single NEJM publication generates more retrievable citation surface than 50 press releases because the engine recognises the source archetype, not just the URL.

How citation propagation works: the chain from press release to stable anchor

Citation propagation follows a consistent chain. A press release on a Phase 3 result, picked up by three or more authoritative medical outlets with independently indexed coverage, begins appearing in Perplexity answers within days — Perplexity's real-time retrieval layer indexes these quickly. If those articles are then cited in a medical society commentary or educational piece, the claim moves into the content that ChatGPT's retrieval draws on when browsing is active. If the underlying trial is published in a PMC-accessible peer-reviewed journal, the claim becomes a stable citation anchor across all three major engines. Brands that skip steps — that publish a press release without ensuring medical outlet syndication, or that publish a trial without ensuring open PMC access — find citation propagation stalls at the early stages. Syndication is not a supplementary activity to content creation; it is the mechanism by which owned content enters the broader citation graph and becomes durable rather than a one-off retrieved result. The investment in each propagation step is the investment in long-term retrievability, not just in short-term news coverage.

What happens when sources disagree: engine behaviour under citation conflict

Source disagreement is common in pharma: a Phase 3 result described accurately in a journal publication, paraphrased inaccurately in a news summary, and further distorted in a patient forum may all be indexed simultaneously. ChatGPT synthesises toward the most common framing — two inaccurate versions can dilute one accurate one. Perplexity surfaces disagreement more explicitly. Gemini defaults to the most recently indexed version regardless of accuracy. Brands with a high volume of slightly inaccurate recent third-party coverage face the greatest accuracy risk on Gemini; brands with accurate but sparse coverage face the greatest risk on ChatGPT. Monitoring and correcting inaccurate third-party content is as important as creating accurate owned content — the accuracy index is determined by the entire indexed ecosystem, not the owned fraction alone.

Want a real audit on your brand? Request a sample report or get the full PharmaGEO Playbook.