GEO citation business-outcome evidence — what’s actually measurable
The question
Beyond Princeton’s GEO-bench paper, what’s the actual measured business-outcome evidence that LLM citations (in Claude / ChatGPT / Perplexity / Google AI Mode responses) drive trackable downstream traffic, signups, or revenue — what attribution methodologies exist (Profound, Peec.ai, AthenaHQ, others), and what’s the credibility floor of their published case studies?
What we already know (from the vault)
- The 2026-04-22-agent-seo-state-of-the-discipline brief explicitly flagged this as the load-bearing evidence gap: Princeton’s KDD 2024 paper measures visibility-in-response (Position-Adjusted Word Count on GEO-bench), not click-through, signup, or revenue. Every business-outcome claim downstream of “appeared in LLM response” is currently anecdotal vendor case study.
- The brief catalogued the tooling landscape — Profound ($99–$399), AthenaHQ (YC-backed), Peec.ai (Berlin Feb 2025 launch), Otterly ($39), Scrunch / Adobe / Semrush — and noted “monitoring is real and useful; ‘optimization’ is mostly automated content-rewrite loops; the optimization work itself is still mostly editorial judgment.”
- The 2026-04-23-generative-engine-optimization-geo canonical concept doc confirms GEO is the founder’s chosen term and frames the underdog effect (rank-5 +115% from Cite Sources) as the strategic insight for raydata.co’s zero-authority regime.
- The 2026-04-22-publishing-for-agents-spec companion piece commits the founder to schema.org + DefinedTerm and “copy for agents” UX as the publishing pattern — but those are inputs to the citation funnel, not measurements of its outputs.
- Skeptic position from Rand Fishkin / SparkToro already in vault: AI search interest is inflated 10–100x relative to actual usage; the discipline is being sold faster than it is being used.
What the web says
- Zero-click is the dominant outcome. 93% of AI search sessions end without a website click; 75% for Google AI Mode specifically (thestacc.com 2026 stats). The base rate of “citation → traffic” is roughly 7% at best, before any conversion math.
- Conversion rates on the traffic that does click are reportedly 4–17x higher than organic search. Microsoft Clarity’s analysis of 1,277 publisher/news domains reported LLM-aggregate Sign-Up CTR of 1.66% vs Search 0.15% (~11x); Copilot 17x, Perplexity 7x, Gemini 4x (Microsoft Clarity blog). Opollo’s 312-firm IT/tech study reported AI 14.2% vs Google organic 2.8% (~5x).
- No platform’s case study traces a complete funnel. Of the five vendor studies reviewed, zero present an end-to-end “appeared in LLM response → user clicked → user signed up → user paid” with reliable attribution. AthenaHQ’s flagship 30-day showdown explicitly stops at “answer share” and “content mentions” (AthenaHQ showdown page — see methodology dissection below).
- Attribution is fundamentally broken at the platform level. Most AI platforms strip referrer headers; ChatGPT, Claude, and Gemini answers that are read-and-not-clicked produce zero GA4 signal; click-through traffic frequently lands as Direct rather than as identifiable AI referral. The category is honestly described as “directional surveillance, not precision analytics” by Meltwater, one of the more candid vendors (Meltwater LLM tracking).
- The methodology gap is structural, not technical. Vendor monitoring tools use either synthetic prompt sampling (run 1,000 prompts/day, parse responses for brand mentions) or client-side mimicry (simulate browser sessions). Both measure share of mention in LLM output, not user behavior. They cannot measure who saw the citation, who clicked, who converted — because those events do not pass back through the LLM provider’s API.
- Independent academic measurement of business outcomes is missing. Princeton’s GEO-bench (Aggarwal et al., KDD 2024) is still the only rigorous public study, and it measures visibility, not downstream conversion. No follow-up paper from Princeton, Stanford, MIT, or any major lab has measured citation-to-revenue attribution as of 2026-05-09.
- Aggregate AI-referred traffic is still <1% of total traffic for most sites, per Microsoft Clarity’s own data. Even the “4–17x conversion rate” multiplier is small absolute volume.
Attribution platforms reviewed
Profound ($99 / $399 / Enterprise)
- What it measures: Share-of-answer across 10+ AI engines (ChatGPT, Claude, Perplexity, Gemini, Copilot, DeepSeek, Grok, Meta AI, Google AI Mode, AI Overviews).
- Methodology: Client-side mimicry — simulates browser sessions to capture the full rendered LLM response including local packs and injected content. Marketed as “real user-facing data from front-end interactions” but the prompts are still synthetic; “real” means real-browser-render, not real-user-behavior.
- Published case studies: Profound’s blog publishes share-of-answer movement charts. None of the cases I could find connect share-of-answer to signups or revenue with reliable attribution.
- Credibility assessment: Best-in-class for the monitoring layer. Honest about what it measures (visibility in LLM output). Does not claim to measure business outcomes. The 10+ engine coverage is the strongest differentiator.
AthenaHQ ($595/mo, YC-backed)
- What it measures: Same surface as Profound (answer share, mentions) plus an automated content-rewrite “optimization” loop and an Athena Citation Engine (ACE) that scores content for citation probability.
- Methodology: 1,000 simulated buyer questions per platform across B2B SaaS / Pro Services / E-commerce, daily monitoring, weekly deep analysis. Tested brands and websites are not disclosed, so the test is not independently replicable.
- Published case studies: The flagship “30-day showdown” page claims +45% answer share gain (vs Peec.ai +8%, Profound -1%) and a “75.6x ROI” by dividing $13/percent-gain into the 45% lift. The ROI math is fictional — it asserts a percent-gain-to-revenue conversion factor that the test did not measure. The page measures answer share only; zero linkage to signups or revenue.
- Credibility assessment: Vendor self-promotion masquerading as research. Author is the winning vendor; competitor brands are mentioned by name; tested brands are anonymized; no third-party audit; no business-outcome traceability. The “75.6x ROI” claim should be treated as marketing copy, not evidence.
Peec.ai ($199/mo, Berlin)
- What it measures: Brand mention tracking and visibility analytics across major AI engines.
- Methodology: Pure analytics layer — clean UX, but the same fundamental measurement (synthetic prompts, share-of-answer).
- Published case studies: Limited public material. Cleanest UX for early-stage teams per the April brief.
- Credibility assessment: Low marketing-noise vendor; the cheapest credible monitoring entry. Same evidence ceiling as Profound — measures presence, not outcomes.
Otterly ($39/mo)
- What it measures: Mention tracking across ChatGPT, Google AI Overviews, Perplexity. “Mention.com for AI.”
- Credibility assessment: Beginner tier. Same monitoring-only ceiling.
Scrunch / Adobe LLM Optimizer / Semrush AI Visibility / Bluefish
- What they measure: AI-visibility modules bolted onto existing SEO suites.
- Credibility assessment: Useful if you already pay one of them; otherwise their evidence quality is no better than the focused players.
LLM Pulse, ALM Corp’s “2 million sessions” reports, Meltwater
- A second tier of vendors publishing aggregate “AI search statistics” reports. Useful for trend signal; same underlying methodology limit (synthetic prompts or third-party clickstream that can’t trace conversion to citation).
Convergences and contradictions
Where the platforms agree (and are probably right):
- Share-of-answer in LLM output is measurable with synthetic-prompt sampling — this is real, replicable, and useful as a directional signal.
- LLM citations do meaningfully drive traffic for some categories of brand and query, especially long-tail / underdog domains (consistent with Princeton’s rank-5 +115% finding).
- Click-through traffic from AI assistants converts at materially higher rates than organic search (4–17x, depending on study) — this is reported by Microsoft, Opollo, and Contentsquare independently, with different methodologies, so the directional signal is robust even though the multipliers vary widely.
Where they disagree or fail to deliver:
- None of them publish a case study tracing “citation → click → signup → revenue” with reliable attribution. The full funnel is unmeasured.
- Vendor showdowns (AthenaHQ vs Profound vs Peec) show wildly different “winners” depending on who runs the test — strong evidence the comparison methodology is not stable.
- Conversion-rate multipliers differ by 5–10x across studies (4x to 23x), suggesting the underlying samples are highly heterogeneous and category-dependent.
Where independent analysis cuts against vendor claims:
- 93% zero-click rate means the denominator for any LLM-citation-to-conversion ratio is misunderstood by every vendor case study that measures only the click-through tail.
- Microsoft Clarity is the closest thing to independent measurement (1,277 domains, 8 months) but is still a vendor blog post promoting Microsoft’s own tracking tool, with no published methodology document, no peer review, and no defined “smart events for conversion detection” — the credibility floor is “directionally trust the multiplier, don’t trust the precision.”
- Most AI-influenced traffic shows up as Direct in GA4 because referrer headers are stripped; the entire attribution category is structurally undermeasuring its own claimed outcomes.
- No academic follow-up to Princeton has published business-outcome attribution. The evidence base for “GEO drives revenue” is currently zero rigorous studies + a stack of vendor case studies that don’t measure revenue.
Synthesis for RDCO
Recommendation: do not invert the engine. Augment X-first with a small, time-boxed GEO test and treat the discipline as an editorial constraint, not a load-bearing distribution bet.
The honest read of the evidence:
- Princeton’s paper measures visibility-in-response, not revenue. The Princeton-validated techniques (Quotation Addition, Statistics Addition, Cite Sources) are good writing techniques regardless. Adopting them costs little and the underdog effect is real for raydata.co’s zero-authority regime — that’s the April brief’s call and it still stands.
- No vendor has demonstrated the full funnel. Every “GEO drives revenue” case study reviewed here either (a) measures share-of-answer and asserts ROI by extrapolation (AthenaHQ), or (b) measures click-through conversion rates without proving the click came from the LLM citation (Microsoft Clarity, Opollo). The credibility floor on “GEO → revenue” is currently zero rigorous public studies.
- The base rate is brutal. 93% zero-click means the visibility-to-traffic ratio is ~7%. Even with a 10x conversion multiplier on the click-through tail, the absolute volume is small until LLM usage scales 5–10x further.
- Inverting the engine would be a high-cost, low-evidence bet. X-first delivers measurable engagement today; blog-first-for-LLM-citation delivers an unmeasurable signal tomorrow. The asymmetry of evidence does not support inversion.
The cheapest reversible test before committing:
- Ship 3–4 keystone reference pieces over 60 days, built to GEO standards (question H2s, 40–80 word answer capsules, ≥3 sourced statistics, ≥5 outbound citations, FAQPage + Article + Person schema, llms.txt, named concepts coined from the vault — MAC, macro-vs-micro data quality, the agent-deployer thesis).
- Stand up Peec.ai or Profound monitoring at the cheapest tier ($99–$199) for those 60 days. Seed with 20–30 brand and concept queries (raydata.co, Ben raydata, “macro vs micro data quality,” “model acceptance criteria,” etc.). Baseline at week 0; re-baseline at week 8.
- Concurrently track GA4 referrer-stripped Direct traffic spike, AI-referrer hostnames, and signup conversions tagged by suspected AI-influence (a “where did you hear about us” question on signup is currently the only reliable instrument).
- Decision rule at day 60: If share-of-answer for ≥2 coined concepts moves from 0 to non-zero AND raydata.co Direct traffic shows a measurable lift correlated with publish dates, expand to 6 pieces/quarter cadence. If neither moves, the thesis is wrong for raydata.co’s current authority and the right move is Reddit/HN/podcast guesting until authority exists. Do not go all-in on blog-first before this gate.
The X engine is a known good. Don’t break it on a thesis the evidence base can’t carry yet. The vault contains ~30 pieces of raw material that compound into long-life reference content; converting them at 1–2 pieces/month is cheap insurance against the GEO bet being right, while preserving the X cadence that’s already working. The asymmetric upside (Princeton’s underdog effect) makes the test worth running. The asymmetric downside (paying tooling tax, writing for the wrong audience, killing the working surface) makes inversion the wrong shape.
Watchlist signal that would change this answer: if a peer-reviewed paper or a Stanford / MIT / Princeton follow-up publishes a controlled study tracing LLM citation → conversion with real attribution, revisit this immediately. That paper does not currently exist. When it does, the evidence threshold for inversion will be met. Until then, augment, don’t pivot.
Open follow-ups
- Run the cheapest reversible test above — pick Peec.ai or Profound, seed 20–30 queries, baseline now (May 2026), re-measure July 2026.
- Build the “where did you hear about us” instrument on Sanity Check signup and any future raydata.co conversion surface — this is the only direct way to capture LLM-attributed signups absent platform support.
- Audit raydata.co server logs for GPTBot / ClaudeBot / PerplexityBot crawl frequency; if the bots aren’t crawling, the GEO bet’s input layer is broken regardless of content quality.
- Track the Princeton group + adjacent labs for a follow-up paper measuring citation-to-conversion attribution. That’s the rigorous-evidence threshold for engine inversion.
- Open question for product: is there a market for an open-source GEO monitoring stack, given the closed-source vendor tax? Could be a future RDCO bet, not just a tool to buy.
- Re-examine in October 2026. AI search usage is doubling roughly annually; the base-rate math (93% zero-click) and absolute traffic volumes will look different in 6 months.
Sources
- Princeton GEO paper (Aggarwal et al., arXiv 2311.09735, KDD 2024) — the only rigorous study; measures visibility, not revenue
- Microsoft Clarity — AI Traffic Converts at 3x the Rate of Other Channels — 1,277 domains, 8 months; closest to independent; methodology gaps disclosed in critique above
- AthenaHQ — 30-day GEO Platform Showdown — vendor self-promo; measures answer share only, no business outcome traceability
- thestacc — AI Search Referral Traffic Statistics 2026 — source for 93% zero-click base rate
- Microsoft Clarity blog (publisher-side conversion data) — Copilot 17x / Perplexity 7x / Gemini 4x signup-conversion claims
- Opollo — 2026 AI Search Benchmark Report — 312-firm IT/tech study; 14.2% AI vs 2.8% organic conversion
- Meltwater — How to Track LLM Visibility — “directional surveillance, not precision analytics” — most candid vendor admission
- Profound — Best GEO Tools 2026 — Profound’s own positioning; 10+ engine coverage
- Writesonic — Profound vs PEEC vs AthenaHQ — third-party-ish vendor comparison
- Contentsquare — What Is AI-Referred Traffic? 2026 Benchmarks — independent-ish aggregate benchmarks
- SparkToro / Rand Fishkin — AI search overstated — skeptic baseline from the April brief
- Prior vault: 2026-04-22-agent-seo-state-of-the-discipline, 2026-04-23-generative-engine-optimization-geo, 2026-04-22-publishing-for-agents-spec