Research & Papers

From Citation Selection to Citation Absorption: A Measurement Framework for Generative Engine Optimization Across AI Search Platforms

Perplexity and Google cite more sources, but ChatGPT's answers are more influenced...

Deep Dive

A new arXiv preprint from researchers Zhang Kai and Yao Jingang introduces a measurement framework for Generative Engine Optimization (GEO) that goes beyond simple citation counts. The study analyzes the geo-citation-lab dataset covering 602 controlled prompts across ChatGPT, Google AI Overview/Gemini, and Perplexity, with 21,143 search-layer citations and 23,745 citation-level feature records. The key finding: citation breadth and depth diverge significantly across platforms.

Perplexity and Google cite more sources on average, but ChatGPT shows substantially higher average citation influence among fetched pages. High-influence pages tend to be longer, more structured, semantically aligned, and rich in extractable evidence such as definitions, numerical facts, comparisons, and procedural steps. The authors argue that GEO should measure answer-level absorption as a separate outcome from citation counts, providing a more nuanced understanding of how AI search engines actually use content.

Key Points
  • Study analyzed 602 prompts across ChatGPT, Google AI Overview/Gemini, and Perplexity with 21,143 citations
  • ChatGPT cites fewer sources but shows higher average citation influence than competitors
  • High-influence pages are longer, structured, and rich in extractable evidence like definitions and numerical facts

Why It Matters

Content creators must optimize for absorption, not just citations, as AI search engines prioritize evidence-rich pages.