Summarize this article with AI
Before you optimize a single page, you need to know where the AI is reading from. ChatGPT pulls heavily from Bing’s top 10 and Wikipedia. Perplexity pulls from Reddit, G2 and live web. Claude pulls from technical docs, research papers and B2B reports. Gemini pulls from Google’s index. Grok pulls from X. Google AI Overviews pulls from the SERP top 50. The source mix per engine is not symmetric, and the brands that win show up in the right sources for the right engines.
Below: the 4-step reverse-engineering anyone can run in 30 minutes, the 6 source types AI engines actually pull from, the measured per-engine weights, and the editorial moves to engineer your appearance in those sources.
Why source mapping is the foundational GEO research
Backlink research, keyword research, competitor analysis: all of those map a space. Source mapping maps a channel. It tells you which third-party platforms an AI engine retrieves from, weighted by the engine’s training and indexing logic.
- AI citation sets are narrower than SERPs. Google’s top 10 surfaces ~10 domains per query. AI answers pull from 3 to 6.
- Sources rotate. 40 to 60% of cited domains change every month.
- Per-engine preferences are stable. While individual sources rotate, the source-type mix per engine is remarkably consistent.
The 4-step source mapping
Pick 10 prompts in your category
Run each prompt in 4 engines
Tag each cited URL by source type
Compute the source-type mix per engine
The 6 source types AI engines pull from
| Source type | Examples | What makes it citation-friendly |
|---|---|---|
| Owned-domain content | Your blog, comparison pages, docs | Direct control, but only 9 to 15% of total citation share |
| Third-party content | Reddit threads, newsletters, podcasts | Highest share for commercial-investigation queries |
| Reviews | G2, Capterra, TrustRadius | High weight in B2B, refreshable quarterly |
| Encyclopedic | Wikipedia, Wikidata | Wikipedia accounts for 27% of ChatGPT citations |
| Technical | Documentation, GitHub, arxiv papers | Heavy weight in Claude for B2B SaaS |
| News / media | Forbes, TechCrunch, Bloomberg | High weight in Gemini and Grok, ephemeral |
The 9 to 15% owned-domain ceiling is structural. Most B2B teams optimize only the 9 to 15% and wonder why their citation share won’t move. Source mapping fixes that gap.
Per-engine preferences (measured weights)
| Engine | Top source type | Second | Third |
|---|---|---|---|
| ChatGPT | Bing top 10 (87% match) | Wikipedia (27%) | Reddit (15%) |
| Claude | Owned domain (9.1%) | Technical / B2B docs | Comparison content |
| Perplexity | Reddit (18-25%) | G2 / Capterra (12-18%) | Wikipedia (8-12%) |
| Gemini | Google index | Schema-rich pages | News / media |
| Grok | X / Twitter | Real-time news | |
| Google AI Overviews | Top SERP top 10 (38%) | Top SERP top 100 (62%) | YouTube (18%) |
Three implications:
- For Claude, optimize your owned domain hardest. Highest owned-domain rate of the 6 engines.
- For Perplexity, win Reddit and G2. 30 to 43% of Perplexity’s citation share lives there.
- For ChatGPT, win Bing and Wikipedia. Bing top-3 rank is the single largest lever.
How to engineer your appearance in those sources
Owned-domain (Claude move)
Reddit (Perplexity move)
G2 / Capterra / TrustRadius (Perplexity + ChatGPT move)
Wikipedia (ChatGPT move)
Newsletter mentions (Perplexity + Gemini move)
Technical / B2B docs (Claude move)
What’s next
For the framework to identify which competitors win which sources, read Competitor Citation Analysis.
For the tactic-level moves to earn citations, read How to Get Cited by AI Search Engines.
For the cross-engine pillar, read How to Do GEO in 2026.
The brands that win in AI search aren’t the loudest. They’re the ones who show up in the right sources at the right time.







