01Study Overview

Background and Objectives

"Does a strategy that works for ChatGPT also work for Gemini?" — this is the first question any company pursuing GEO (Generative Engine Optimization) must confront. Yet, to our knowledge, no Japanese-language study has measured the degree of citation source overlap across engines using large-scale data.

In this study, we measured the overlap rate of domains cited by four major AI search engines in response to identical queries, and conducted a structural analysis of "at which layer" and "to what degree" commonality exists.

Dataset

Table 1 — Study Dataset Overview
ItemValue
Domain citations (domain_citations)9,891
Observation records (observations)1,793
Target keywords37
Measurement period22 days (Feb 27 – Mar 29, 2026)
Engines analyzedChatGPT / Gemini / Perplexity / Google AI Overviews
Note: This study was independently conducted by the GEO Research Team at LIFE, Inc. The 37 target queries span multiple industries including beauty, HR SaaS, finance, recruitment, and education, all in Japanese. Queries were submitted to each engine in a manner that replicates the human search experience.

02Domain Citation Overlap — Approximately 10%

Identical queries were submitted to all four engines, and the Jaccard coefficient (intersection over union) of cited domain sets was calculated for each engine pair. The results are as follows.

Domains cited by all 4 engines: only 81 out of 1,843 unique domains (4.4%)

The proportion of domains cited by all four engines for the same query on the same day averaged 1.2%, with a median of 0.0%.

Pairwise Overlap Rates

Table 2 — Pairwise Domain Citation Overlap (Mean Jaccard Coefficient per Query-Day)
Engine PairMean JaccardMedian
Perplexity × AIO18.9%16.7%
Gemini × Perplexity13.2%11.1%
Gemini × AIO10.8%9.1%
ChatGPT × Perplexity6.9%0.0%
ChatGPT × Gemini6.2%0.0%
ChatGPT × AIO6.0%5.0%
Figure 1 — Domain Citation Overlap Heatmap (Mean Jaccard per Query-Day)
ChatGPT
Gemini
Perplexity
AIO
ChatGPT
6.2%
6.9%
6.0%
Gemini
6.2%
13.2%
10.8%
Perplexity
6.9%
13.2%
18.9%
AIO
6.0%
10.8%
18.9%

The Structural Isolation of ChatGPT

The heatmap reveals a striking pattern: ChatGPT references markedly different sources from the other three engines. The overlap between ChatGPT and every other engine falls in the 6–7% range, and the median for ChatGPT × Gemini is 0.0% — meaning that for more than half of all queries, not a single domain was shared between the two.

By contrast, Perplexity × AIO exhibits the highest overlap at 18.9%, which is attributable to their shared reliance on Google Search as a foundational index. Even so, 18.9% means that approximately 80% of cited domains remain distinct.

Figure 2 — Structural Relationships Among Engines
AIO 18.9% Perplexity 10.8% – 13.2% Gemini
Google Search Infrastructure Group
← 6–7% →
ChatGPT
Independent Sources

03Brand Recommendation Overlap — Approximately 14%

Beyond domains (source URLs), we also measured the overlap rate of brand names recommended by each engine. Brand names were extracted from 998 response texts, and Jaccard coefficients were calculated in the same manner.

While domain overlap averages approximately 10%, brand overlap averages approximately 14%.

Brand overlap exceeded domain overlap across all pairs (with the exception of Perplexity × AIO), by an average of +4.6 percentage points.

Table 3 — Domain Overlap vs. Brand Overlap (Mean Jaccard per Query-Day)
Engine PairDomainBrandDifference
ChatGPT × Gemini6.2%11.9%+5.7pt
ChatGPT × Perplexity6.9%12.1%+5.2pt
ChatGPT × AIO6.0%11.4%+5.4pt
Gemini × Perplexity13.2%16.5%+3.3pt
Gemini × AIO10.8%16.4%+5.6pt
Perplexity × AIO18.9%16.6%-2.3pt

This finding carries significant implications. Even when engines consult entirely different domains (information sources), they may arrive at recommending the same brands. In other words, AI recommendations are not contingent on any particular URL; rather, they appear to be shaped by the aggregate of information distributed across the web — what we refer to below as "perception."

04Recommendation Context Alignment — Approximately 37%

We deepened our analysis one further step, measuring the alignment rate of "why each engine recommends a given brand" — the recommendation context (reason tags). From 135 response texts, we extracted recommendation context tags (e.g., "cost-effective," "beginner-friendly," "proven track record," "highly reviewed") and calculated pairwise Jaccard coefficients.

Recommendation context alignment averages 37% — substantially higher than domain (10%) or brand (14%) overlap.

Each engine exhibits a distinct "selection axis": ChatGPT favors authority, Gemini favors taxonomic categorization, Perplexity favors user reviews, and AIO closely mirrors Google Search results. Nevertheless, approximately 40% of recommendation contexts are shared.

Table 4 — Pairwise Recommendation Context Alignment (Mean Jaccard Coefficient)
Engine PairContext Tags
Gemini × AIO42.5%
Gemini × Perplexity39.1%
ChatGPT × Gemini37.9%
Perplexity × AIO37.4%
ChatGPT × AIO32.2%
ChatGPT × Perplexity31.5%

ChatGPT's isolation trend persists at the context level (31–38%), but compared to the domain level (6–7%), the gap narrows considerably. This indicates that despite drawing from entirely different information sources, ChatGPT's rationale for recommendations shares a meaningful degree of commonality with other engines.

Industry-Specific Variation in Context Alignment

The degree of context alignment varies substantially depending on industry characteristics.

Table 5 — Context Alignment by Industry Category
CategoryExamples4-Engine Shared ContextCharacteristics
High alignmentOnline brokerages, recruitment agencies44–50%Established top 3 players; clear evaluation criteria
Medium alignmentCosmetics, e-commerce tools15–40%Some shared criteria, but also many unique perspectives
Low alignmentNail schools, social media influencers0–9%Strong regional factors; subjective queries with no definitive answer

05The Perception 3-Layer Structure

The analyses above reveal that cross-engine "commonality" exists at three distinct layers. We term this the "Perception 3-Layer Structure."

Figure 3 — The Perception 3-Layer Structure
37%
Context Layer — Why it recommends
Formed from web-wide perception (cognitive structure). Highest commonality across engines
14%
Brand Layer — What it recommends
Selected through context, but outcomes differ due to each engine's citation behavior
10%
Domain Layer — What it cites
Each engine selects from its own index. Lowest commonality across engines

What the 3-Layer Structure Means

This structure carries important practical implications.

In summary, the strategies that are effective across all engines operate at the Context and Brand layers — that is, building a state in which the brand is discussed across diverse sources in varied contexts (perception reinforcement). Placement on any single domain is effective only for the engines that cite it.

06Comparison with International Research

We compare our findings with similar studies published internationally.

Table 6 — Cross-Engine Overlap Rates: International Comparison
SourcePeriodLanguageCross-Engine OverlapMetric
ConvertMate2024English11%Domain
Profound2025–2026English40–60% monthly variationMonthly citation source volatility
SparkToro / Gumshoe.aiNov–Dec 2025EnglishList exact match <1%Brand lists
This study (LIFE)Feb–Mar 2026Japanese~10% (domain) / ~14% (brand)Domain + Brand + Context

ConvertMate's English-language data (11% domain overlap) and our Japanese-language data (approximately 10%) are closely aligned. The finding that cross-engine domain citation overlap hovers around 10% appears to be a universal tendency, independent of language.

SparkToro's finding of "less than 1% exact match in brand lists" is also consistent with our results. Their study further noted that "top-tier brands appear consistently," which captures the same structural phenomenon as our finding that "brand overlap is higher than domain overlap" — observed from a different analytical perspective.

The original contribution of this study lies in measuring Brand Overlap (14%) and Context Alignment (37%).

Existing international research has been confined to domain-level or brand-list-level analysis. To our knowledge, this is the first study to extend the analysis to "why engines recommend" (the Context Layer).

07Implications for Businesses

Cross-Engine Strategy: Perception Reinforcement

The most significant practical takeaway from the 3-Layer Structure is that "perception reinforcement" is the strategy that works across all engines.

Specifically, the following activities operate at the Context and Brand layers:

Engine-Specific Strategy: Addressing the Domain Layer

While perception reinforcement is essential as a baseline, it may be insufficient in certain cases. ChatGPT, in particular, exhibits extremely low domain overlap with the other three engines (6–7%), necessitating targeted responses to ChatGPT's distinctive citation tendencies (preference for authoritative sources, English-language data, etc.).

For a detailed analysis of engine-specific citation tendencies, see Why ChatGPT and Gemini Recommend Different Brands.

Prioritization by Industry Characteristics

08Detailed Methodology

Data Collection

Japanese-language queries were submitted to each AI search engine, and both response texts and cited URLs were collected. Multiple measurements were conducted per query to construct a high-confidence dataset.

Overlap Measurement

Table 7 — Summary of Measured Variables
VariableData SourceExtraction MethodSample Size
Domain citationsDomains extracted from cited URLsURL parsing9,891
Brand namesResponse textsText analysis998
Recommendation context tagsResponse textsText analysis135
Regarding reproducibility: Brand name and context tag extraction involved text analysis, and therefore perfect reproducibility is not guaranteed. However, all extraction results have been recorded in full, and identical extraction outputs were applied consistently for each input. Domain citation overlap rates were calculated through purely mechanical processing and are highly reproducible.
Data usage: The data presented in this article are based on original research conducted by LIFE, Inc. Citation and reference of this data are welcome, provided that "LIFE, Inc." is credited as the source. For detailed research data or industry-specific investigations, please contact us via the form below.