1. AI Search Did Not Kill Google Search

Despite dramatic headlines, AI chatbots have not replaced traditional search – they have extended it. Large clickstream studies show that people tend to use AI tools and Google together, not instead of each other, and the overall search “pie” is growing rather than shrinking.

Semrush analyzed 260 billion rows of clickstream data between January 2024 and June 2025 and found that users who start using ChatGPT do not reduce their Google searches; if anything, their Google usage remains stable or ticks slightly upward after adoption.

This pattern holds across multiple cohorts and a control group of users who never used ChatGPT, indicating that ChatGPT sessions are added on top of existing search behavior rather than cannibalizing it.

Datos and SparkToro’s “State of Search 2025” data further shows Google still holds around 95% market share in traditional search even as user behavior fragments toward AI tools, social platforms and zero-click experiences. Together, these findings imply that user journeys are becoming multi-touch: many people now ask a chatbot for options, then search Google to verify, compare and click into websites, so brands must think in terms of blended AI-plus-search visibility rather than an either/or bet.

2. No AI Visibility Tool Can “Put” You Into AI Answers

The promise that a tool can “get you into AI answers” echoes early SEO-era claims about guaranteed first-page rankings and is just as misleading today. AI visibility platforms can surface opportunities, but the work of actually earning brand inclusion in model answers is fundamentally human and strategic.

Modern GEO tools can uncover the prompts where your category appears, show how often your brand is mentioned relative to competitors and highlight on-page content gaps or weak entities that make your site less usable as a source. They can also reveal missing coverage on third-party sites, such as reviews, marketplaces or Wikipedia, where LLMs frequently source contextual information.

What they cannot do is force their way into a model’s training data, secure off-site mentions on authoritative domains or safely auto-edit your site at scale without risking SEO conflicts and brand safety issues.

Real gains come from in-house teams and agencies improving information architecture, crafting entity-rich, expert content, building thought leadership and PR, and deliberately shaping how the brand is discussed across the wider web – tools amplify this work, but they never replace it.

3. Nobody Truly Knows “Prompt Search Volume”

Search professionals love precise numbers, but with LLMs, “prompt demand” is inherently squishy.

Model providers do not expose the kind of transparent, query-level data that SEOs are used to from tools like Search Console, so any numbers you see are modeled estimates, not ground truth.

OpenAI and other assistant providers keep granular usage data private, so third-party GEO tools must infer volume from panel-based telemetry, browser and extension clickstream logs, scraped prompt galleries or opt-in chat exports.

These signals can be cleaned and extrapolated to provide directional insights, such as which use cases are growing, whether certain commercial prompts are trending up and how intent clusters differ between AI assistants and classic search.

However, they cannot reliably tell you “this exact prompt is searched 2,400 times per month” in the way keyword tools try to for Google, and any charts suggesting precise prompt counts should be treated more like a forecast or audience model than an audited metric. The practical takeaway is to use LLM “prompt volume” as a prioritization hint and sanity check, not as a KPI you optimize to the decimal place.

4. AI Visibility Is Not the Same as Rankings

Traditional SEO is grounded in rankings: a deterministic, ordered list of pages where positions can be tracked and compared over time. AI visibility is different because LLM answers are generated on the fly, making your presence probabilistic, context-dependent and highly variable between users and sessions.

In classic search, you can usually rely on relatively stable ranking positions, modest personalization and predictable visibility curves across the top results, which makes monitoring straightforward.

In LLMs, outputs are conditioned on prompt wording, conversation history, user profile and even interface design, so two people asking what appears to be the same question may see different brands, different evidence and different levels of detail.

This generative behavior is also why hallucinations occur: models are optimized to provide plausible, fluent answers even when uncertain, which introduces noise into any attempt to track your brand’s presence.

To cope with this, current GEO measurement tends to follow two paths. Crowd-style averaging aggregates outputs across many users, panels and extensions to estimate overall share of voice, but it can hide strong differences across audience segments.

Persona-based sampling fixes a detailed persona (job role, region, preferences), then runs repeated queries to identify the “stable mode” for that profile – effectively asking, for this exact audience, how often do we appear in the core answer.

Both approaches help, yet both operate under the same constraint: because AI answers are probabilistic and context-sensitive, no tool can offer perfectly deterministic coverage of when and where your brand appears.

5. Off-Site Signals Matter More Than On-Site Tweaks for GEO

Most GEO dashboards make on-page optimization look like the main lever because it is measurable and under direct control.

In reality, the strongest signals for appearing inside AI answers are typically off-site – how widely and in what context your brand is discussed across the broader web.

Recent analyses of AI surfaces show that brand web mentions correlate more strongly with AI Overview presence than classic link-based metrics, suggesting that linkless mentions and brand salience are key ingredients in how models learn which entities to recommend.

The highest correlated factors tend to be brand-centric, raw mentions, branded anchors and branded search volume, rather than marginal technical tweaks.

At the same time, Semrush and other studies of AI citations indicate that domains like Reddit, Wikipedia, YouTube, Google-owned properties, Yelp, Amazon and Linkedin are among the most frequently referenced sources in generative answers, which means the ecosystems you do not control often define your AI narrative.

For GEO, this leads to a clear implication: your brand story on reviews, Q&A threads, social feeds, vertical directories and niche communities often matters more than micro-optimizing title tags or schema on your own site.

On-page work is still important for being a credible, citable source, but the decisive factor in whether your brand name is spoken inside an answer tends to be how richly you exist across third-party environments.

6. The KPI That Actually Matters: Brand Mentions in Answers, Not Just Citations

Being cited as a source in an AI answer is nice, but on its own it usually does little to move revenue, pipeline or sign-ups. For most brands, the real lever is whether the assistant actually names them in the core response, especially in high-intent, commercial scenarios.

Cloudflare’s CEO has noted how skewed the traffic equation has become: OpenAI and Anthropic reportedly crawl orders of magnitude more pages per click than Google, with ratios in the thousands-to-one or worse, which means that LLMs consume huge volumes of content while sending relatively little referral traffic back.

Early studies of Google’s AI Overviews similarly suggest that even top cited links behave more like lower organic positions (around the visibility of a classic “position 6” result) than like the top three blue links, with click-through rates dropping sharply.

Platforms like Reddit have publicly said that, despite heavy LLM citation, AI chatbots are not yet meaningful traffic drivers compared with Google Search and direct visits, underscoring the gap between being mentioned as a source and actually winning visits.

For GEO, this shifts the focus toward three things: whether your brand is named in the primary answer, whether that mention appears in commercially valuable contexts (such as vendor shortlists or “best tools for X” recommendations) and whether you become the default suggestion for specific personas and use cases.

Citations are still useful as a sign of authority and a potential traffic trickle, but brand inclusion in the narrative is the KPI that more closely reflects real business impact.

7. GEO Without SEO Alignment Can Hurt More Than It Helps

Chasing AI visibility in isolation can backfire if it undermines the SEO foundation that still drives the bulk of qualified traffic and conversions. Because GEO changes often affect structure, formatting and copy, careless implementation can erode rankings while delivering only modest gains in AI exposure.

Consider a page that currently ranks well for “how to choose the best accounting software for small businesses” and brings in around 2,000 organic visits each month thanks to strong SEO and external authority.

If a GEO tool recommends structural changes to improve snippet extraction for LLMs – for example, by aggressively templating sections or stripping nuanced context – you might see AI mentions and chatbot referrals climb from zero to a few hundred visits per month.

But if those same changes weaken on-page relevance, dilute user engagement or confuse internal linking, the page could slide down the organic results, cutting Google traffic from 2,000 visits to 200 and leaving total performance flat or worse.

This scenario is already emerging as AI visibility tools optimize primarily for their own dashboards rather than for your overall acquisition mix.

To avoid this, any GEO strategy should treat SEO as the backbone, ensure that proposed changes are validated against organic rankings and conversion paths, and tie AI visibility metrics back to downstream business outcomes like brand search growth, leads and revenue rather than abstract “LLM share of voice.”

Why GEO Needs Its Own Measurement Model

GEO sits on top of the same web, crawlers and many of the same signals that SEO has always used, but the mechanism of visibility has shifted from fixed ranking to dynamic generation.

What shows up in an answer now depends on context, intent and who is asking, which breaks a lot of the assumptions built into classic rank tracking.

Early research indicates that there is meaningful but incomplete overlap between strong Google rankings and LLM visibility: many pages and brands that dominate organic results still fail to appear prominently inside AI answers, revealing a gap that traditional SEO metrics do not explain.

This is why GEO requires its own frameworks – entity-centric, persona-aware and heavily weighted toward off-site signals – instead of simply rebranding SEO rank tracking with new labels.