seojuice

Where LLMs Actually Find Your Brand When People Stop Searching Google

Lida Stepul
Lida Stepul
May 06, 2025 · 12 min read

TL;DR: When a reader asks ChatGPT, Perplexity, or Google's AI Mode "what should I use for X", the answer doesn't come from Google's top-10. It comes from a different surface set — Reddit, YouTube, podcasts, Hacker News, Stack Overflow, topical forums — that the model saw at training time and re-cites at inference time. If your brand isn't present on those surfaces, the model can't surface you, and ranking number one on Google won't change that. The fastest way to see the gap is a one-evening audit: pick three queries you'd want to surface for, ask three LLM panels the same prompt, capture which competitors get cited and which sources the model leans on. Most operators are surprised by which surface dominates for their category; few are surprised by the gap. The investment priority isn't "do everything everywhere." It's the one surface where the audit shows the steepest gap between competitor presence and yours.

The question I keep getting

Two operators ask me the same thing in the same week. The first runs a B2B SaaS, ranks number one on Google for their primary keyword, and pulls steady organic traffic. The second runs a consumer brand in a crowded category, ranks third or fourth on Google, and is losing share to a competitor whose Google rankings are objectively worse. Both ask me a variation of the same question: "I'm doing fine on Google, but ChatGPT keeps recommending the competitor when somebody asks. What gives?"

I have spent the past six months poking at this for clients and for vadimkravcenko.com, and the short answer is that Google rankings and LLM citations are reading from different graphs. Google's graph is links plus signals around the URL itself. The LLM graph is a citation graph built from a different set of sources: places where humans talk about brands without the brand owner controlling the conversation. Reddit threads, YouTube videos, podcast episodes, Hacker News submissions, Stack Overflow answers. The model trained on those at training time, and at inference time it still leans on them. If you have no presence in the citation graph, your Google ranking is mostly invisible to the LLM, no matter how high it is.

This piece is the map and the audit. The map names the five off-site surfaces that carry signal today, with a rough sense of which one carries which weight by query type. The audit is the one-evening method I use to figure out where the gap actually sits for a specific brand. The investment priority section closes the loop: if the audit returns X, invest in Y first, not all five at once.

The citation graph isn't Google's link graph

Rand Fishkin made the point most directly in a SparkToro piece on appearing in AI answers, calling out the two surfaces that "seem to still hold particular sway with AI answers":

Reddit and YouTube (which seem to still hold particular sway with AI answers).

Fishkin's read matches what I keep finding on client audits, with two surfaces to add: podcasts and developer forums. The reason is mechanical. The major LLMs train on Common Crawl plus licensing deals (OpenAI's Reddit deal, Google's Reddit deal, both signed in 2024) plus content the model providers scrape directly (YouTube transcripts, podcast transcripts via Apple / Spotify / Listennotes, Stack Overflow's content). When the model is asked a question at inference time, two systems work together: the parametric memory (what the model learned in training) and the retrieval layer (a live search index over a similar set of sources, augmented for freshness). Both lean on the same surface set. Both lean less on the open web than Google does, because the open web is noisier and the citation-grade sources are higher signal per token.

The notable absence in this graph is X. Pre-2023 tweets are in Common Crawl. Post-2023 X content is firewalled: no Common Crawl access, no OpenAI deal, no Google deal as of this writing. X-native brand mentions have low surfacing weight in current LLM citations. Operators putting effort into X for LLM discovery are spending against a closed door. That can change; X's data could open up tomorrow. Today, it hasn't.

What each surface actually carries

Five surfaces carry most of the weight in 2026. The relative weight shifts by query type, which is the thing the map is for.

Map of five off-site surfaces — Reddit, YouTube, podcasts, Hacker News, Stack Overflow — with rough weight bars by query category covering SaaS, B2B, consumer, and local intent
The five off-site surfaces by query category. Bars are rough observation, not survey data.

Reddit carries the heaviest weight for "best X for Y" and "is Z any good" queries across most categories. The OpenAI and Google licensing deals (both 2024) explicitly include Reddit. Subreddit relevance matters: r/SaaS, r/Entrepreneur, r/marketing carry weight for B2B; r/BuyItForLife and product-category subs for consumer. A mention in a small but topically relevant subreddit weighs more than a mention in a high-traffic but irrelevant one.

YouTube transcripts are scraped and indexed. Cited heavily for "how do I X" and "X vs Y" queries. The cited unit is the specific video transcript, not the channel. A single comparison video can drive citations across dozens of related queries for a year after publishing. Channel-level brand presence helps the model build a picture of you over time; video-level keyword coverage triggers specific citations.

Podcast episode transcripts feed both training data and retrieval indexes. Cited disproportionately for B2B expertise queries like "who's an expert on X" or "what's the right way to think about Y". The signal here is recurring third-party validation. A founder who has been a guest on twelve category-relevant podcasts over two years looks completely different to the model than one who hasn't.

Hacker News is a high-signal source for technical and SaaS topics. Both the submitted URL and the comments are indexed. Disproportionate cited-source frequency for developer-tool and infrastructure queries. The signal is conversational density on a URL, not the URL itself. A submission that gets fifty thoughtful comments weighs more than one that gets a hundred upvotes and no discussion.

Stack Overflow remains the technical-question authority. Cited for "how to fix X" code-specific queries. Weight has softened post-2024 as Stack Overflow's own traffic declined under the AI-answer pressure, but for code-specific surfaces it still holds. Topical forums (industry-specific communities, indie message boards, Discourse instances) act as a sixth catch-all. Weights vary by community and there's no clean generalization beyond "domain authority and recurring community presence both matter."

SurfaceWhat it carriesQuery types where it dominates
RedditRecommendation-style discussion + reviews"best X", "is Y any good", "X alternatives"
YouTubeWorkflow walkthroughs + comparisons"how do I X", "X vs Y", tool demos
PodcastsLong-form expertise signal"who's an expert on X", B2B services framing
Hacker NewsTechnical + SaaS conversation densityDeveloper-tool + infra-tool queries
Stack OverflowCode-specific authoritative answers"how to fix X" code questions
Topical forumsDomain-specific recurring presenceNiche-industry queries

How the signal actually moves from mention to citation

The mechanism matters because the lag matters. A brand mention on Reddit posted today doesn't surface in a ChatGPT answer tomorrow. The signal moves through two paths with different cadences.

Pipeline diagram showing how a brand mention on Reddit, YouTube, or a podcast enters LLM training data, gets indexed into retrieval, and surfaces at inference time as a cited source
Two parallel paths from mention to citation: training-data ingestion (months) and retrieval indexing (days to weeks).

Path one is training data. A mention gets scraped or licensed, ends up in a training corpus, gets weighed by training time, and shows up in the model's parametric memory. Lag here is measured in months at minimum, usually six to eighteen months from mention to surfacing. The signal is durable once it's in — a mention from 2023 still influences a 2026 inference, but the path is slow and weighted toward repeated, varied mention rather than single events.

Path two is retrieval. The major LLMs have a live search-index layer (Perplexity makes this most explicit, ChatGPT's browse mode also leans on it, Google AI Mode obviously does). The retrieval layer indexes fresh content faster, on the order of days to weeks, and is queried alongside the parametric memory at inference time. A recent mention can surface within a few weeks via this path, but the retrieval citation tends to be more URL-specific and less durable than the training-data signal.

The practical implication: editorial cadence matters more than any one mention. A brand discussed in four Reddit threads, three YouTube videos, and two podcast episodes over a year looks different to both paths than a brand mentioned once. The audit you'll run in a minute is reading the cumulative picture, not any one moment.

The one-evening audit

The audit is cheap, defensible, and the part of this piece you should actually run before doing anything else. It takes about three to four hours the first time and an hour each quarterly re-run. Five steps.

Five-step audit flowchart: pick three queries, run on three LLM panels, log cited sources, read the gap, pick one surface to invest in first
The audit in one picture. Three queries, three engines, log the citations, pick one surface.

Step one: pick three queries you'd want to surface for. Be specific. "Best project management tool for a five-person remote agency" beats "good PM software" because the specific query is closer to the way a real user types into a real LLM. Pick one head query (what you'd love to win) and two long-tail queries (what you'd realistically expect to win first).

Step two: run each query on ChatGPT, Perplexity, and Google AI Mode. Capture the full answer plus the cited sources. If an engine doesn't cite sources by default, prompt for "what sources did you base this on?" Most will give you them on request.

Step three: log the citations. For each cited source, note which surface it sits on (Reddit thread, YouTube video, podcast episode, Hacker News submission, Stack Overflow answer, brand-owned page, neutral article, competitor-owned page). The pattern that emerges across nine answer-source pairs (three queries × three engines) is usually clear within twenty minutes.

Step four: read the gap. Two readings matter. The first: which surface(s) dominate the citation pool for your specific query type? The second: which competitors get cited, and from which surfaces? If your competitor shows up in seven of nine answers and is cited from Reddit four times, the gap is Reddit. If they show up cited from YouTube videos, the gap is YouTube. The map at the top of this piece will tell you whether the cited-surface pattern matches the typical category pattern.

Step five: pick ONE surface to invest in first. The temptation is to invest in all five surfaces at once. Don't. Pick the surface where the gap between competitor presence and yours is steepest, AND which is consistent with your category's typical pattern. Investment energy spreads thinly across all surfaces returns less than focused energy on the right one.

If you want a more rigorous version of this audit with persistent tracking over time, the AI visibility audit methodology piece walks the longer-form approach. The one-evening pass above is enough for a first read.

Reading the audit and picking the investment

The audit returns a surface gap. The next move is investment, and investment looks different surface by surface. What the AI engines actually reward isn't ranking — it's recurring, varied, third-party mention. A brand that gets discussed naturally across Reddit threads and YouTube comparison videos for two years looks completely different to the model than a brand that's only present on its own marketing pages, regardless of how those marketing pages are optimized.

The "recurring, varied, third-party" framing is the key. Promotional behavior on any of these surfaces gets penalized harder than absence. The investment that works is earned: contributing to subreddit conversations as a real participant, supporting category-relevant YouTube creators with sponsorship or guest appearances, going on podcasts as a guest with category expertise, engaging on Hacker News as someone who has technical opinions rather than promotional ones.

Two-column investment-priority table mapping audit results — competitors cited from Reddit, YouTube, podcasts, Hacker News, or Stack Overflow — to the first off-site surface to invest in
The investment-priority decision table. Map your audit result to the first surface to invest in.
Audit resultFirst investment
Competitors cited from Reddit threads, you're absentEarned Reddit presence in 3-5 category-relevant subs over 6 months
Competitors cited from YouTube videosSponsored mentions on existing category channels OR channel building with monthly cadence
Competitors cited from podcast episodesGuest appearances on 6-10 category podcasts over a year
Competitors cited from Hacker News submissionsOriginal technical writing posted to HN by team members with track records
Competitors cited from Stack Overflow answersEngineering-team time on high-signal threads in your tool's category

Two practical notes from the past six months of watching operators try this. First, the timeline is six months minimum to register, twelve months to see meaningful shift in audit results. Anybody who promises faster is selling you something. Second, the surface investment isn't a tool-purchase decision; it's a team-time decision. A Reddit gap can't be closed by a Reddit tool. It gets closed by a person with category credibility spending two hours a week being a real participant in 3-5 subreddits.

Where this gets messy

The five-surface map is a snapshot of the current state, not a permanent picture. Three things move underneath it.

One, the weights shift as training data shifts. The next model generation might weigh podcasts heavier and Reddit lighter, or might shift toward retrieval over parametric memory in ways that favor one surface over another. The map you see above is mid-2026 reading; mid-2027 might read differently.

Two, surfaces themselves shift. The X firewall is a recent change. Stack Overflow's softening is a recent change. Bluesky, Threads, and Mastodon are not yet at meaningful weight in any audit I've run, but a year from now one of them might be. The map needs a re-read every couple of quarters; the audit method survives the re-reads.

Three, promotional behavior gets caught and penalized at a faster rate than the underlying signal accumulates. A brand that hires a Reddit agency to seed posts in a dozen subs in a month doesn't see surfacing benefit and often sees penalties — Reddit's mod culture, Reddit's own anti-spam, and the LLM training pipelines all catch the pattern. The investment that works is slower and more boring than any agency pitch will tell you.

The honest framing for the reader: this map will read differently in twelve months. The audit method won't. Re-run the audit quarterly, update your read of the map, and adjust the investment priority. Don't lock in a multi-year contract with anyone who promises to deliver "ChatGPT citations" — they can't, the surfaces shift, and the investment that actually works requires team time more than agency time.

What this doesn't replace

Off-site brand presence is a complement to on-site optimization, not a substitute. The schema markup, structured data, clean page hierarchy, internal linking, and content quality that the on-site LLM-optimization pieces cover all still matter. They make the brand mention parseable when it does show up, and they ensure that when a reader clicks through to your site from an LLM citation, the destination page reinforces rather than degrades the signal.

The framing that works for me: off-site presence makes the LLM aware of you; on-site quality determines whether being aware of you converts to citation rather than to dismissal. Both are required. Most operators I hear from have over-invested on-site and under-invested off-site, because the on-site moves feel concrete and the off-site moves feel like marketing-team territory. The AI-first search optimization piece covers the on-site side; the foundational AI-driven SEO piece covers the joint framing.

If you came out with a surface gap

Two next-action reads, depending on where the audit pointed.

If the gap was broad, with competitors cited from three or more surfaces while you're absent across most of them, the multi-source brand pickup piece is the right next read. It walks the broader frame of multi-platform brand presence without the specific surface-by-surface map. If the gap was narrow and concentrated on one surface, the surface-specific deep dives are the next reads: podcast SEO for the podcast-investment path, video SEO for YouTube, brand-in-ChatGPT for the narrower angle, and the GEO mentions-in-AI piece for the conceptual frame underneath all of this.

The piece you don't need is the one promising "ChatGPT optimization in 30 days." The audit is the thing. Run it, read the gap, pick one surface, commit a year of focused team time, and re-audit in three months. The surfaces will keep shifting; the method will keep working.

<script type="application/ld+json"> { "@context": "https://schema.org", "@type": "FAQPage", "mainEntity": [ { "@type": "Question", "name": "Does posting on Reddit and YouTube actually move the needle for LLM citations?", "acceptedAnswer": { "@type": "Answer", "text": "Yes, but the timeline is slower than most operators want. Six months minimum to register, twelve months to see meaningful shift in citation patterns. The signal that works is recurring, varied, third-party mention — not paid placement. Promotional behavior gets caught and penalized faster than the underlying signal accumulates." } }, { "@type": "Question", "name": "Why don't LLMs cite Twitter or X anymore?", "acceptedAnswer": { "@type": "Answer", "text": "Pre-2023 tweets are in Common Crawl and still influence training data. Post-2023 X content is firewalled — no Common Crawl access, no OpenAI deal, no Google deal as of mid-2026. X-native brand mentions have low surfacing weight in current LLM citations. This can change; X's data could open back up. As of this writing, operators investing in X for LLM discovery are spending against a closed door." } }, { "@type": "Question", "name": "How often should I re-run the audit?", "acceptedAnswer": { "@type": "Answer", "text": "Quarterly. The weights shift slowly; the surfaces shift slowly; re-running monthly is overkill and noise-prone, but yearly misses platform-rise events. A quarterly cadence catches the meaningful shifts without burning team time." } }, { "@type": "Question", "name": "Should I pay to be mentioned on podcasts or YouTube channels?", "acceptedAnswer": { "@type": "Answer", "text": "Yes, if the channel is category-relevant and the placement is a substantive sponsor mention in long-form content. Banner-ad placements and pre-roll ads don't carry citation signal. A sixty-second sponsor read in a podcast where the host genuinely engages with what you do carries signal; a generic ad slot does not. Cost-benefit only works for channels with category audience overlap." } }, { "@type": "Question", "name": "What about Bluesky, Mastodon, and Threads — do those count?", "acceptedAnswer": { "@type": "Answer", "text": "Not yet at meaningful weight in any audit I've run through mid-2026. Track them as potential future surfaces but don't invest first. The audit method will catch a meaningful weight shift if one of these surfaces rises; until then, your team time goes further on the five established surfaces." } } ] } </script>