SEO
AI Search Tracking: How to Measure Citations Across ChatGPT/Perplexity/AI Overviews

Why Google Analytics is blind to AI search tracking
AI search tracking is the practice of measuring how often your brand appears in ChatGPT, Perplexity, and Google AI Overviews. Google Analytics captures none of this by default. This post walks through a tested, multi-platform workflow using free and paid tools, so you can measure citations, share of voice, and sentiment without guessing.
Key Takeaways
Google Analytics captures roughly 0% of AI-referred traffic because most AI platforms strip referrer data before passing outbound clicks.
Citation rate, share of voice, and sentiment polarity are the three metrics that replace session volume for AI visibility.
A 20-prompt manual log takes about 45 minutes per week and provides ground-truth data to validate any paid tool output.
AI assistants cite content that is on average 25.7% newer than traditional search results, so refresh cadence directly affects citation retention.
Paid tracking tools sit around an industry average of $337 per month, with mid-market options like Otterly running $200 to $800.
Google Analytics shows essentially zero traffic from AI assistants because most platforms do not pass recognizable referrer strings or UTM parameters. A week where ChatGPT cites your site twenty times looks identical, inside GA4, to a week where it cites you zero. That is the core measurement gap that makes AI search tracking a separate discipline from standard web analytics.
According to AEO Engine's published analytics work, 0% of AI traffic is visible in Google Analytics by default. ChatGPT's browsing mode strips referrer data. Perplexity sends partial referrers inconsistently across device and session types. Google AI Overviews reduce clicks by design, which means impression data inside Search Console is the only click-side signal you reliably get.
The practical consequence: session counts undercount AI-driven interest by a wide margin, often completely. If you are looking at GA4 to evaluate AI visibility, you are looking at the wrong dashboard. For context on how AI Overviews specifically reshape click behavior, see our breakdown of Google AI Overviews and B2B click-through.
What three metrics replace pageviews in AI search?
Citation rate, share of voice, and sentiment polarity are the metrics that replace session volume in AI search measurement. Citation rate tracks how frequently your brand appears across a sampled prompt set. Share of voice compares that rate against named competitors running the same prompts. Sentiment polarity flags whether responses frame your brand positively, neutrally, or negatively.
A high citation rate paired with negative framing actively suppresses conversion, which is why sentiment cannot be treated as cosmetic. Rankability's tracking data on Grammarly showed the brand appearing in 85% of relevant AI answers with 77% positive sentiment, a figure that would be invisible in any session-based report. That combination, volume plus framing, is what you are actually measuring.
The three definitions worth committing to memory:
Citation rate: the percentage of sampled prompts where your brand is named, across a defined prompt library
Share of voice: your citation rate divided by total mentions across all tracked competitors on the same prompt set
Sentiment polarity: positive, neutral, or negative framing flagged per mention
Coverage gap: prompts where neither you nor any competitor is cited, which often signal content opportunities nobody owns yet
How to start tracking AI citations for free
A functional AI citation baseline costs nothing to build. The method is a fixed prompt set, run weekly across ChatGPT, Perplexity, and a Google incognito session, with results logged in a shared spreadsheet. It is slow and non-statistical. But it gives you real ground-truth data to validate paid tool outputs against, and it often reveals which prompts actually matter before you spend money automating the wrong ones.
Ahrefs documents a similar approach in their guide to tracking AI Overviews, with AI Share of Voice metrics now appearing in their Brand Radar dashboard. The free version of the workflow is the same logic, executed manually, with Google Search Console filling the gap on AI Overview impressions and CTR.
Steps for a usable manual setup:
Define 15 to 25 category-level prompts your buyers actually use, based on top-of-funnel search terms, not branded queries
Run each prompt weekly in ChatGPT (web interface), Perplexity, and Google, noting whether your brand, a competitor, or neither was cited
Log the cited source URL when one appears, not just the brand mention itself
Pull Google Search Console's AI Overviews filter for impression and CTR data on pages already showing in Overviews
Hold a hard rule: one manual run per week is one data point, not a rate, so wait at least four weeks before drawing conclusions
For deeper context on what actually drives those citations, our analysis of six factors that drive Perplexity citations is the companion piece.
Which paid AI tracking tools actually deliver?
Paid AI tracking tools differ on which platforms they cover, how often they run prompts, and whether they include sentiment analysis. Most B2B teams either overpay for enterprise depth they do not need, or buy a single-platform tool and miss cross-platform trends. The table below maps the main options against the factors that matter for a 50 to 300 employee SaaS or B2B team.
According to The Smarketers' AEO measurement breakdown, Otterly covers the big four engines at roughly $200 to $800 per month, positioning it as a mid-market option rather than an enterprise platform. Rankability's pricing analysis pegs the industry average for AI search tracking tools at around $337 per month, which is the benchmark to evaluate everything else against.
Tool | AI platforms covered | Prompt frequency | Price range | Best for |
|---|---|---|---|---|
Rankability | ChatGPT, Perplexity, Gemini, AI Overviews | Daily | Below $337/mo avg | Mid-market B2B |
Peec AI | ChatGPT, Perplexity, Gemini | Daily | Varies | SaaS plus agencies |
Otterly | ChatGPT, Perplexity, Gemini, Copilot | Configurable | $200 to $800/mo | Multi-brand teams |
Profound | All major LLMs | High-frequency | Enterprise | Orgs needing deep model data |
Mangools AI Search Watcher | ChatGPT, Perplexity | On-demand | Lower cost | Solo operators, small teams |
The filtering question before any purchase: do you need sentiment scoring, or only citation rate? That single distinction cuts the relevant shortlist in half before you sit through a demo.
What does a three-layer AI tracking workflow look like?
The most reliable AI tracking setup combines a weekly manual prompt log, a paid tool for statistical volume, and Google Search Console for AI Overview click data. Each layer catches something the others miss. Running only one produces a partial picture that can mislead content decisions more than no data at all.
Layer one is the 20-prompt weekly log we covered above. Layer two is a paid tool, typically Otterly or Peec AI, configured to run daily on the same prompt set, which converts your manual spot-checks into a statistical citation rate over weeks and months. Layer three is Search Console's AI Overview impression and CTR data, pulled weekly and joined against layer two to detect cases where you are cited but not clicked. Our 90-day Perplexity and ChatGPT citation tracking dataset shows what that joined dataset looks like in practice.
One operational note that breaks most tracking setups: AI assistants cite content that is on average 25.7% newer than traditional search results. Your tracking should flag pages that have not been updated in six months before they drop out of citation rotation. We add a simple "last updated" column to layer one and review it monthly.
How do you turn citation data into content decisions?
Citation data is only useful when it produces a specific, schedulable content action. A citation rate drop points to a refresh. A share of voice gap identifies a competitor page to pull apart. A sentiment flag points to a claim that needs stronger sourcing. Without a decision rule tied to each metric, the dashboard becomes a reporting artifact rather than a working tool.
In Gravidy audits, the most common pattern is teams that have implemented a tracking tool but never tied its outputs to a content cadence. The data sits in a Slack channel. Nothing changes on the site. The four decision rules we ship with every audit:
Citation rate falls week over week: check the publish date of the previously cited page, compare against the competitor page that replaced it, refresh with newer sources
Share of voice gap on a specific prompt: identify the competitor post being cited, audit what it contains that yours does not (structural depth, author credentials, cited data, schema)
Sentiment negative or neutral: find the exact sentence the AI is drawing from, trace it to the page, rewrite or reinforce with sourced claims
Citation trend versus pipeline: match high-citation months against inbound lead volume to begin a citation-to-revenue attribution model
For a wider view of how this fits into modern search work, our AI in SEO 2026 overview covers the strategic frame.
Frequently Asked Questions
How do I track traffic from AI search?
Standard analytics cannot capture AI-referred sessions reliably. The practical method is to track citation rate across a weekly prompt log and a paid tool such as Peec AI or Otterly, then add Google Search Console's AI Overview impressions filter for the click-side signal. Session volume in GA4 is not a meaningful proxy for AI visibility.
What tools measure AI citations?
Free starting point: a manual prompt log plus Google Search Console. Paid options with multi-platform coverage include Otterly ($200 to $800 per month), Peec AI, Rankability, and Profound for enterprise-scale needs. Most B2B teams at the 50 to 300 employee stage start with Peec AI or Otterly, validate the data against their manual log for two to four weeks, then expand the prompt set.
How do I track brand mentions in AI?
Define 15 to 25 prompts that reflect real buyer questions in your category, not branded searches. Run them weekly across ChatGPT, Perplexity, and Google. Log whether your brand is mentioned, what source is cited, and the framing. A paid tool automates this at volume. The manual version provides the calibration data that keeps the tool honest.
Conclusion
GA4 is blind to AI traffic. Citation rate, share of voice, and sentiment are the metrics that fill the gap. Start with a 20-prompt manual log, add a paid tool when volume demands it, and layer in Search Console for AI Overview click data. The workflow takes about a week to set up and produces useful data within a month.
Most B2B sites we audit have three to five citation gaps sitting in plain sight, usually pages that ranked once and quietly stopped being cited. If you want to know which fixes are draining your traffic, book a Free SEO Audit Call. Thirty minutes, specific findings, no slide decks.


