What PR's Measurement Playbook Teaches Us About LLM Visibility
In June 2010, more than 200 PR professionals gathered in Barcelona for a measurement summit organized by AMEC, the international association for communication evaluation. The industry had spent a decade defending its budgets with a metric called Advertising Value Equivalency: take the column inches of press coverage, multiply by what the same space would have cost as a paid advertisement, and present the resulting number as the “value” of PR. Everyone in the room knew AVE was indefensible. Clients had started saying so publicly. The summit produced the Barcelona Principles: seven measurement standards that replaced AVE with outcome-based metrics like brand awareness lift, share-of-search correlation, and audience engagement. The PR industry did not solve attribution that week. What it did was establish a measurement framework rigorous enough to survive a budget conversation without pretending to have a direct line to revenue. That framework has sustained a $19 billion global industry for 15 years. LLM visibility measurement needs the same thing.
TL;DR
PR is a $19 billion industry that has never proven direct revenue attribution — yet it sustains $5K-$50K/month retainers because it built a layered measurement framework: brand lift studies, share-of-search correlation, and media mix modeling. LLM visibility is where PR was in 2005: the clipping-services phase, where monitoring dashboards show that mentions happened without proving they changed anything. The branded search bridge exists (0.664 correlation across 75K brands), but it is being asked to carry the entire attribution burden alone. The AI brand insights market needs its Barcelona Principles — quasi-experimental SOV measurement, branded search correlation with named confounders, and multi-signal evidence cases — before CFO patience runs out.

The $19 billion industry that never proved direct ROI
Public relations is a $19 billion global industry as of 2024, according to the International Communications Consultancy Organisation. No PR firm in that $19 billion has ever drawn a direct, attributable line from a press placement to a closed deal. Not once, in the entire history of the profession. The closest anyone has come is correlational: media coverage volume correlates with brand awareness lifts, which correlate with purchase intent, which correlates with revenue. Every link in that chain has confounders. Every PR professional knows this.
Yet PR sustains $5,000 to $50,000 monthly retainers at scale, across thousands of agencies, for thousands of clients, year after year. The channel has survived every recession, every budget review cycle, and every CFO who asked “what are we getting for this?” It survived because the industry built a measurement framework that was honest about what it could prove, rigorous about the signals it tracked, and structured in a way that let each signal corroborate the others.
AI brand insights and LLM visibility measurement are facing the same challenge today, under the same budget pressure, with less measurement infrastructure than PR had in 2005. The parallel is not metaphorical; it is structural.
Four measurement layers that made PR budget-defensible
Before Barcelona, PR measurement was AVE: counting press clippings and multiplying by ad rates. After Barcelona, the industry built a layered framework where each signal measured something different and compensated for the blind spots in the others. Understanding these layers matters because each one has a direct analog in LLM visibility measurement.
| PR Measurement Layer | What It Proves | LLM Visibility Equivalent |
|---|---|---|
| Media clipping services | Where coverage appeared; volume and reach | SOV monitoring dashboards: where and how often AI platforms mention a brand |
| Brand lift studies | Whether perception changed among target audiences | Quasi-experimental SOV measurement: whether content changes moved AI brand perception |
| Share-of-search correlation | Whether awareness translated to demand (branded query volume) | Branded search trend analysis correlated against LLM visibility changes |
| Media mix modeling | PR's contribution to outcomes alongside paid, organic, and direct channels | Multi-signal attribution: AI referral segments + branded search + SOV as overlapping evidence |
The power of this framework is that no single layer carries the attribution burden alone. A brand lift study showing improved perception is more convincing when branded search volume moved in the same period. A share-of-search increase is more defensible when clipping data shows concurrent media coverage growth. Each signal is incomplete; together, they constitute a defensible case.
PR agencies present this as a narrative with evidence, not a dashboard with a single number. The Barcelona Principles explicitly state that AVE is invalid; single-number summaries are rejected in favor of multi-dimensional measurement. The framework works because it names what each signal can and cannot prove, rather than pretending any one metric captures the full picture.
LLM visibility is where PR was in 2005
The current state of AI brand insights measurement maps precisely onto pre-Barcelona PR measurement. The GEO monitoring market reached $848M in 2025, with 27 platforms offering SOV tracking across ChatGPT, Gemini, Perplexity, and Google AI Overviews. This is the clipping services layer: counting where your brand appears, how often, and in what context. It is necessary, valuable, and insufficient for a budget conversation on its own.
When PR was in its clipping-services phase, practitioners presented “media impressions” as the justification for their budgets. Media impressions told you that coverage happened; they did not tell you whether it changed anything. The identical pattern is playing out in LLM visibility: platforms show that AI mentions happened, that SOV moved up or down, that competitors rank differently across platforms. The consistent complaint in reviews of every major GEO monitoring tool is the same sentence: “tells you what is happening but offers very little guidance on what to do about it.”
That complaint is structural, not a product failure. It is what happens when you have a monitoring layer without a measurement framework above it. AI brand insights dashboards are showing the equivalent of media impressions, and practitioners are asking the equivalent of “so what?” because the brand lift and share-of-search layers have not been built yet.
The branded search bridge: PR's share-of-search, applied to LLM visibility
Share-of-search is the most significant measurement innovation in PR's post-Barcelona era. The concept, developed by Les Binet at the IPA, is straightforward: the share of branded search queries for a category is a strong leading indicator of market share. If your brand captures 30% of the branded searches in your category, you are likely to hold approximately 30% market share within 12 months. Binet's research across multiple categories showed the correlation holds consistently enough to function as a planning and evaluation metric.
PR agencies adopted share-of-search as a bridge metric: when media coverage increases, branded search typically increases with it, and that branded search increase is a measurable, verifiable signal that the coverage generated real-world awareness. The mechanism is not direct attribution; it is correlated evidence with a published, peer-reviewed basis.
The same bridge exists for LLM visibility. The Ahrefs study of 75,000 brands found a 0.664 correlation between branded web mentions and AI platform visibility: when AI engines cite a brand more frequently, users search for that brand more in Google. This is the share-of-search analog for AI brand insights. A brand that improves its LLM visibility through content optimization should expect a correlated increase in branded search volume, measurable through Google Search Console.
The 0.664 correlation is meaningful; in social science terms, it explains roughly 44% of the variance. PR's share-of-search correlation with market share is of comparable strength, and it has been sufficient to justify millions in annual PR spend across the industry. The key is that PR never presents the correlation as proof of causation. It presents it as one signal in a multi-layered evidence case, explicitly named as correlational, with the other layers providing corroboration.
The brand lift gap in AI brand insights
PR has the Barcelona Principles, the AMEC Integrated Evaluation Framework, and two decades of published brand lift methodologies. LLM visibility has SOV monitoring and before-and-after charts. The gap between these two states is where the measurement problem lives.
Brand lift studies in PR follow a structured methodology: measure perception before a campaign, execute the campaign, measure perception after, and use control groups or statistical techniques to isolate the campaign's contribution from background noise. The studies do not prove that a specific Forbes article generated $47,000 in revenue. They prove that brand perception moved in the desired direction during the campaign period, among the target audience, by a statistically significant margin, after controlling for confounders. That is enough.
LLM visibility measurement lacks this layer entirely. When a brand optimizes its content for AI engines and its SOV increases, the current measurement infrastructure cannot distinguish the content's contribution from model updates, competitor actions, or the 40-60% monthly citation source volatility that occurs independently of any content changes. That is the equivalent of a PR campaign claiming credit for a brand lift that happened to coincide with a product launch, a competitor scandal, and a seasonal demand spike. No credible PR firm would make that claim; it is standard practice in GEO monitoring.
Building the brand lift equivalent for LLM visibility requires quasi-experimental design: within-brand query controls that hold platform and model version constant, hierarchical Bayesian estimation that separates signal from noise, and confidence intervals that tell you how certain you should be about each SOV movement. This is the measurement methodology PR invested 20 years developing. LLM visibility needs it in 2, because the budget conversations are happening now.
What the budget conversation actually requires
Every PR professional who has defended a budget to a CFO knows the conversation follows a specific structure. The CFO does not ask “can you prove this generated $X in revenue?” They know the channel does not work that way. The question is: “what evidence do you have that this investment is contributing to business outcomes?” The answer that works is layered: we tracked brand awareness before and after the campaign, it increased by Y% in our target segment; branded search volume grew Z% in the same period, consistent with increased awareness; website traffic from brand-related queries increased by W%, and that traffic converted at a higher rate than non-branded.
That answer survives because it presents multiple independent signals, each measuring a different aspect of the same underlying phenomenon, each with named limitations. The CFO understands that no single signal is conclusive. The CFO also understands that three correlated signals moving in the same direction, independently measured, constitute a stronger case than any single metric could.
The AI brand insights budget conversation needs the same structure. The current state of AI search attribution forces practitioners to rely on a single proxy, usually branded search or survey forms, rather than the layered evidence framework that PR has validated over two decades. The proxy is asked to carry the entire attribution burden; it buckles under the weight because no single signal can justify a channel investment alone.
The path forward is clear because it has already been walked. Build the brand lift layer for LLM visibility measurement: quasi-experimental SOV tracking with confidence intervals. Connect it to the share-of-search layer: branded search correlation tracked over time with confounders named explicitly. Add the referral layer: GA4 AI traffic segmentation for the 29.4% that is visible, conversion benchmarking, and engagement quality metrics. Present all three together, each with its limitations stated, and let the overlapping evidence do the work that no single metric ever could.
PR had 20 years to build this. LLM visibility has 2.
The Barcelona Principles came 15 years after the PR industry started taking measurement seriously. The AMEC framework took another 5 years after that to reach widespread adoption. PR could afford that timeline because the channel was established, budgets were locked in, and the measurement pressure was gradual. The measurement framework was built after the market was already mature.
LLM visibility does not have that luxury. Consumer sentiment is at a 10-year low; 94% of enterprises are increasing AI search spend into a market that cannot prove the spend is working. The CMOs on Reddit asking “how are we proving ROI?” are not writing thought leadership pieces; they are preparing for quarterly budget reviews. The measurement framework needs to exist before the market matures, or the budget conversations will kill the investment before it has a chance to prove itself.
This is the compressed timeline problem. PR built measurement infrastructure over two decades because the industry could survive without it. LLM visibility and AI brand insights measurement must be built in two years because the budget pressure is arriving before the measurement infrastructure is ready. The question is whether the AI brand insights market will build its Barcelona Principles before CFOs run out of patience, or whether the channel will contract before its measurement catches up.
What we are building
Sill is building the three-layer measurement framework that maps onto PR's proven playbook. The monitoring layer is live: daily SOV tracking across ChatGPT, Gemini, Google AI Overviews, and Perplexity, covering 86 industries. The brand lift layer is in development: quasi-experimental measurement with within-brand query controls, hierarchical Bayesian estimation, and placebo-calibrated confidence badges that distinguish real LLM visibility movements from background noise. The share-of-search bridge is next: branded search trend analysis correlated against SOV changes, with confounders named explicitly, following the methodology that has justified PR budgets for 15 years.
The output will look like PR's best measurement reports: overlapping evidence from independent signals, each with stated limitations, together constituting a defensible case. That is what survives a budget conversation. The PR industry proved it works. We are building it for LLM visibility.
Start building the evidence case now
PR agencies learned that the measurement foundation must be in place before the budget conversation starts. Sill tracks your LLM visibility daily so when the CFO asks, you have months of baseline data, competitive benchmarks, and trend analysis ready.
References
- AMEC. “Barcelona Principles 3.0.” International Association for the Measurement and Evaluation of Communication, 2020. amecorg.com
- ICCO. “World PR Report 2024.” International Communications Consultancy Organisation, 2024.
- Binet, Les and Peter Field. “Share of Search as a Predictive Measure.” IPA Effectiveness Awards, 2020.
- Ahrefs. “LLM Brand Visibility Study: 75,000 Brands.” Ahrefs Blog, 2025. ahrefs.com
- SparkToro. “Dark Traffic: How Much of Your Analytics Is Missing.” SparkToro, 2025.
- Incremys. “2026 GEO Statistics: Applications, Market and Future Outlook.” Incremys, 2026. incremys.com
- Aggarwal et al. “GEO: Generative Engine Optimization.” KDD 2024. arxiv.org
- AMEC. “Integrated Evaluation Framework.” International Association for the Measurement and Evaluation of Communication, 2016. amecorg.com
- Search Engine Land. “7 Hard Truths About Measuring AI Visibility.” Search Engine Land, March 2026.
Get Your Report
Request your first analysis today to see where you stand.
