Measurement

AI Share of Voice (SOV) Tracking: A Complete Guide for Marketers

April 23, 2026 · Teehoo Martech · 13 min read

Traditional Share of Voice measured how much of the advertising airtime in your category your brand owned. AI Share of Voice (AI SOV) measures something more important in 2026: how often ChatGPT, Claude, Perplexity, and Gemini mention your brand when a customer asks a category question.

It's the single most important marketing metric nobody tracked in 2024 and everyone will track by 2027. This guide covers the methodology, the math, the common mistakes, and how to tie SOV to actual pipeline.

Why AI SOV Is the Metric That Matters

14.4% → 2.5%
AI search conversion rate vs traditional Google organic (6x improvement)

Three reasons traditional marketing metrics don't capture the AI-search shift:

SOV is the metric designed for the reality that most brand discovery happens inside a conversation with an AI assistant.

The SOV Formula

The math is simple. The discipline is in the sampling.

Basic SOV

SOV = (# prompts mentioning your brand) / (total prompts tested) × 100

Run 50 category-intent prompts across 4 AI engines. Your brand appears in 18 responses. Your basic SOV is 18/50 × 100 = 36.

Weighted SOV (Preferred)

Basic SOV treats all mentions as equal. But an engine with 300M weekly users matters more than one with 3M; a brand that's the first recommendation matters more than one that's the seventh. Weighted SOV corrects for both:

Weighted SOV = Σengine wengine × Σprompt position_weight / total_prompts

A common weighting scheme:

EngineWeekly usersWeight
ChatGPT300M+0.45
Perplexity~50M0.20
Gemini~400M (bundled w/ Google)0.20
Claude~30M0.15

And position weights within a given answer:

Mention positionWeight
First brand mentioned1.0
Second0.7
Third0.5
Fourth+0.3
Passing mention (not recommendation)0.1

This produces a 0-100 score where being the first-mentioned brand in a ChatGPT answer contributes far more than a passing mention in Claude.

Sampling matters more than math. The biggest SOV mistakes are not in the formula — they're in picking the wrong prompts. A 25-prompt test with the right prompts is far more useful than a 200-prompt test with the wrong ones.

How to Pick the Right Prompts

Your prompt set should reflect how real customers ask AI about your category. Not how your SEO team writes title tags. Four prompt categories to include:

1. Category-Intent Prompts (60-70% of your set)

The user knows they want something in your category but not which brand. These are the gold-standard SOV prompts.

Examples for a project management SaaS:

2. Comparison Prompts (15-20%)

The user is deciding between two specific brands. These test how you fare in head-to-head.

3. Problem Prompts (10-15%)

The user describes a pain point, not a product category. Tests whether AI surfaces your brand for adjacent intent.

4. Brand Validation Prompts (5-10%, lower priority)

The user already knows your brand name. Tests how AI describes you.

Brand validation prompts are the least predictive of SOV lift because the brand is already named in the query. Useful for reputation monitoring, not for tracking visibility.

Benchmarking: What's a Good SOV Score?

No universal benchmark — it depends on category. Rough guidelines from what we've seen:

CategoryLeaderChallengerLong tail
B2B SaaS (horizontal)60-8030-50<20
B2B SaaS (vertical)40-6020-40<15
DTC E-commerce40-6015-30<10
Local services20-4010-20<5
Fashion / lifestyle35-5515-30<10

The more concentrated the category, the higher the top SOV scores. Horizontal B2B SaaS (CRM, project management, video conferencing) has a handful of winners — each takes 60-80. Fragmented categories (DTC apparel, local services) have no single dominant brand and the top SOV is lower.

Your absolute SOV matters less than the delta between you and your named competitors — and your trajectory over 3-6 months.

Tracking SOV Over Time: The Discipline

A monthly tracking process that works:

Month 1 — Baseline

  1. Lock in your prompt set (25-50 prompts, distributed across the 4 categories above)
  2. Lock in your competitor set (3-5 named competitors)
  3. Lock in your engines (typically ChatGPT, Claude, Perplexity, Gemini)
  4. Run the scan. Record SOV for you + each competitor.
  5. Record per-engine breakdown — often you're strong on 1-2 engines and weak on 1-2.

Months 2-12 — Same Prompts, Re-Run

The prompt set shouldn't change. If it changes, you're measuring a moving target and can't compare month over month.

Every month, capture:

Quarterly — Prompt Audit

Once a quarter, review whether the prompt set still reflects how customers talk about the category. New use cases, new intent phrases, new comparison sets. Add 5-10 new prompts. Retire the stalest 5-10. This keeps the set current without destroying month-over-month comparability entirely (keep the 70% core stable).

Tying SOV to Pipeline

The #1 question from CFOs: "Does SOV correlate with revenue?"

Our position: yes, but with a lag, and triangulated. Four patterns we've observed:

  1. Direct/branded search lift. ChatGPT users often double-check an AI recommendation by Googling the brand. A 10-point SOV lift typically produces a 15-25% increase in branded-search impressions within 30-60 days.
  2. "Direct" traffic in GA4. Because AI referrals often come through without a clean utm_source, they show up as Direct. Sustained SOV improvement → sustained direct-traffic lift to the matching pages.
  3. Self-reported attribution. Add "How did you hear about us?" to every conversion form. An explicit "ChatGPT / AI" option catches what no technical attribution can.
  4. Sales call anecdata. Track how many sales calls start with "I saw you recommended by ChatGPT when I asked about..." In early-stage GEO, this is qualitative but very telling.

SOV is a leading indicator, not a revenue-accurate one. Treat it like brand awareness from the pre-digital era: you know it matters even if you can't attribute every sale to a specific mention.

The 5 Most Common SOV Tracking Mistakes

  1. Running the same prompt twice in one day and assuming the result is stable. AI responses vary across sessions; take the aggregate of several runs, not a single snapshot.
  2. Only tracking ChatGPT. It's the biggest single engine, but Claude, Perplexity, and Gemini behave very differently. Missing 3 of 4 engines means missing 55-75% of AI traffic.
  3. Changing the prompt set quarterly. Kills comparability. Lock the core, iterate the edges only.
  4. Not tracking competitors. Absolute SOV means less than relative. A +5 jump while competitors jumped +10 is a loss, even though your number went up.
  5. Chasing a single-month regression. One bad month often reflects an engine update, not a real regression. Use 3-month rolling averages for signal, and investigate only if two consecutive months both regress.

Setting Up SOV Tracking in 30 Minutes

Fast path:

  1. Identify your category. Write down 25 prompts an ideal customer might ask.
  2. Name your top 5 competitors.
  3. Run a tool like Teehoo Martech's free Brand Scan across all 25 prompts × 4 engines. ~2 minutes.
  4. Record the SOV + per-engine scores + competitor deltas in a spreadsheet. ~10 minutes.
  5. Set a calendar reminder for the same day next month. Same prompts. Same tool. Re-run.

That's it. The discipline is in doing it every month for 6 months — the first month is a baseline, months 2-6 are where the story emerges.

FAQ

What is AI Share of Voice (SOV)?
AI Share of Voice is the percentage of AI-generated answers in your category that mention your brand, relative to competitors. If you run 100 category-intent prompts across ChatGPT/Claude/Perplexity/Gemini and your brand appears in 42 of them while your top competitor appears in 67, your SOV is 42 and theirs is 67. It's the direct AI-era analogue of Google ranking share.
How do I calculate my AI SOV?
Basic formula: (# of prompts where your brand is mentioned) / (total prompts tested) × 100. For a weighted version, weight by engine (ChatGPT's weekly user count » others) and by mention position (first-mentioned brand counts more than fifth). A 25-50 prompt test across 4 engines is the minimum viable sample size; 100+ prompts gives higher confidence.
How often should I track AI SOV?
Monthly for most brands; weekly if you're running an active GEO optimization campaign; daily only if you're a high-velocity launch (new product, PR moment, major pivot). The AI models themselves update every few weeks — weekly cadence catches engine-specific regressions before they compound into a bad monthly number.
What's a good AI SOV score?
Benchmarks vary by category. In software/SaaS, category leaders typically hit 60-80 SOV; challengers 30-50; everyone else <20. In e-commerce, top DTC brands hit 40-60 SOV in their niche. What matters more than absolute number: your delta vs named competitors, and your direction of travel over 3-6 months.
Can I track AI SOV without a tool?
Yes, manually — but it's time-consuming. You'd run 25+ prompts on 4 engines, log each response, and count mentions in a spreadsheet. Roughly 2-3 hours per monthly cycle. A tool like Teehoo Martech's free Brand Scan automates this end-to-end in under 2 minutes and gives you an SOV score plus engine breakdown.

Get your AI SOV baseline in 2 minutes

Free scan across ChatGPT, Claude, Perplexity, and Gemini. See your SOV score and competitor ranking.

Run a Free Scan or See agency features

Related reading
· How to Rank in ChatGPT: 12 Ranking Factors
· GEO for Agencies: The Complete Playbook
· AI Brand Monitoring for Agencies
· The AEO Audit Checklist: 14 Points

About Teehoo Martech
AI visibility platform for brands and agencies. Track and improve how AI recommends you. Learn more.