Skip to main content

Solutions

Service BusinessesProduct BrandsComing SoonEnterprise B2B

Research

AI Recommendation IndexIndustry BenchmarksMethodologyThe Recommendation Gap Explained
RankingsHow It WorksPricingAboutGet Your Score

METHODOLOGY

How the AI Recommendation Index is built

The AI Recommendation Index measures which businesses ChatGPT, Claude, Perplexity, and Gemini recommend when buyers ask for help choosing a service provider. Every ranking comes from real queries submitted to these platforms every month. No estimates, no models, no paid placement. This page explains exactly how.

Version 1.0 | Published March 2026 | Last updated April 2026

What the index measures

The AI Recommendation Index tracks how often a business appears in AI responses to buyer-intent queries. We ask ChatGPT, Claude, Perplexity, and Gemini who they recommend for a given industry in a given city, then we count which businesses get named, how often, and on which platforms.

ARS

AI Recommendation Share

The percentage of tracked queries in which a business was named across the four AI platforms. A 48.3% ARS means the business appeared in responses to 48.3% of the 60 queries run that month. ARS is the primary ranking signal because it directly measures how often AI recommends a firm to buyers.

CIS

Composite Index Score

A weighted score from 0 to 1 combining four signals: recommendation frequency (50%), platform coverage (25%), position in results (15%), and citation context quality (10%). CIS rewards consistency of recommendation across platforms, not just raw volume. See the full breakdown in Section 3.

Why two scores

Frequency alone does not distinguish a business recommended confidently across all four platforms from one that appears heavily on a single platform. A firm ranked first on ChatGPT alone may have a higher ARS than a firm mentioned across all four platforms at lower frequency, but the latter is a more durable AI recommendation signal. CIS catches that nuance.

How we collect the data

Every month, we run 60 distinct buyer-intent prompts across four AI platforms for every tracked city, giving 240 total responses per city per month. Queries use web-search-enabled API access to match consumer behaviour. No scraping of chat interfaces, no browser automation, no fake accounts.

Four AI platforms

Platform Provider Access method Model used
ChatGPT OpenAI API with web search tool enabled GPT-4o with web_search_preview
Claude Anthropic API with web search tool enabled claude-sonnet with web_search
Perplexity Perplexity API, search is default sonar
Gemini Google API with search grounding enabled gemini-2.0-flash with google_search grounding

Each platform uses a different training approach, different retrieval model, and different citation handling. Tracking all four gives a more accurate picture of which businesses are recommended consistently across AI systems, not just on one.

60 queries per city per month

For each industry and city combination, we submit 60 distinct buyer-intent prompts to each of the four platforms. Prompts use varied framings: "who should I hire", "best", "most recommended", and specific contextual variants. Each prompt runs on all four platforms, giving 240 total AI responses per city per month.

Prompts are locked at the start of each city's tracking program and never modified. This ensures month-on-month comparisons are genuine rather than artefacts of prompt changes. Prompts are published on request to journalists and researchers.

Controlling for response variance

AI platforms produce variable responses. The same prompt run twice can return different businesses in different orders. We control for variance three ways:

  • Prompt set locked. Every city runs the same 60 prompts each month. Changes are logged in the update history section below.
  • Platform parameters locked. Model version, temperature, and tool configuration are recorded per run. If OpenAI or Anthropic deprecates a model, we note the switch date.
  • Monthly cadence. Running monthly rather than weekly smooths short-term noise. Real trends emerge over 3 to 6 month windows, matching how AI recommendation infrastructure actually evolves.

How the CIS is calculated

The Composite Index Score is built from four components, each weighted by how strongly it signals durable AI recommendation authority. A business can have a high ARS but a modest CIS if it only appears on one platform or always appears late in lists. CIS catches that nuance.

50%

Recommendation frequency

How often the business appeared across all 60 queries and all four platforms. A business cannot score well overall without being recommended consistently.

25%

Platform coverage

How many of the four platforms independently recommended the business. Appearing on all four is a much stronger signal than appearing heavily on just one. It means multiple AI systems, trained on different data, have converged on the same recommendation.

15%

Position weighting

Where in the response the business appeared. Being named first carries more weight than being named fifth. When a buyer asks for a recommendation, the first name gets the call.

10%

Citation context quality

Whether the AI cited the business with specific, confident context (known for small business tax advisory) or included it in a generic list. Specific citations signal that AI has substantive knowledge of the business, not just its name.

Worked example

HLB Mann Judd, Adelaide

CIS 0.62

Frequency (50%)0.24

Appeared in 29 of 60 queries (48.3% ARS)

Platform coverage (25%)0.31

Recommended across multiple platforms independently

Position weighting (15%)0.05

Average position across responses where it appeared

Citation quality (10%)0.02

Cited with specific context in a portion of appearances

Sum: 0.24 + 0.31 + 0.05 + 0.02 = 0.62

How we parse AI responses

Every AI response is raw text. To turn text into rankings, we parse for business names, resolve them to real entities, and de-duplicate across platforms. The parser is conservative: when a name is ambiguous, we log it and exclude it rather than risk wrong attribution.

How parsing works

After each query runs, the raw response is stored verbatim. A parsing pass extracts candidate business names using pattern matching tuned to each industry (for example, accountant firm names often include "Partners", "Chartered", or "CPA"). Candidate names are then matched against a known-entity table built from Google Places, public business registries, and prior run history. Matched entities carry the rank position and citation context from the response into the intelligence_results table.

What we exclude

Out-of-geography businesses. If a Sydney query returns a Melbourne firm, that result is logged but excluded from the Sydney ranking. Cross-geography responses typically indicate the AI misunderstood the query.

Out-of-industry businesses. If a tax accountants query returns a legal firm, excluded. Industry misclassification is rare but does happen.

Ambiguous names. If a name could resolve to two or more different firms (for example "Smith and Associates" in multiple cities), we log the ambiguity and exclude the result from the ranking pending manual resolution.

Non-business entities. References to government agencies, regulatory bodies, or industry associations are parsed but not ranked. These appear in responses frequently and would distort business rankings.

Accuracy tradeoffs

The conservative approach means true positives are high confidence, but we may under-count firms with common names. Over time we improve entity resolution as we see which responses correctly identify specific firms versus which are ambiguous. Parser changes are logged in the update history.

What the index covers

The Recommendation Index currently covers 10 professional service industries across 7 cities in 1 country. Coverage expands month by month as new cities complete their baseline runs.

Industries tracked

We prioritise industries where AI recommendation has high buyer-decision weight: professional services where buyers typically research 3 to 5 options before choosing, and where AI systems are already being used as a research starting point.

Countries tracked

How new markets are added

New cities and industries are added based on (1) buyer query volume, (2) data availability for parsing, and (3) specific requests from businesses or journalists. A new city takes approximately one month to go live: baseline run, parser tuning, and entity resolution. A new industry takes longer because we need to develop query patterns and entity recognition for that vertical.

See the live index at www.ai-seeyou.com/australia/ or pick a specific industry from the list above.

The research that shapes this index

The methodology draws on peer-reviewed GEO research, published AI platform documentation, and empirical citation analyses. Most GEO vendor claims have no evidence base. We only use techniques backed by published research or direct platform documentation.

Princeton GEO paper (KDD 2024)

Aggarwal et al., "GEO: Generative Engine Optimization", KDD 2024. The landmark study of what actually improves AI citation likelihood. Tested nine optimisation tactics across 10,000 queries. Key finding: citing credible sources produced a 115.1% visibility increase for lower-ranked sites. Adding statistics delivered up to 40%.

arxiv.org/abs/2311.09735 →

Ahrefs AI Overview citation study (2025)

Ahrefs analysed sources cited in Google AI Overviews. Finding: 74% of AI Overview citations come from top-10 organic search results, meaning AI recommendations lean heavily on strong SEO foundations. Reinforces that AI visibility is not separate from search visibility.

ahrefs.com/blog →

OpenAI ChatGPT Search documentation

OpenAI's own documentation on how ChatGPT Search ranks sources. Notable quote: "Ranking in ChatGPT Search is based on a number of factors designed to help users find reliable, relevant information. There is no way to guarantee top placement."

help.openai.com →

Anthropic Claude web search documentation

Anthropic's platform docs on Claude's web search capabilities, including how Claude formats responses with inline citations. Useful reference for understanding what Claude retrieves and how it weighs sources.

docs.anthropic.com →

Why monthly cadence

The Princeton paper confirmed AI recommendations shift as training data, retrieval, and underlying search indexes update. OpenAI updates its index frequently. Anthropic re-indexes Claude periodically. Monthly cadence is the minimum required to detect genuine change versus random response variance.

Using and citing the data

The AI Recommendation Index data is free to cite with attribution for editorial, journalistic, and academic use. Commercial use (reselling the data, embedding it in a competing product, training an AI model) requires a separate agreement. Contact dave@ai-seeyou.com for commercial terms.

Attribution format

Source: AI See You Recommendation Index, {month year}.
https://www.ai-seeyou.com/{industry-slug}/{country-slug}/{city-slug}/

For journalists: immediate use

Every industry-country hub page (for example, /tax-accountants/australia/) includes a "What the data shows" section with copy-paste-ready quotable paragraphs and an attribution line. For city-level data, see the individual city ranking pages. All hub and city pages are free to cite.

For journalists: data requests

For specific analyses, advance access to upcoming runs, or custom cuts of the data (for example, "which Brisbane firms ranked in the top 5 every month of 2026?"), email dave@ai-seeyou.com with the angle and publication. Response within 24 hours on weekdays.

For researchers

Academic researchers can request the full prompt set, parser configuration, and anonymised response corpus for peer review or replication studies. Subject to a data use agreement that prohibits re-identification of specific firms for commercial purposes. Email with institutional affiliation and intended use.

Disclaimer: AI responses exhibit natural variance. Single-run results should be interpreted as indicative, not definitive. Month-on-month trends over 3 or more runs are the statistically sound signal. Rankings do not constitute endorsement, recommendation, or legal advice. Every business listed in the index either appeared in AI responses during the tracking period or did not.

Version history

Methodology changes are logged here so journalists, researchers, and clients can track what changed and when. Material changes (for example, new AI platforms, changed prompt sets, scoring formula updates) trigger a major version bump.

Date Version Change
March 2026 1.0 Initial methodology publication. Four platforms tracked: ChatGPT, Claude, Perplexity, Gemini. 60 prompts per city. CIS formula locked at 50/25/15/10 weights.
Formula lock. The CIS weights (50% frequency, 25% platform coverage, 15% position, 10% citation quality) are locked for the first 12 months to ensure month-on-month comparisons are valid. Any weight changes will be published here with a 30-day notice and re-baselining of historical data.

Questions people ask AI about this

Google ranks web pages. The AI Recommendation Index ranks businesses based on how often AI platforms mention them when asked to recommend service providers. Ahrefs data from 2025 showed AI Overviews pull 74% of citations from top-10 organic search results, so AI rankings correlate with SEO but are not the same. A firm can rank page 2 on Google but still win AI recommendations if its entity signals, review aggregation, and content depth line up. The index measures the AI-layer outcome directly, not the ingredients.

ChatGPT, Claude, Perplexity, and Gemini represent the four largest general-purpose AI platforms buyers currently use for service provider research. Each has distinct training data and retrieval: OpenAI uses Bing plus its own indexing, Anthropic uses Claude web search, Perplexity runs its own index, and Gemini uses Google's infrastructure. Tracking all four gives a cross-platform view rather than over-indexing on any single provider. If new platforms reach meaningful adoption (for example, Meta AI or xAI's Grok), they will be considered for inclusion in future versions.

No. The AI See You Recommendation Index cannot be paid into. Rankings come only from actual AI responses to real prompts. No brand, advertiser, or sponsor affects the data. This is deliberate: paid rankings dominate Google, Bing, and every major directory. The value of this index is precisely that it cannot be bought. Princeton's 2024 GEO research concluded that credibility signals outweigh commercial signals for AI citation outcomes, so the methodology follows the research.

Do AI engines recommend your business?

Get a free Recommendation Score in 30 seconds. See exactly where AI recommends you, where it doesn't, and what to fix.

Get My Free Score