
AI Presence
Methodology & Scoring Framework (v1.0)
Last updated: December 2025
Status: Public methodology
Applies to: AI Readiness Audits and Competitive Visibility Runs
Purpose
AI Presence exists to help organizations understand how modern AI systems interpret, trust, and recommend them.
As search has shifted from ranking pages to generating answers, visibility is no longer determined by keywords alone. It is determined by how clearly, consistently, and credibly an organization can be understood by large language models (LLMs).
This document explains exactly what AI Presence measures, how scores are produced, and what the system does not claim to do.
What We Mean by "AI Visibility"
In this context, AI visibility refers to:
- The likelihood an organization is recognized as a distinct entity
- The likelihood it is retrieved for relevant questions or problems
- The likelihood it is preferred over alternatives when AI systems generate answers
AI visibility is inferred from observable, repeatable signals — not from direct access to AI model internals.
AI Presence does not measure traffic, rankings, or guaranteed outcomes.
Readiness vs Competitive
- Readiness — evaluates an organization in isolation using public signals. No live AI scraping for scoring.
- Competitive — introduces competitors by design; evaluates relative positioning.
- Scores are not rankings or predictions. See FAQ: Are AI Presence scores rankings?
- No guarantees. See FAQ: Can AI Presence guarantee inclusion?
- No live AI scraping for scoring. See FAQ: Does AI Presence scrape live AI responses?
Two Score Types
AI Presence uses two distinct scores, each serving a different purpose.
1. AI Readiness Score
What it is
A standalone assessment of how understandable and trustworthy an organization appears to AI systems in isolation.
What it measures
- Entity clarity
- Content structure and coverage
- Reputation indicators
- Machine-readability signals
What it does not do
- Compare against competitors
- Use live AI prompt testing
- Reflect relative market position
This score is stable unless the underlying signals change.
2. Competitive AI Visibility Score
What it is
A normalized score that reflects relative AI visibility when an organization is evaluated alongside competitors.
How it differs
- Uses the same underlying signals as the AI Readiness Score
- Scores are normalized across the competitive cohort
- Scores may be higher or lower than the standalone readiness score
This difference is expected and intentional.
Signals We Evaluate
AI systems infer trust and relevance from recurring patterns across the web.
AI Presence evaluates the following signal categories.
Entity Clarity
What it represents
How clearly an organization can be identified as a unique, consistent entity.
Examples
- Consistent business naming
- Clear location and service associations
- Dedicated, canonical pages
Why it matters
Entity ambiguity leads to retrieval errors or exclusion.
Reputation Signals
What it represents
Evidence that an organization is trusted by real people.
Examples
- Review volume and recency
- Review platform diversity
- Aggregate sentiment indicators
Why it matters
AI systems strongly weight social proof when selecting recommendations.
Directory & Citation Presence
What it represents
Consistency and corroboration across trusted third-party platforms.
Examples
- NAP consistency
- Verified listings
- Breadth of directory coverage
Why it matters
Directories act as external validation layers for entity trust.
Content Coverage
What it represents
Depth and clarity of explanatory information.
Examples
- Service-specific pages
- FAQs and educational content
- Practitioner or staff information
Why it matters
AI systems extract answers from structured, explanatory content.
Structured Signals
What it represents
Machine-readable annotations that reduce ambiguity.
Examples
- Schema markup
- Clear semantic hierarchy
- Consistent page structure
Why it matters
Structured data reduces inference uncertainty for AI systems.
Mentions & External Indicators
What it represents
External references that reinforce legitimacy.
Examples
- Brand mentions
- Social profiles
- Contextual citations
Why it matters
External mentions provide corroboration beyond owned properties.
How Scoring Works (High Level)
- Scores are composite, not single-factor
- Signal categories are weighted within defined ranges
- No single signal can fully dominate a score
- Competitive scores are normalized across the cohort
- Scores are not rankings
- Scores are not probabilities
- Scores are not guarantees
AI Presence intentionally avoids false precision.
Canonical AI Visibility Readiness Stages
AI Readiness Scores are mapped to five readiness stages that reflect how AI systems interpret and trust your brand. These stages help contextualize scores and set realistic expectations.
1️⃣ Emerging (0-39)
What it means
AI models have limited, fragmented, or inconsistent information.
Brand mentions may exist, but confidence is low.
Facts are often inferred or partially missing.
Plain-English framing
"AI is aware of you, but cannot yet speak confidently or consistently."
Important note
This is where most brands start.
Emerging ≠ failure
Emerging = opportunity
2️⃣ Developing (40-59)
What it means
AI can answer some questions accurately.
Evidence exists, but coverage is incomplete.
Answers may vary between models or prompts.
Plain-English framing
"AI can describe you in parts, but not reliably as a whole."
Key signal
Inconsistency, not invisibility
3️⃣ Established (60-79)
What it means
Core facts are verifiable and repeatable.
AI answers are mostly accurate across models.
Competitive positioning is visible.
Plain-English framing
"AI can describe your brand clearly and correctly most of the time."
Strategic meaning
This is the first level of real AI credibility
4️⃣ Strong (80-89)
What it means
High confidence, high consistency.
Clear differentiation from competitors.
Fewer unknowns or speculative claims.
Plain-English framing
"AI understands who you are and why you matter."
Strategic meaning
You are influencing AI answers, not reacting to them
5️⃣ Rare (90-100)
What it means
Exceptional clarity and authority.
AI defaults to your brand as a reference.
Very few competitors reach this level.
Plain-English framing
"AI treats your brand as a trusted source of truth."
Important
Rare is intentionally hard to reach.
Scarcity here is a feature, not a flaw.
Evidence Sources
Evidence sources shown in reports are illustrative, not exhaustive.
They exist to:
- Demonstrate signal presence
- Increase transparency
- Support interpretability
They do not represent:
- Crawl completeness
- AI training data
- Guaranteed citation sources
What AI Presence Does Not Do
AI Presence does not:
- Scrape live AI responses continuously
- Guarantee inclusion in AI-generated answers
- Influence or control AI models
- Train AI systems
- Measure keyword rankings
- Provide real-time monitoring
AI Presence evaluates readiness and relative likelihood, not outcomes.
Known Limitations
- AI behavior changes over time
- Many AI systems do not cite sources
- Geographic bias varies by platform
- Emerging standards are still evolving
- Visibility does not equal conversion
This system measures clarity and trust signals, not business performance.
How to Use This Tool
AI Presence is intended as:
- A diagnostic framework
- A prioritization guide
- A planning tool
Recommended cadence:
- Standalone audit: as needed
- Competitive comparison: quarterly or after major changes
Obsessive daily monitoring is discouraged.
Versioning
- This methodology is versioned
- Changes are documented
- Historical scores remain interpretable
- Backward compatibility is prioritized
Current version: v1.0
Closing
AI Presence exists to reduce ambiguity.
In an era where AI systems increasingly define what gets repeated, clarity and specificity are no longer optional.
This methodology reflects that reality.