Understand every metric in Parse. Learn how we measure AI visibility, what each score means, and what good performance looks like.
Score and rate metrics are graded on a five-tier scale. This scale applies to all 0-100 scores and percentage rates across Parse.
Metrics that measure a brand's visibility and performance in AI responses.
Overall AI visibility
Weighted composite of strength (40%), reach (30%), and authority (30%).
A single 0-100 score that combines three signals with explicit weights: strength at 40% measures competitive performance in niches and pairwise comparisons, reach at 30% uses a log-normalized mention score to reduce long-tail compression, and authority at 30% measures citation volume and citation breadth.
Type: Score (0-100)How well you compete
Competitive strength from niche performance and pairwise wins.
Measures how strongly a brand competes in AI recommendations by combining niche-level percentile performance with smoothed pairwise win rate against co-mentioned competitors.
Type: Score (0-100)How broadly AI mentions you
Log-normalized score of how often AI mentions the brand.
Measures how frequently the brand appears in AI responses using a log-normalized mention score so large outliers do not flatten the rest of the distribution. Higher scores mean broader AI mention volume.
Type: Score (0-100)How often the brand appears in ChatGPT and Google AI Overview responses.
Combined score showing the brand's presence across major AI platforms. Calculated from visibility rates in both ChatGPT and Google AI Overview.
Type: Score (0-100)Competitive standing versus tracked peers in your category.
Shows how the brand's visibility compares to direct competitors and similar brands in the same industry. This highlights competitive position in AI responses.
Type: Score (0-100)Percentage of prompts where this brand is mentioned among peer brands.
The percentage of relevant queries where the brand appears compared to peer brands. A higher mention rate indicates stronger presence in the category.
Type: Rate (percentage)Average ranking position when the brand appears in AI responses.
When the brand is mentioned, this shows where it typically appears in the response (1st, 2nd, 3rd, etc.). Lower numbers mean more prominent placement.
Type: Rank (lower is better)Ranking position among all brands in the index.
Where the brand sits compared to every other brand that we track. A lower rank means it appears more frequently across the full index.
Type: Rank (lower is better)Ranking position among brands in the same industry.
Competitive standing inside the brand's industry cohort. Use this to see whether it is gaining or losing ground relative to peers that serve the same market.
Type: Rank (lower is better)Number of AI responses that include this brand during the selected period.
Raw count of how often the brand shows up in AI answers. Track this alongside visibility and rank metrics to spot meaningful volume changes.
Type: CountMetrics about how AI models cite and reference sources.
Number of times AI models cite the domain as a source.
Counts how often AI responses attribute information to the domain. More citations signal that models trust and rely on the content.
Type: CountEstimated impact of closing this citation gap.
A categorical rating (HIGH, MED, LOW) estimating how much the brand's authority score would improve by gaining citations on this source, based on the source's authority rank and citation volume. This is not a numeric score — it is a three-level classification.
Total number of pages from this domain referenced by AI models.
Measures how many individual pages from the domain were cited over the selected period. Useful for spotting domains that provide deep coverage on relevant topics.
Type: CountNumber of unique URLs from the domain cited by AI models.
The breadth of pages from the domain that AI models reference. A higher count shows that multiple pieces of the content are cited, not just a single flagship page.
Type: CountUnique prompts that cite or reference this source.
Counts how many distinct prompts include this citation across AI responses. Higher values mean the source appears across a broader set of user intents.
Type: CountMost recent time AI models referenced this source.
Shows how fresh the citation is. Recent timestamps signal that the source is still being referenced frequently.
Type: DateMetrics about brand performance within specific AI prompts.
Composite score based on mention rate and position ranking.
A weighted score that combines how often the brand appears (mention rate) with how prominently it's positioned in responses. Brands mentioned more frequently and in higher positions score higher. Use it to identify prompts where the brand has strong overall presence.
Type: Score (0-100)Number of unique brands mentioned in responses to this prompt.
The diversity of brands that models mention for this prompt. A higher number means answers are pulling from a broader competitive set.
Type: CountHow many times this brand is mentioned in responses to the prompt.
Total mention count for the brand when people ask this prompt. Pair it with visibility to understand both frequency and share of voice.
Type: CountWhether the focus brand appears in an individual response.
Shows if the selected brand is referenced in a specific AI response so you can quickly zero in on answers that include you.
Type: BooleanChange in metric value over the selected time period.
Represents the delta between the current value and the prior comparable period. Positive change shows improvement, negative values indicate decline.
Type: Delta (signed number)