Overview
The Analysis (Full Report) goes much deeper and takes up to 90 seconds. It consists of five steps executed sequentially.Three Analysis Strands
The analysis is based on three parallel or sequential strands:1. Technical Audit (Context/Clarity) – domains only
Goal: How well is your website readable and understandable for AI models?- Website crawl – Your website is crawled, important pages analyzed
- Structured data – Schema.org/JSON-LD, meta tags, heading hierarchy, Open Graph
- llms.txt – Is it present? Does it offer AI models a structured summary?
- robots.txt – Which AI bots (GPTBot, ClaudeBot, PerplexityBot, etc.) are allowed to crawl?
- Image alt texts, canonical, meta robots, HTTPS, response time
2. Visibility (Unseeded)
Goal: Are you recommended in general queries too – without your brand being explicitly mentioned?- Search queries are generated based on industry and offering (navigational, commercial, comparative, optional custom)
- Unseeded = The AI receives no hint about your brand. This tests whether you are organically recommended – as in a real customer query
- Search models: Perplexity Sonar and GPT-4o-mini Search
- Per intent it is checked: Is your brand mentioned? At which position?
- Position scoring: Position 0 = highest score, Position 3+ = lower score
3. Recognition & Risk
Goal: Do models know you? Are there mix-ups or misinformation?- Knowledge mode: Several LLMs are directly asked – e.g. “Do you know [brand]?” – without web search
- Search mode: Same models with web search to include current sources
- Per model: Confidence (how certain?), external sources, ambiguities (mix-ups)
- Aggregation: Recognition rate, average confidence, external authority
Scoring (Full Report)
After the three strands, scores are calculated:| Dimension | Source |
|---|---|
| Recognition | Recognition model results, known_ratio, external_authority_score |
| Presence | Unseeded intent results; for persons: 50% Unseeded + 50% recognition rate |
| Clarity | Technical audit (domain only); person scans: 0 |
| Risk | Recognition aggregation (ambiguities, uncertainties) |
- Recognition: 25%
- Presence: 35%
- Clarity: 30%
- Risk: 10%
- Recognition: 60%
- Presence: 15%
- Clarity: 0% (not applicable)
- Risk: 25%
Action Recommendations
Action recommendations are generated using Flagship models. They draw on the three analysis strands (technical audit, visibility, recognition) and apply best practices to produce prioritized measures:- Each recommendation refers to a finding source (technical audit, visibility analysis, recognition test, score engine)
- E-E-A-T assignment: Experience, Expertise, Authoritativeness, Trustworthiness
- Priority: High, Medium, Low
- Effort: Quick Win, Medium, Strategic
- Expected effect is described
Temporal Sequence
- Context-Clarity (technical audit) – for domains
- Unseeded (visibility) – can run parallel with Recognition
- Recognition/Risk
- Scoring – calculation of final scores
- Actions – generation of action recommendations
Entity Types
| Type | Description | Clarity | Note |
|---|---|---|---|
| Domain | Website (e.g. makemerank.ai) | Yes | Technical audit is performed |
| Person | Name + topic (e.g. “John Smith, B2B SEO”) | No | Specify topic for more precise Presence scoring |
