OptimAI OptimAI

The Complete AI SEO Guide with OptimAI

This AI SEO guide walks through how modern search has shifted - and how OptimAI measures it. AI engines now answer questions your customers used to type into Google. OptimAI is the audit tool built for that shift: we check whether AI engines can extract your content, whether they're actually citing your domain right now, and what they're answering for your target keyword - then hand you a prioritised to-do list of what to fix, why it matters, and how it's prioritised by impact.

Where traditional SEO tools hand you a spreadsheet of metrics, OptimAI scores both halves of modern search in a single scan:

  • AI Search Readiness - 40% of your total score, across six dimensions legacy SEO tools don't measure at all: structured data, Q&A patterns, direct answer density, citation worthiness, entity richness, content freshness.
  • AI Search Visibility - direct citation checking on Perplexity, OpenAI (ChatGPT) and Google. We ask each engine your keyword and tell you whether, where and how your domain gets cited.
  • AI Search Opportunity - reverse-engineers what AI engines are actually answering for your keyword and flags the topics, questions and formats your page is missing.
  • Traditional SEO fundamentals - Content, On-Page, Technical and Performance (60% of the total). Everything you'd expect from a serious audit tool, alongside everything the incumbents leave out.

Every suggestion is personalised to the specific page and keyword you scanned. No generic advice, no filler, no "industry best practices" copy-pasted from a template - the scorecard reads your page and tells you what to do about it.

This is the definitive walkthrough of OptimAI, its scoring methodology and why this is the best tool to get your site visible in AI and improve your Google ranking. Every panel, every dimension, every tip worth knowing - plus a real-world example scan of Salesforce's UK product page end-to-end, so you can see exactly how the scorecard reads on a real site.

If you just want the UI tour, skim section 2. If you want to understand the scoring, head to section 3. If you're evaluating Pro tier, section 4 covers everything. For a working rhythm once you're up and running, see section 5.

1. Getting Started

OptimAI is an AI SEO audit tool. Paste a URL and a target keyword, get a 20-dimension scorecard covering how well the page ranks in traditional Google search and in the new generation of AI answer engines (also called Answer Engine Optimization, or AEO). The free trial gives you enough credits for ten full analyses, no card required.

1.1 Logging in

Empty OptimAI AI SEO audit dashboard after signup with URL + keyword inputs, Analyze button showing -12 credit cost, and a green '120 free credits added' badge.

Sign in with Google or Microsoft. A Standard account is created automatically with 120 free credits - enough for ten full analyses. You'll land on the empty dashboard:

  • Scan bar - URL field + keyword field + Analyze button (showing the credit cost for the scan)
  • Free credits badge - your starting balance
  • Empty state - once you run a scan, every scan you ever run will appear in this table
Tip: Start with your homepage and your most important landing page. These are the pages that win or lose customers - fix them first.

1.2 Run your first scan

Scan bar filled in with URL https://www.salesforce.com/uk/products/ and keyword 'crm software', Analyze button showing -12 credit cost.

Paste a URL. Enter a target keyword. Click Analyze. The Analyze button shows the credit cost dynamically so you always know what you're spending before you spend it.

Tip: A keyword is the phrase you'd want this page to show up for - the words a customer would type into Google or ask ChatGPT when looking for what you sell. For a CRM vendor it might be "crm software" or "best crm for small business". Pick one focused phrase per scan - the tool measures how well this specific page answers that specific query. If you're not sure what your page should rank for, look at the headline and ask: what question does this page answer?

1.3 Your first report

Top of a report view: overall score gauge showing 47, scanned URL + keyword, five category score circles, full dimension heatmap.
Top of the Salesforce worked-example report. Download the full PDF.

A scan takes around 45-60 seconds. The report opens automatically when it's done. You get a single long scrollable view broken into sections:

  • The scan bar at the top showing the scanned URL + keyword, with a Refresh button (runs a fresh scan at the reduced refresh cost)
  • A section nav bar: AI Search / Content / On-Page / Technical / Performance / Keywords / SERP / History - click to jump straight to that section
  • The Report / Action Plan tabs top-right - switch between the full report and a todo-list view of every suggestion across every dimension, prioritised by impact
  • The top panel with your overall score, five per-category scores, and a dimension heatmap
  • Each category section in turn - top panel, per-dimension panels, expandable metrics drawers
  • The Download PDF button bottom-right
Tip: Throughout the system you'll see small (i) icons - click for a plain-English explainer of that item. Every dimension, every metric, every weighting has one. The Content category info icon clicked, revealing a tooltip: 'The quality, depth, and readability of your page content. Covers keyword usage, content quality, readability, and heading structure. Weight: 24% of page total.'

1.4 Real example explained: Salesforce

If you just want the UI tour, skip to Getting Around. This section walks through a real report end-to-end and is the clearest single view of how to read OptimAI - worth 5 minutes.

The Salesforce UK products page scoring 47 is a fair real-world example - big brands aren't immune, especially on AI search readiness. Let's walk through a real scan (salesforce.com/uk/products against the keyword "salesforce products", download the full PDF here) and unpack what each number is telling us.

Tip: A low score often means keyword / page mismatch, not a broken page. A brand hub built for navigation (e.g. /products/) won't score well for a commercial keyword like "best crm software" - that intent belongs on a dedicated landing page. Before fixing the page, check you're scoring it against a keyword it was actually built to win.

The headline: 47/100 - dragged down by AI Search

CategoryScoreWeightContribution
AI Search Readiness3140%12.2
Content7224%17.3
On-Page4520%9.0
Technical5410%5.4
Performance506%3.0
Overall47100%47

If we excluded AI Search entirely and just scored the traditional SEO dimensions (Content + On-Page + Technical + Performance, 60% of our weighting), this page would land around ~58 - solidly amber, a "decent page that needs work" rather than "in trouble". The combined score of 47 isn't an error; it's what happens when you honestly measure a page optimised for 20 years of traditional SEO but zero years of AI search.

Tip: This gap - pure-traditional SEO score vs combined score - is the number that should matter to a forward-looking marketer. It quantifies how much of your current ranking strength is invested in a channel that's shrinking (classic organic, with 58% click drop where AI Overviews appear) vs the channel that's growing (AI citations).

AI Search Readiness: 31 - "low readiness"

This is the single biggest lever.

Tip: Think of AI Search Readiness like dressing for a job interview. Nailing the schema, entity naming and Q&A structure doesn't guarantee AI engines will cite you - their answers shift with every query, user and model update. But turning up in pyjamas (no schema, wall-of-text, no entity clarity) guarantees they won't. There's no "page 2" of an AI answer - engines either pick you or they don't.
DimensionScoreWeight
Structured Data Completeness1527%
Question-Answer Pattern Quality4417%
Direct Answer Density4817%
Citation Worthiness3613%
Entity & Data Richness413%
Content Freshness4413%

Two red flags stand out: Structured Data at 15 (schema markup nearly absent - no FAQPage, no Article, no Organization) and Entity & Data Richness at 4 (almost no concrete data points, no pricing, no customer counts). AI systems favour extractable facts over marketing prose, and this page reads like marketing prose.

Salesforce's page doesn't clear either bar.

Content: 72 - the saving grace

If AI Search is the red flag, Content is why the overall isn't in the red. The page has 1,749 words of substantive copy covering multiple product subtopics. Headings are clean, readability sits in the fair-to-good band, the writing is clear. This is the category where Salesforce's size shows up - they've had editors polishing this copy for years.

On-Page (45), Technical (54), Performance (50) - the "basics overlooked" story

Even at 20 years of SEO investment, the basics have gaps. A few worth flagging:

  • Images: 14 - only 25% of the page's 211 images have alt text. 159 images have no alt attribute. That's a 5-minute-per-image fix and a compounding signal.
  • Keyword not in H1 - "salesforce products" doesn't appear exactly in the H1. (Caveat: we guessed this keyword from the SERP title; Salesforce's real target could differ.)
  • Meta description 61 characters - outside the optimal 50-60 window, so Google will rewrite the SERP snippet itself.
  • External links: 47 - excessive, dilutes link equity.

AI Search Visibility: cited by 2 of 3 engines

When we ask the AI engines "salesforce products":

EngineStatus
PerplexityCited - salesforce.com/pricing as citation 2 of 5
OpenAI (ChatGPT)Cited - salesforce.com/eu/products as citation 1 of 2
GoogleNot found - domain discussed but not cited as a formal source

Two of three engines cite the Salesforce domain - not always the exact /uk/products page we scanned, but a page from the domain. That's the domain-authority effect: AI engines know Salesforce is an authoritative source and pull in some page from the domain, even if it's not the best-optimised one.

Why 47 is the right score

Salesforce is the 1,500-pound gorilla of CRM, ranks #2 in Google SERP for "salesforce products", is cited by two of three AI engines, and has 1,749 words of clean copy. And it scores 47. That's not a flaw in the tool; that's the tool doing its job. Other SEO audit tools would give this page an 80+ and a gold star, because they'd be measuring 20-year-old signals. We're measuring what actually drives discovery in 2026: can AI engines extract, understand, and cite your content? For this page, the honest answer is "partially, and only by leaning on domain authority".

The receipts: OptimAI scoring its own landing page

We run the same tool against our own landing page (optimai.live/free-ai-seo-audit, keyword "free ai seo audit") to prove we eat our own dog food:

Top of the OptimAI self-scan report showing an 82 overall score and green gauges across all five categories.
Top of our own landing page report: 82/100, green across all five categories. Download the full PDF.
CategoryOptimAI landingSalesforceDelta
Overall8247+35
AI Search8131+50
Content8172+9
On-Page7745+32
Technical9154+37
Performance9250+42

The difference isn't domain authority or word count - Salesforce wins both on raw inputs. The difference is deliberate structure: FAQPage and Article schema, explicit entity naming, short paragraphs, concrete data points, cited external sources. Every suggestion the tool gives other pages, we've already taken on our own.

P.S. This scan was run pre-launch, before the site was indexed - so you'll see zero AI citations and no ranked keywords on our own report. That's expected for a brand-new domain. The point of the comparison is the structural score (how well the page is built for AI search and on-page fundamentals), not the visibility score - visibility will build naturally as the site gets indexed and linked.

Tip: If you're evaluating whether an SEO tool's advice is worth following, ask the tool to audit its own website. Any vendor that won't publish its own score is telling you something.

2. Getting Around

A quick tour of the UI surfaces. This is shallow on purpose - the meat is in Section 3.

2.1 Dashboard

OptimAI dashboard with 'Your Pages' table showing multiple scans, score badges, change deltas, scan counts, and a filter input.

After your first scan, the dashboard becomes your home view. From any report, click the OptimAI logo (top-left) or the app back button (next to the Page Report heading) to return to the dashboard at any time.

Report view with the OptimAI logo (top-left) and the app back button (next to the 'Page Report' heading) highlighted - the two ways to return to the dashboard from a report.
From a report, return to the dashboard via the logo or the back-arrow button next to the Page Report heading.
  • Your Pages table - every scan you've ever run (we keep lifetime history). Columns: # / Page / Keyword / Score / Change / Scans / Last Scan / Refresh. Click any row to open the report.
  • Filter input (top-right of table) - filters by URL or keyword, client-side, instant.
  • Refresh button - re-scans the same URL+keyword at the reduced refresh cost.
  • Region banner (top, dismissible) - tells you which country we're scoring against. Change from the settings menu.
Tip: Region matters. If you're optimising for a UK audience but your account is set to United States, search volumes and rankings all reflect US Google instead. We auto-detect from your IP, but you can override it any time.

2.2 Settings menu

Settings menu open from the avatar bottom-left, showing name, email, credit balance, Account / Buy Credits / Transactions / Region / Sign Out links.

Open from the avatar bottom-left of any page. Header shows your name, email, current credit balance. Menu items: Account (opens the Account modal), Buy Credits, Transactions, Region, Sign Out.

2.3 Account modal

Account modal showing name, email, Standard tier badge, credit balance, recent activity, and Upgrade to Pro panel with feature checklist.

Your full account profile: identity, tier, credit balance, recent activity, upgrade path to Pro, and a Deactivate Account option (your scan history and credit balance are preserved - contact support to reactivate).

2.4 Buy credits

Buy Credits modal showing five pack tiles (50, 100, 250 POPULAR, 500, 1000 credits) with GBP pricing and VAT note.

Five credit packs available in GBP and USD. VAT exclusive on GBP transactions (UK 20%), rest of world exempt. Credits never expire. Need more than £1,000? There's a bulk pricing contact link.

Tip: No subscription, no lock-in. Buy credits when you need them, use them whenever. One-page test for a pitch next week? Buy 50, scan once, sit on the rest. They'll be there in six months when you need them.

2.5 Transactions

Transactions ledger showing date, description, type badges (ANALYSIS, TRIAL, PURCHASE), amounts coloured red for debits and green for credits, running balance.

Full ledger of every credit movement on your account. Every scan debit, every purchase, every bonus, every refund - dated, categorised, with running balance. Reachable from settings menu or the Account modal's Recent Activity link.

2.6 Region picker

Region picker modal with a country dropdown (UK selected) and caveat text about existing scans being tied to their scan region.

Region matters. If you're optimising for a UK audience but your account is set to United States, search volumes, keyword rankings and competitor data will all reflect US Google rather than the Google your customers actually use. We auto-detect region from your IP, but you can override it any time from the region picker in the settings menu.

Changing the region only affects new scans. Existing scans stay tied to the region they were scanned under - if you want to re-region a historical report, run a fresh scan after switching.

Tip: Switch region before you scan, not after. If you've already scanned a UK page on UK Google and want to see it on US Google, you'll need to run a fresh scan after switching.

2.7 Same page, different keywords

Dashboard showing two scans of the same URL under different keywords (one 43, one 47) - each is a separate row.

Each URL+keyword pair is treated as a separate scan and gets its own dashboard row. Scanning the same page against multiple keywords is a legitimate and useful pattern - different keywords surface different strengths and weaknesses.

Tip: A page can have multiple "right" keywords - and tracking each separately is the point. Your homepage might be optimised for your brand name and your category. Scan it against both, see which performs better, treat the gap as your roadmap.

3. Report Deep Dive

The single-page report is one long scrollable view broken into sections. Every score on the page comes from the same scan - the dashboard, the report, the PDF and the action plan are all reading the same numbers.

3.3 AI Search Visibility

AI Search Visibility panel - status banner with citation timeline sparkline, three engine panels (Google, OpenAI, Perplexity) showing citation snippets.

Below the 6 dimension panels, the AI Search Visibility panel tells you whether AI engines are actually citing your domain right now for the keyword you scanned.

Layout: status banner (green tick or red X), citation timeline sparkline, three engine panels (Google / OpenAI / Perplexity), and a Competitor Analysis table comparing your page vs the average of the top cited pages.

Each engine panel shows a badge: Cited (green), Mentioned (amber - in text but not formally cited), Not Found (red), or Unavailable (engine API errored). Cited engines expand to show the actual citation snippets where available.

Tip: A "Mentioned" badge is closer to a win than it looks. The AI engine knows your domain exists and considers it relevant - it just didn't formally cite it. Usually the fix is structural: a clearer answer to the question in extractable format (lists, schema, direct opening). One pass through Direct Answer Density and Q&A dimensions usually flips Mentioned → Cited within a few weeks.

Caveat on Google: the way we query Google differs from how we query Perplexity and OpenAI - which means for Google we can confirm whether your domain was cited but not always show the exact snippet that referenced you. For Perplexity and OpenAI you get the full context around your cited reference.

AI search results are non-deterministic. Re-running a scan can produce different results because the engines themselves don't always cite the same sources. Patterns matter more than single data points - which is why the timeline bubble chart exists.

Citation History: Best of Day bubble chart with the 27 Mar column highlighted - Google 1, OpenAI 3, Perplexity 0 citations across 3 scans that day.
Citation History bubble chart: read across a row to see how often each engine cited your page over time, and across the columns to compare engines on a given date.

You'll only see the bubble chart timeline after you have run more than one scan. It's possible to run more than one scan in a day, each of which may get entirely different citation results. The bottom row of the bubble chart shows how many scans were run on a given date - in the screenshot above you can see the highlighted column shows 3 scans run on 27th March. We show the highest count of citations from the scans run that day - here you can see one scan got 3 citations from OpenAI and 1 from Google, which could have come from different scans that day.

Citation History bubble chart with tooltip showing OpenAI 3 citations on 27 Mar sourced from the 17:26 scan, best of 3 scans that day.

If you click one of the green bubbles you'll get a tooltip telling you which scan from that day got you the citations - click the timestamp link where it says "Source" to view that report in full, so you can check the exact details.

Competitor Analysis panel - table comparing your page vs the average of the top cited pages across Structured Data and Content Freshness dimensions, with concrete actions.

When possible (this depends on whether AI engines return any citations at all, which isn't guaranteed) we also analyse the other sites that were cited against the same AI Search Readiness metrics as your page. We distill this down into a comparison of which AI Search Readiness dimensions other cited sites are achieving higher scores in than you - giving you a real-time AI Search competitor analysis.

Tip: The Competitor Analysis table is gold dust. It's literally telling you "the pages getting cited have X, you have Y, do this." Treat it as a prioritised punch list - work through dimensions in order, refresh after your changes, watch the bars move.

3.4 AI Search Opportunity

AI Search Opportunity panel - coverage stats, colour-coded suggestion panels by type (Content Gap red, Topic Expansion amber, Format Change teal, New Content dark teal), expandable questions/topics lists.

The companion to AI Search Visibility. While AI Search Visibility tells you whether AI engines are citing your domain, AI Search Opportunity reverse-engineers what AI engines are actually answering for your keyword and flags what's missing from your page.

How it works: when you scan, we query Perplexity, OpenAI and Google for your keyword and analyse their responses. We extract the questions AI engines are answering, the topics they cover, and the domains they cite. Then we compare against your page content and flag the gaps.

Tip: Questions and topics aren't categories the AI engines themselves return - we detect them in the raw response text (think People Also Ask, but extracted from what Perplexity, OpenAI and Google actually wrote about your keyword). Treat them as strong signals, not exact ground truth - occasionally a "topic" is a proper noun the pattern matcher grabbed, or a question has odd phrasing. The value is the coverage gap, not the literal wording.

The panel shows:

  • Coverage stats: "X of Y topics covered", "X of Y questions answered" - your gap at a glance
  • Suggestion panels in four types, each colour-coded:
    • Content Gap (red) - subjects cited sources cover that you don't. Add this content to compete.
    • Topic Expansion (amber) - topics you mention but don't cover deeply.
    • Format Change (teal) - structural changes (FAQ schema, lists, tables) cited sources use.
    • New Content (dark teal) - standalone pages worth creating to capture adjacent queries.
  • Expandable lists of all questions and topics - items already on your page show a "you" badge
  • Cited domains - the actual sites AI engines are pulling from
Tip: Start with the high-impact Content Gap panels - those are the biggest unlocks. Topic Expansion is your second priority (faster wins because content already exists). Format Change is the lowest-cost fix per impact. Save New Content for when the existing page is fully optimised.

Tip - the full AI search picture: AI search is already eating the clicks traditional SEO used to capture, and the tools most agencies still run were built for a pre-AI world.

Sections 3.2 AI Search Readiness, 3.3 AI Search Visibility and 3.4 AI Search Opportunity together give you what those tools don't:

  • Readiness - what your page looks like to an AI engine
  • Visibility - whether you're actually being cited right now
  • Opportunity - what you'd need to cover to compete

Topic Expansion specifically maps to the way AI engines "fan out" a user's query into related sub-questions before retrieval - the more of that fan your page meaningfully covers, the stronger a citation candidate you become across the whole conversation, not just the single keyword.

Most other AI-search tools we've seen stop at brand-mention tracking - telling you you're losing without telling you how to win. Combining a structural readiness score, a live citation check and a gap-against-cited-peers analysis in one scan is pretty powerful stuff - and it's how you stay ahead as AI answers reshape organic search, rather than watching the traffic drop and wondering why.

3.5 Content (24% of total)

Content category top panel with overall score and four dimension panels: Keyword Usage, Content Quality, Readability, Heading Structure.

Category top panel opens with a plain-English diagnosis, an explainer, the category weight tag, and a horizontal bar chart of the 4 dimensions.

3.5.1 Keyword Usage (32% of Content)

Where your target keyword appears: H1, title tag, meta description, URL, first 100 words, body copy. Also measures keyword density (target 0.5-2.5%) and semantic variants.

Tip: The easiest category to move. If you're missing the keyword from your H1, fix it today - often shifts the score 15-20 points in a 10-minute change. But don't stuff - density above 3% reads as spammy and we'll flag it.
3.5.2 Content Quality (32% of Content)

Word count, topic depth (distinct H2 subtopics), and presence of supporting elements (lists, examples, data points). Thin content (<500 words) gets dinged; 1000+ words covering the topic from multiple angles scores well.

Tip: Word count is a proxy for thoroughness, not a target. Don't pad. Open the top 3 pages currently ranking for your keyword and check what subtopics they cover that you don't. Add those.
3.5.3 Heading Structure (18% of Content)

H1 presence and uniqueness, keyword in H1, hierarchy (H1 > H2 > H3 with no skipped levels), total heading count, descriptive vs vague heading copy.

Tip: Your H1 does two jobs - tells humans what the page is about and tells search engines (and AI) what to index. It should contain your target keyword and be the only H1 on the page.
3.5.4 Readability (18% of Content)

Flesch reading ease score (target 60+ for general audiences, 50+ for B2B), sentence length, syllables per word, paragraph length. Hard-to-read content loses both human readers and AI extraction quality.

Tip: Most B2B copy is over-formal because the writer is signalling expertise. AI engines (and humans) prefer plain language. Quick wins: split sentences over 25 words, replace jargon, shorten paragraphs to 2-3 sentences.

3.6 On-Page (20% of total)

On-Page category top panel with five dimension panels: Meta Tags, Images, Internal Links, External Links, Site Structure.

The HTML mechanics search engines and AI crawlers read first. Less glamorous than AI Search but cheap to fix and essential to get right.

3.6.1 Meta Tags (25% of On-Page)

Title tag length (50-60 chars optimal), meta description length (150-160 chars), keyword presence in both, OpenGraph and Twitter card tags for social previews.

Tip: Your title and meta description are your free advert in the search results. If your title gets truncated mid-word in the SERP, it's too long. If your meta description is missing or auto-generated, you're letting Google pick the snippet for you (badly).
3.6.2 Images (20% of On-Page)

Alt text coverage (target 100%), keyword in alt text where relevant, image dimensions/sizing, lazy loading, modern formats (WebP/AVIF).

Tip: Alt text is for accessibility, SEO and AI extraction. Three wins for one job. Describe what the image actually shows in plain language. Don't keyword-stuff.
3.6.3 Internal Links (20% of On-Page)

Number of internal links (10-30 typical), descriptive anchor text, keyword in anchor text where relevant, links to related pages on your site.

Tip: Internal links pass authority around your site. Every important page should have at least 3-5 internal links pointing to it from related pages. "Our CRM pricing" beats "click here" every time.
3.6.4 Site Structure (20% of On-Page)

URL cleanliness (lowercase, hyphens not underscores, no parameters, keyword in path), URL depth (2-3 levels typical), HTTPS, canonical tag presence.

Tip: Plan URLs at creation time: lowercase, hyphenated, keyword in the slug, shallow depth. Renaming existing URLs bleeds authority slightly through the redirect.
3.6.5 External Links (15% of On-Page)

Number of external links, nofollow vs follow, links to authoritative sources, broken external links.

Tip: Linking out feels counterintuitive but is one of the strongest credibility signals - both for traditional SEO and AI search. Aim for 2-5 external links to high-quality references in any substantial article.

3.7 Technical (10% of total)

Technical category top panel with overall score 98 and three dimension cards: Technical SEO, Accessibility, Best Practices.

The behind-the-scenes plumbing search engines and AI crawlers check first. If we can't measure a dimension during the scan - usually because the page requires login, hides content behind JavaScript, or the server times out - we assign a neutral score of 50 (the exact midpoint) so it doesn't pull your overall score up or down.

3.7.1 Technical SEO (55% of Technical)

Robots.txt, sitemap.xml, canonical tags, hreflang, structured data validity, mobile-friendliness, indexability.

Tip: Mostly "set once, ignore" but easy to break accidentally. Common gotcha: a developer sets <meta name="robots" content="noindex"> on production. The Pro Site Audit tool is perfect for spotting these across a whole site.
3.7.2 Accessibility (30% of Technical)

Automated accessibility audit - colour contrast, ARIA labels, keyboard navigation, semantic HTML, form labels.

Tip: Accessibility isn't just ethical - it's an SEO multiplier and increasingly a legal requirement (UK and EU regulations are tightening). Aim for 90+ out of 100. This is one metric worth going all the way for.
3.7.3 Best Practices (15% of Technical)

Automated best-practices audit - HTTPS, no console errors, valid source maps, no deprecated APIs, secure form handling.

Tip: Low weight by design - hygiene factors, not ranking levers. Fix the obvious stuff (HTTPS, console errors), then move on. Don't spend a week chasing 100 here when AI Search needs work.

3.8 Performance (6% of total)

Performance category top panel with overall score 53, Core Web Vitals and Page Speed dimension cards, and the full 6-metric drawer expanded showing LCP, TBT, CLS, FCP, Speed Index and TTI values.

Page load speed and rendering quality measured during the scan, split into Core Web Vitals (Google ranking signals) and supplementary speed metrics. Same fallback as Technical: if a dimension can't be measured (login-gated pages, heavy JavaScript, server timeouts), it gets a neutral score of 50 so it doesn't skew your overall score.

3.8.1 Core Web Vitals (60% of Performance)

LCP (Largest Contentful Paint, target <2.5s), TBT (Total Blocking Time, target <200ms), CLS (Cumulative Layout Shift, target <0.1). Google uses these as direct ranking signals.

Tip: Performance is a long-tail of small fixes, not a single big lever. Biggest wins usually: lazy-load below-fold images, defer non-critical JS, use a CDN, size images correctly. 80%+ in Core Web Vitals is competitive; chasing 100 is rarely worth the time.
3.8.2 Page Speed (40% of Performance)

FCP (First Contentful Paint), Speed Index, TTI (Time to Interactive). Don't directly rank but reflect perceived speed.

Tip: If Performance is dragging you down but the rest of the report is healthy, the highest-leverage fix is usually the hosting layer, not the page itself. Switching to a modern CDN-fronted host (Vercel, Netlify, Cloudflare Pages) often jumps Core Web Vitals 20+ points with zero code changes.
Note on variability: page speed measurements are sensitive to network conditions at the moment of the scan, so you'll sometimes see Performance wobble between otherwise identical scans. At the extreme this can nudge your overall score up or down by about 1 point - it's the nature of synthetic speed testing, not a real change on your site. Nothing to worry about unless the shift is persistent across several scans.

3.9 Keyword Opportunity

Keyword Opportunity panel - coverage gauge, stats row, related keywords variants table with volume/difficulty/CPC, and People Also Ask list.

After the 5 scoring categories, the Keyword Opportunity panel lists related keywords people search for alongside your target, with a coverage score showing how many already appear on your page.

Why it matters: each related keyword is a real search query where people could find your page - but only if your content actually mentions it. Every variant in the table that your page doesn't cover is a door you're not opening. Adding the high-volume, low-difficulty ones you're missing is one of the cheapest ways to get ranked for more searches at once - you're not writing new pages, you're just making sure the page you already have is visible to more of the traffic that's already looking for what you do.

Layout: a coverage gauge (top), stats row, insight callout, variants table (best related keywords with monthly search volume, difficulty 0-100, CPC, and "On Page" yes/no - top opportunities flagged automatically), and a People Also Ask list from Google's PAA box. If there isn't enough related search data for your keyword (usually very narrow branded queries), the panel simply doesn't appear on the report.

Tip: This panel is your editorial roadmap. The "Top opportunity" rows are keywords with real search demand and beatable competition that your page doesn't yet cover. The PAA list is even more direct: each item is a literal question Google has decided is related to your topic.
Tip: We've done the maths for you - rows tagged "Top opportunity" are ones where "Found on page: No" + "Difficulty: easy (under 40)" + "Volume: high (1000+/mo)" all align. Demand + low competition + you're not covering it = the easiest wins in SEO. Drop everything and add these keywords first.

Tip: Keyword specificity matters as much as keyword breadth - ambiguous keywords pull variants from the wrong topic universe entirely.

Example: a UK planning page scanned for "article 4" returns data about US Constitutional law (Article IV - interstate commerce, fugitive clause) instead of UK Planning Article 4 Directions. Same two words, two completely different topics.

If your suggestions look off-domain, add a disambiguating qualifier so the engines know which universe you're in ("article 4 directions", "article 4 planning", "article 4 HMO").

This applies across the whole report, not just the Keyword Opportunity panel - AI Search Opportunity and SERP Intelligence are equally susceptible to homonyms.

3.10 SERP Intelligence

SERP Intelligence panel - rank position gauge #2, comparison stats vs top 10, top 10 results table with You badge.

The "what does the SERP actually look like for this keyword" view. Pulls live Google results for your keyword + region, finds your page's position (if it ranks), and shows the top 10 with competitor metrics.

Keyword specificity: one page, two keyword variants

To show how keyword choice alone can move a ranking, here's the same Planning Direct page (/find-out-about/permitted-development/article-4-directions) scanned against two related keywords. Same HTML, same content, same domain authority - only the keyword changes.

Variant 1 - ambiguous keyword "article 4": ranks #4, but the SERP is polluted with off-topic results. Wikipedia's "Article 4 direction" (#3, planning-relevant), Ealing Council (#5), NATO's "consultation process and Article 4" (#7, completely different Article 4), Equality & Human Rights' "Article 4: Freedom from slavery and forced labour" (#9, yet another unrelated Article 4). Google doesn't know which "Article 4" the searcher means, so it serves a mix.

SERP Intelligence - Planning Direct ranks #4 for 'article 4' alongside Wikipedia, NATO, Equality Human Rights - a SERP mixed across unrelated Article 4 topics.

Variant 2 - specific keyword "article 4 directions": rank jumps to #2 (up from #4) and the competitor set cleans up completely - legislation.gov.uk, Dorset Council, London.gov.uk, Bristol Council, gov.uk. All UK planning, all authoritative, all directly relevant. Adding one qualifying word moved a real page 2 positions up the SERP AND eliminated four off-topic competitors.

SERP Intelligence - Planning Direct ranks #2 for 'article 4 directions' with a 100% UK planning authoritative competitor set (legislation.gov.uk, london.gov.uk, gov.uk, etc.).

The specificity lesson: Variant 1 → Variant 2 is the move most pages have waiting for them. Not a content change, not a schema change, not a link-building project - just picking a keyword that matches the page's specific intent. Now you know the perfect target, apply it and optimise for it by following OptimAI scan suggestions and get cited more by AI and rank higher in Google!

The "not found" diagnostic: competitive SERP defeat

Specificity isn't the only trap. Here's a different Planning Direct page - /find-out-about/permitted-development/loft-conversion, a page explicitly about loft conversions - scanned against the obvious keyword "loft conversion". Result: not found in the top 94.

SERP Intelligence - 'not found' state for 'loft conversion' on Planning Direct's loft-conversion page: rank gauge showing a dash, top 10 dominated by local pack and authority domains (fmb.org.uk, hoa.org.uk, reddit.com).

The keyword-page fit is correct this time - the page is literally about loft conversions. But the SERP is dominated by high-authority sites (Federation of Master Builders, HomeOwners Alliance, reddit) and local-pack business listings that Planning Direct can't outweigh on domain authority alone. This is a different failure mode: the page belongs in the SERP by topic, but the competitive ceiling is too high.

The diagnostic lesson: "Not found" doesn't always mean wrong keyword. Check the top 10 competitors - if they're all huge authority domains or local-pack listings you can't outrank, the realistic path is a long-tail variant ("loft conversion permitted development rules UK", "when do you need planning permission for a loft conversion") rather than the head term. SERP Intelligence shows you the ceiling before you spend months trying to reach it.

Tip: Competitive intelligence. Click into the top 3 results, study their structure (headings, length, FAQ presence, schema use), and compare to yours. If they all have a feature you're missing, that's a structural pattern Google is rewarding - copy it.
Tip: "Not found in top 100" is a hard ceiling - the keyword is too competitive for your current domain authority. Either pick a less competitive variant (long-tail) or accept this is a multi-quarter project.

3.11 History

History panel - current score gauge with delta vs first analysis, sparkline trend line, and per-scan table with date, overall, rating, SERP position, AI, AI Visibility, Content, On-Page, Technical, and Performance columns plus a View button on each row.

The History panel only appears when you have 2+ scans of the same URL+keyword pair. It tracks score changes over time so you can see whether your edits are moving the needle.

Layout: current score gauge with delta vs first scan, sparkline of overall score over time with red/amber/green bands at 40 and 70 thresholds, and a per-scan table with every historical scan (date, overall score, rating, SERP position, per-category scores, and a View button to open that historical report).

Tip: History is what turns OptimAI from a one-shot audit into an optimisation discipline. Set a cadence: scan, fix top 3 suggestions, refresh in 2 weeks (Google needs time to recrawl), watch the sparkline.
Tip: The per-category trend is more useful than the overall score for diagnosis. If overall has flatlined but AI Search dropped while Content rose, you probably accidentally broke schema markup while editing copy. The View button lets you reload the full report from any historical point to see exactly what changed.

3.12 Action Plan

Nearly last but certainly not least, the Action Plan is the original reason we built OptimAI and probably the most useful feature.

Action Plan tab on a Salesforce products scan - every dimension suggestion grouped by category (AI Search dimensions shown), each item with impact score and status badges.

Top-right of the report has two tabs: Report (the deep view we've just walked through) and Action Plan - a stripped-down to-do list of every suggestion across the report.

Top-right of the report view with the Report and Action Plan tabs highlighted - the switch between deep-dive view and todo-list view.

The Action Plan offers two views, one sorted by category & dimensions (see above) and the other sorted by priority (see below). This helps you focus your work by moving systematically through dimensions, for example overhauling Content across a set of pages in one push, or by sweeping pages for the most critical fixes that will move the needle fastest.

Action Plan tab in Priority view - every suggestion across the whole report ranked by calculated impact score (HIGH 18.36 at the top, descending), ignoring category grouping.

You can tick off suggestions as you work through. OptimAI tracks these between scans so you can see at a glance whether each item is resolved, still needs work (the changes you made weren't significant enough to resolve the root cause), or is new - generated since the last scan.

Tip - a few important things to remember:

  1. Suggestions are dynamic. The system updates progressively as you optimise your pages - you don't need to obsess about resolving every single one. Aim to get each category green overall; the list will surface the most significant issues naturally.
  2. Fixes often overlap. Resolving one issue can quietly resolve others at the same time - don't chase perfection.
  3. Work methodically: focus on a related batch of changes at a time (either by type or priority), rescan, see the progress, then pick the next batch.

3.13 Download PDF

Download PDF button highlighted in the bottom-right of the report view footer.

Finally the Download PDF button - does what it says on the tin! The bottom-right of every report has a Download PDF button which generates a PDF copy of the full report - same scores, same explanations, same tip text - rendered as a clean print-friendly document. It also includes the full Action Plan so you have a handy cheat sheet to work from.

Page 1 of a Salesforce products scan PDF export - overall 47 score, full heatmap of all 20 dimensions, and per-category score lists.
Page 1 of the Salesforce products PDF report: overall score, full dimension heatmap, and per-category breakdown.

Tip: Pro/Partner tier unlocks white-label PDFs - your branding (logo, colours, our footer removed) instead of OptimAI's. Covered in detail in Section 4 below.

The PDF is arranged with page breaks between each section so you can easily cut out pages you don't need using a tool like ilovepdf.com.

4. Pro Tier

Everything covered so far works on every tier. This section covers what Pro and Partner tiers add: whole-site audits, batch scanning, branded reports, and team workspace sharing. If you're scanning more than a handful of pages a month, or reselling audits to clients, these are the features that make the work repeatable and presentable.

Tip: Pro isn't required to use OptimAI. If you're running occasional single-page scans to validate specific edits, Standard is all you need. Pro kicks in when you start thinking whole site: "audit all 80 pages on our marketing site and give me a prioritised list" or "generate a client-branded PDF for every report I ship this week."

4.1 Standard vs Pro vs Partner at a glance

StandardProPartner
Site audit + page tree view-yesyes
Sitemap export-yesyes
Batch scanning-yesyes
Batch PDF download-yesyes
White-label PDFs-yesyes
Unlimited summary reports-yesyes
Domain intelligence-yesyes
Domain keyword opportunities-yesyes
Site health audit-yesyes
Team report sharing-yesyes
Credit pooling-yesyes
Billingcredit packscredit packsmonthly subscription
Credit rate1x1x1.5x
Early access research--yes
Early access new features--yes

The biggest line isn't a feature - it's the credit rate. Partner subscribers get 1.5x credits for their money compared to pack purchasers. £500/month on Partner buys 750 credits; £500 of credit packs buys 500. If you're running more than ~20 analyses a month, Partner is almost always the better economics and brings scan and refresh costs back to parity with Standard tier pricing.

4.2 Partner subscriptions

Partner is our self-serve monthly subscription for agencies, consultants and in-house teams running regular audits. Here's how it works:

  • Credits arrive on your billing day each month
  • Credits never expire - unused credits roll forward indefinitely
  • Cancel any time via the Stripe customer portal - your balance stays, you drop to the Pro tier
  • Had a busy month and need just a few more scans? You can always top up mid-month with regular credit packs for that bit extra.
  • VAT applies at 20% on GBP transactions; USD is VAT-exempt (rest of world)
  • Need more than the self-serve tiers? Contact us for custom pricing
  • Keep all your historic report data regardless of tier. You paid for it so it's yours forever.
Tip: Partner isn't a lock-in. Cancel whenever and your credits are still there - we don't wipe balances on downgrade. Safe to subscribe for a busy month, cancel when work eases off, resubscribe when it picks up. We designed it this way so you have flexibility and are always in control of your own budget.

4.3 Pro Dashboard

Pro tier has a slightly different dashboard with two levels: the top level shows your scans grouped by domain with aggregated values. Click on a domain row to drill down into a table listing the scans run for just that domain - otherwise the same as the standard tier dashboard, including the filter to quickly find scans.

Pro tier domain-grouped dashboard with Your Domains table - four domains listed with aggregated Pages, Scans, Avg Score, Avg Change, Credits, Last Scan and a Site Audit button per row.

From the domain dashboard you can also add a domain directly to run a site crawl. This will display the Site Audit page tree (more on this below). You can also enter this view for any existing domain in the list by clicking the Site Audit button in the table.

If you've already run a single-page scan, the domain will already be listed here and you'll find the scans on the drill-down for that domain - no need to re-add via the Add Domain button.

4.4 Site Audit - Pages tab

The page tree (or Pages tab) within Site Audit is your main place to overview the latest state of your scanned pages for that domain (or subdirectory/section) and run bulk operations.

At the top of every tab within Site Audit you will see domain level metrics: backlinks, domain rank, referrers and anchor text, etc. All standard stuff, except you don't get charged extra for it, unlike with some other tools! Next there is a large amber panel explaining how to use the bulk tools - you can dismiss this once you are familiar, and if you need a reminder just pop back here to these help documents.

Site Audit overview for www.salesforce.com/uk/media: domain metric tiles (backlinks, referring domains, domain rank, organic keywords, traffic value), Top Anchor Texts panel, and the amber 'How to Run the Site Audit Tools' help panel above the page tree.

After the explainer panel (or after the domain stat cards once you've dismissed it, as in the image below) you'll see the page tree, which is a sitemap displayed in a similar way to a file and folder tree on your computer. You use this to select pages you want to perform bulk operations on (e.g. batch scan, summary reports or batch PDF downloads) by clicking the checkboxes.

Page tree for www.salesforce.com/uk/media with the explainer panel dismissed: indented sitemap listing 12 discovered pages with checkboxes, path, and target keyword. The automated-production-broadcast-software row shows its multi-keyword dropdown open with two prior scans ('media workflow management' and 'streamline production and operations').

Tip: You can expand and collapse the page tree at subdirectories (i.e. "sections") using the arrow for quicker scrolling. Subdirectory pages also show the count of "child" sub-pages (including the top level "parent") in the pill badge on the left. Selections cascade through all child pages.

The top of the page tree table also shows you the count of pages which have been scanned and the total number of unlisted pages.

Each row in the table shows the page path (with an icon to open the target page in a new tab), the page title we detected, and a text input for the keyword. If a scan has been performed on that page before, you will also see the Score Bubble for the most recent scan matching that page and keyword pair, and the date of that scan. Clicking the Score Bubble will take you to that specific page report.

If you have run a scan against a page for multiple keywords, the keyword input will become a drop down and allow you to select which keyword report to include. (In the screenshot above you can see this for the page /uk/media/automated-production-broadcast-software.)

For pages which have not been scanned we auto-detect a keyword suggestion (derived from the page's H1 heading, with fallback to the meta title or the URL slug if no H1 is present) which will be shown in grey text. For pages which have been scanned the keyword which was used will be shown in black text. Any text you enter here will be black and used for the batch scan when you trigger it.

4.4.1 Site Audit - Pages -> Export

Site Audit footer toolbar with the Export button highlighted in the bottom right, sitting alongside Download PDFs and Refresh Sitemap.

Click the Export button (bottom right) to generate a Google Sheets version of the page tree sitemap. This will open in a new tab from where you can make a copy to edit or download as an Excel file. It includes checkboxes to select pages and a formula which calculates the credits required (including new scans or refreshes of previous scans at the different costs).

Google Sheets export of the site audit page tree showing 'OptimAI Site Audit - www.salesforce.com - uk - media' header, selected pages + estimated credits row, and columns for Include checkbox, Score, Last Scan, Page Path, Keyword, Credits, and Flags.

This is great to share with your team and plan your keyword strategy - for example listing out keyword variations you may want to run (just duplicate the original row). The file is formatted to follow the same grey/black keyword colouring as the live page tree, indicating suggested versus user-entered keywords.

4.4.2 Site Audit - Pages -> Download PDFs

Select some pages in the tree map and the Download PDFs button (next to the Export button, bottom right) will be enabled. Click it to download all the individual page report PDFs for each page selected in one zip file. (Note: if you have selected pages for which no scan has been run these will not be included, as there is no PDF generated from non-existent scans!)

4.4.3 Site Audit - Pages -> Scan

Site Audit sticky budget toolbar highlighted, showing 'Selected: 8 pages, 5 new (-80) + 3 refresh (-30)', 'Balance: 64 credits', and the Scan (-110) and Report (-80) action buttons on the right.

With some pages selected, the Scan and Report buttons will be enabled. They are found at the bottom of the page in what we call the Budget Bar - this does the maths for you and calculates the cost based on new scans and refreshes, showing the total cost in the Scan button itself and also your current balance (red if insufficient credits).

Click the Scan button to generate new page reports or refresh existing ones for the pages you have selected in the page tree. This runs as a batch operation which you can leave running in the background, as it may take some time. You can opt to receive an email notification when the batch is completed.

Remember to check your keywords before clicking! Once you click, the credits are deducted. (By the way, if any scans fail at any time you will get those credits automatically refunded.)

4.4.4 Site Audit - Pages -> Report

Budget Bar highlighting the '5 new (-80)' counter on the left and the Report (-80) button on the right.

Reports run from the Site Audit page tree view are executive summary reports collating and aggregating the data from multiple individual page reports. You can generate an unlimited number of these reports for free, so you can slice and dice your site across whatever pages you like in as many ways as you like. We'll dive into further details about the summary reports below.

Tip: Running summary reports is free but you still need to have the individual page scan data to run a summary report against! The summary will use the most recent scan for the pages you have selected in the page tree. Remember to check keywords and switch to the variant you actually want for the summary if there are multiple historic scans against a page (using the drop down as explained above).

If you have selected pages that have not had a scan run yet, then you will be charged credits to run the initial scan so it can be included in the report. This is why you see a charge of 80 credits in the image above corresponding to the 5 new scans required by the pages selected in this case.

Also, don't forget to check the recency of previous scans before running a report - you may want to refresh some of these for the latest data first. The Report button will not automatically run refreshes for any existing scans.

4.5 Site Audit - Reports tab

Site Audit with the Reports tab highlighted in the top-right tab switcher, showing four saved executive summary reports for www.planningdirect.co.uk each with overall score, category scores, distribution bars, improvement deltas, and AI citation counts.

Clicking the Reports tab at the top right of the Site Audit page will take you to a listing of all previously-run executive summary reports. You will see a summary panel for every report showing headline results. Click the panel to view the full summary report. Click the trash can icon in the top right of a panel to delete a summary report (don't worry, all individual page reports will be preserved).

4.6 Site Audit - Keywords tab

Site Audit Keywords tab for evolve-businesssolutions.co.uk, with the Keywords tab button highlighted in the top-right tab switcher and the Opportunities / Unscanned / Clusters sub-navigator highlighted in the centre.

The Keywords tab is your domain-level keyword intelligence view - what the whole site is ranking for, which pages are pulling the rankings in, and where the biggest untapped opportunities sit across the entire domain. Essentially this is the site-wide version of the Keyword Opportunity panel in single page reports (see Section 3.9 above).

There are three sections, you can jump to each from the page nav highlighted in the image above:

  • Opportunities - keyword-page pairs you're already on page 2 or 3 for and could get to page 1 of Google with a push
  • Unscanned - pages on your site already ranking for keywords but not yet audited in OptimAI
  • Clusters - thematic groupings of keywords, a lens on which topics your domain has authority in and where the gaps sit
Tip: Treat this tab as your weekly or monthly "what's moving on my domain" dashboard. It refreshes with every scan, and the Opportunities view alone will usually surface more actionable wins in a single afternoon than most paid keyword-research tools hand you in a month.

4.6.1 Site Audit - Keywords -> Opportunities

Keywords tab Opportunities view listing keyword+page pairs ranked by achievability, with position, volume, difficulty, CPC, target page and per-row Scan action.

The Opportunities view lists keywords your domain is nearly ranking well for - typically sitting just outside the first page. Each row is a specific keyword + page pair: the phrase, where you rank today, search volume, difficulty, CPC and the page pulling the ranking.

Why this matters: moving a keyword from position 14 to position 4 is vastly cheaper than getting a new page to rank at all - Google already knows you for that query, you just need a targeted optimisation push. Click Scan on any row to run a full OptimAI audit for that page+keyword pair and get a prioritised to-do list for getting it over the line.

Tip: Watch for the Top opportunity badge - those are the rows where the ranking difficulty is realistic and you're already within striking distance of the first page. That triage logic is the expensive part of SEO work - most tools hand you the raw list and leave you to rank it by hand. We've done it for you, so start at the top and work down.

4.6.2 Site Audit - Keywords -> Unscanned

Keywords tab Unscanned view: pages on the site that are already ranking but haven't been audited yet, expandable to reveal ranking keywords per page with per-row Scan action.

Unscanned lists pages on your domain that are already ranking for one or more keywords but haven't been through an OptimAI audit yet. Expand any row with the triangle to see which keywords that page is ranking for, their positions, and any opportunity badges attached. Click Scan to run a in-depth report and add this page+keyword combo to your optimisation workflow.

This view is where you catch the outliers - pages you'd never have thought to scan. That old blog post from 2021 quietly ranking #3 for something you never targeted on purpose. The news announcement that flukes on the exact right recipe for AI search readiness. Every domain has a handful of these accidental winners, and they're often the fastest pages to convert into genuine traffic assets once you know they exist. Picks up directly on the point in section 4.4: you don't need to scan every blog or news post to find the nuggets - this view surfaces them for you.

Tip - we've done the prioritisation: rows and keywords tagged Top opportunity are where ranking position, search volume and keyword difficulty all line up such that a small optimisation push should land a real result. No spreadsheet crunching, no scoring rubric to memorise - just scan the page, work the suggestions, refresh, watch the position move. That prioritisation ranking is what every serious agency builds a custom spreadsheet for. We hand it to you on a plate for free!

4.6.3 Site Audit - Keywords -> Clusters

Keywords tab Clusters view: thematic groupings of keywords (e.g. 'managed print') with counts and best positions, expand triangle on each cluster reveals the individual keywords, positions and opportunity badges.

Clusters group the keywords your domain ranks for into thematic buckets - related phrases that all circle the same underlying topic. Each cluster shows its label, how many keywords sit inside it, how many pages on your site contribute to it, and the best ranking position across the group. Click the triangle on the left of any cluster to expand it and see the individual keywords, positions and opportunity badges inside.

Why clusters matter: they're the clearest view of where your domain has topical authority and where the gaps sit. A dense cluster (lots of keywords, lots of pages, strong positions) is a topic Google already sees you as credible on - extending it with a hub page or supporting article typically compounds fast. A sparse cluster hinting at demand you're barely addressing is a signal to either create new content or weave the variant keywords into existing pages for additional pull without building anything new.

Tip - two badges, two different plays:

  • Top opportunity - realistic difficulty AND already within striking distance of the first page. Worth your time: scan the page, work the suggestions, push it up.
  • Quick win - a narrower subset: keywords sitting just off page 1 where a small nudge often flips them over into real traffic. Lowest effort-to-impact ratio in the whole tab.

Mental model: Top opportunity = worth doing. Quick win = do this first.

4.7 Site Audit - Site Health tab

Site Audit Site Health tab: category summary tiles (Broken, Redirects, Canonical Issues, Orphan Pages, Unlisted Pages) with expanded URL lists for each category.

The boring hygiene stuff - but more important than most people realise. Every item in this tab is an issue quietly bleeding ranking authority, wasting crawl budget, or keeping pages out of Google's index entirely. None of it is glamorous to fix. All of it tends to be overlooked - which is precisely why it's a reliable win when you don't.

Five categories, each summarised by a count tile and expanded below into the full list of offending URLs (full tooltip on each tile if you need a refresher):

  • Broken - URLs returning a 4xx error. Dead links signal neglect to both Google and users, and waste crawl budget on every scan.
  • Redirects - URLs that 3xx somewhere else. Individual redirects are fine, but chains (redirect-to-redirect) leak link equity and slow page load.
  • Canonical issues - misconfigured or conflicting canonical tags. Google will pick one - usually not the one you'd have picked - and split the ranking weight across the others.
  • Orphan pages - pages in your sitemap with no internal links pointing to them. They exist but nothing's sending them authority, so they rarely rank.
  • Unlisted pages - pages discovered by crawl but missing from your sitemap. Google may never index them - and sometimes it surfaces pages that shouldn't be live at all (stray test pages, content missing from the sitemap by accident).

4.8 Summary Reports

Executive Summary report for a 20-page domain audit showing overall score 61, category score gauges, full dimension heatmap, distribution/improvement/AI citations summary panels, and Improvement Tracking with portfolio average chart and per-page score deltas.

Summary reports - also called Executive Summary reports - take a whole audit (a batch of page scans, a section of a site, a client engagement) and condense it into a single scrollable page you can share, present or hand off. Every panel is an aggregate of the pages you chose when the report was generated, so the output matches exactly the slice of the site you care about.

The top panel (shown above) is the headline view: overall weighted score across the whole set, the five category gauges, a distribution bar (how many pages sit in good / fair / poor), Improvement headline stats (how many pages moved, up or down, since the previous scan), AI Citations headline stats, and the full Dimension Heatmap - this time showing the average score per dimension across the whole set, so a red tile means the set of pages included in the report is weak on that signal, not just one page.

At the top of the page is a section nav which allows you to jump directly to that section of the report, each of which we will walk through below.

Tip: This is the artefact most agencies actually hand to clients. Every panel is designed to be readable by a non-technical stakeholder at a glance - traffic-light gauges, plain-English recommendations, one-line issue summaries. Exports to branded PDF (see 4.10 Branding below) so you can rebadge and send.

4.8.1 Summary -> Improvement

Summary Improvement section with Improved/Declined/Unchanged/Avg Delta headline stats, portfolio average score sparkline over time, and a per-page list showing score deltas (+6, +3, +1, 0, -1) with dates.

Headline stats first: how many pages have improved, declined or stayed unchanged since the last scan, and the average score delta across the whole set. Beside that, a sparkline of the page set average score over time (populated once you have more than one scan run), then a per-page list of movers with the score change per page.

Tip: This is the panel to open first in any progress review. That single "+1 avg delta across 20 pages" figure is the cleanest way to quantify whether the sprint actually moved the needle. Better than promising vague gains - hand the stakeholder a number.

4.8.2 Summary -> AI Citations

Summary AI Citations section with domain-wide citation stats (pages cited / citation rate / total citations), engine filter pills, average citation rate history chart, and per-page rows with mini sparklines and hit/miss/gain/loss indicators per engine.

Domain-level citation intelligence across the page set. Headline stats at the top (pages cited / overall citation rate / total citations). Buttons for each engine showing total citation count which you can click to filter to Google / OpenAI / Perplexity individually. Alongside that, an average citation rate history chart shows how the rate has moved over time - builds up as you run multiple scans.

Below there's a table listing each included page with a mini citation sparkline per row, plus hit / miss / gain / loss indicators for each engine. Expand any page to see its full Best-of-Day bubble chart - same pattern as the single-page AI Search Visibility panel, so no new mental model to learn.

Tip - watch the trend, not the number: AI engines are inherently non-deterministic - a single scan's citation rate bounces. What matters is the direction over weeks. If your citation rate is climbing while your traditional rankings stay flat, the AI search optimisation work is already paying off - just before Google shows you the credit.

4.8.3 Summary -> Recommendations

Summary Strategic Recommendations panel: numbered list of five domain-wide plays (FAQ section, concrete data, untargeted keywords, cannibalization, sourced statistics) with High/Medium/Low priority pills and a short rationale each.

Next we've reviewed all the reports for each page and distilled this to critical strategic plays - the "if you only did five things across this whole audit" list. Each recommendation is tagged High / Medium / Low priority with a one-paragraph rationale based on the specific pages and patterns that triggered it. Hand this to whoever's briefing the next optimisation sprint - it's the executive summary of the executive summary.

4.8.4 Summary -> Keyword Opportunities

Summary Keyword Opportunities table: keywords scoped to the report's page set, with volume, position, difficulty, CPC, On Page status and Top Opportunity badges. Hover tooltip shows which pages surfaced a given keyword.

Same pattern as the single-page Keyword Opportunity panel and the domain Opportunities view - just scoped to the pages in this report. Keywords with real search demand, with the usual Top opportunity badges flagging the achievable-and-valuable ones.

Click any row for a popup that reveals which pages from the report surfaced the keyword, so you can click through to the page and run a fresh report for the new opportunity keyword on that page. This is the same "from insight to action" flow as everywhere else in the tool.

4.8.5 Summary -> Cannibalization

Summary Keyword Cannibalization panel showing a cluster like 'planning appeals' with two competing pages, severity badge (MILD) and Strongest indicator on the higher-scoring page.

Flags cases where two or more pages in the set are chasing the same keyword and fighting each other for rankings. The severity pill (MILD / MODERATE / SEVERE) tells you how bad the overlap is; the Strongest badge tells you which page is currently winning, so you know which page to consolidate onto.

Tip: Cannibalization is the silent killer of site structure. Two pages both targeting "widgets" pull roughly half the authority each - consolidate into one canonical page plus a 301 from the other, and the combined page typically outranks either did on its own. Classic "less is more" SEO.

4.8.6 Summary -> Top 5

Summary Top 5 Strongest Pages panel listing the highest-scoring pages in the set (scores 66, 64, 64, 64, 62) with path, keyword and short commentary, alongside the start of the Top 5 Priority Fixes panel below.

Two panels back to back in the report flow: Top 5 Strongest Pages (your current winners - the templates, content patterns and structural choices that are already working) and Top 5 Priority Fixes (the highest-impact pages to work on next, sorted by expected movement). The strongest list doubles as a brief for replication: whatever those five pages are doing, do more of it on the rest of the site.

4.8.7 Summary -> Common Issues

Summary Common Issues panel showing recurring suggestions across the page set with counts like 20/20, 17/20, 16/20 indicating how many pages each issue affects. Shown here all from AI Search category.

The recurring suggestions across the whole page set - issues that hit a meaningful share of pages, not just one-offs. Each row shows how many pages are affected (e.g. "20/20" = every page, "16/20" = 80%). Ordered by severity then by reach. Capped at the top 10 most significant issues and filtered to anything affecting at least 25% of the set, so it's always actionable rather than overwhelming.

The panel covers all five categories (AI Search / Content / On-Page / Technical / Performance) - in the example above everything happens to be AI Search because this particular site is strong on the fundamentals and weak on AI readiness, but the list spans the full scorecard whenever issues are spread across dimensions.

Issues surfaced here update as you fix them. Once a given issue drops below the 25% pages-affected threshold it falls out, and whatever's next most common takes its place. That turns this panel into a live scoreboard rather than a fixed report - each time you regenerate the summary after a work sprint, the list reflects what's left to do.

Tip - this is the core value prop of OptimAI: a personalised, prioritised, actionable to-do list for an entire section of your site.

Other audit tools (Semrush Site Audit, Ahrefs Site Audit - we're looking at you both) do roughly the same crawl-hygiene job our own Site Health tab does - broken links, redirects, canonical issues, orphans - plus a few hundred extra technical checks. Useful, and a fair chunk of any audit.

Where they stop is right there: a giant tracker of crawl issues for you to triage.

They don't do the per-page content + AI-search analysis that produces this Common Issues list (99 ranking metrics across 20 dimensions, run against each URL + keyword pair, then aggregated up). They don't aggregate it into a prioritised action plan. They hand you a spreadsheet and wish you luck. We hand you a sprint plan.

4.8.8 Summary -> Breakdown

Summary Categories Breakdown table: every page in the audit with SERP rank, overall score, and per-category scores for AI Search / Content / On-Page / Technical / Performance. Each row shows scan date, keyword and path.

The per-page matrix. Every page in the set, with its SERP rank, overall score and scores for each of the five categories. Click any row to open the full single-page report underneath.

This is the panel for comparison over time. Rerun the summary report for the same page set each month and drop the tables side by side - the shape of the scores tells you which categories are moving and which are static, per page. It's invaluable for month-by-month client reporting: same visual, different numbers, clear trajectory.

4.9 Team Sharing (domain workspace)

Settings menu open from the avatar bottom-left, with the Team Sharing option highlighted.

OptimAI does team sharing differently: no invite flows, no seat management, no per-user billing. We use your email domain as the workspace boundary.

Sign in with your company email (alice@agency.co.uk). Anyone else signing in with the same domain (bob@agency.co.uk, charlie@agency.co.uk) is automatically part of your team.

Tip: Generic "public" domains that are available to anyone such as gmail.com or outlook.com are blocked from sharing - you must use a custom "owned" domain for your business running on Google or Microsoft platforms. To be clear, you may not realise that your work email is either Google or Microsoft but in 99.99% of companies it will be - ask your IT support to confirm which if necessary. You do NOT need to set up a gmail.com or outlook.com email account to access OptimAI - and nor should you if you want to use the team sharing feature!

To enable sharing, find the Team Sharing option from the settings menu and toggle it on there.

Sharing requires 2-way consent: if members of your team have it enabled and you don't then they won't see your reports and you won't see theirs. Once enabled both ways reports are shared (per page and summaries), branding (if enabled, see below) is applied for PDF exports and your team can operate a shared credit pool.

The shared credit pool will automatically deduct credits for reports run by any team member from the account with the highest credit balance - so, for example a manager could have an account and control budget through credit purchases as desired, while other members of the team can use the tool freely within that budget limit. There is of course nothing to prevent you or other team members adding credits too, but these won't get used until all other team members have less credits on their balance.

Tip: When enabled you will see a badge next to reports in listings for the "owner" i.e. team member who ran that report - this applies to single page reports and executive summary reports. On the transaction page if you run a report using the credit pool it will show as a pool transaction; if you have the "manager" account with the budget that is getting deducted, you will see reports run by other team members with their username in your transaction ledger so you can easily keep track of who is spending budget.
Tip: Team sharing is free. Collaboration doesn't cost credits - scans do. Adding five team members costs nothing extra; they just all use the same pool and see the same history.

4.10 Branding (white-label)

Settings menu open from the avatar bottom-left, with the Branding option highlighted.

Pro users can configure a brand profile that applies to all exported PDFs - individual reports, batch reports, and executive summaries. The profile is domain-level: configure once at @yourcompany.com, every team member using that domain sees the same branded output. The first user in your company to enable and save these settings will become the "brand owner" and these settings will be locked out for other members of the team to prevent un-approved editing.

Brand Settings modal: Apply custom branding toggle, logo upload, primary and secondary hex colour pickers (#02605C / #2A8A81), header and body font dropdowns with weight selector, live Main Header Style preview, Save Brand Settings button.

Configurable:

  • Logo - PNG, JPG, SVG, or WebP (up to 2MB). Replaces the OptimAI logo in the PDF header.
  • Primary colour - hex value. Scores, section headings, accent bars.
  • Secondary colour - hex value. Secondary accents.
  • Header font - any Google Fonts family available from the built-in font-picker.
  • Header font weight (100-900) - where available for the chosen font.
  • Body font - any Google Fonts family available from the built-in font-picker.

Note: the OptimAI footer will also be removed completely.

Tip: Pick colours that look good on white (PDF backgrounds) AND on your brand's own collateral - the branded PDFs are designed to drop into client decks and printed reports, so the palette needs to travel.
Tip: Both single page report and summary report PDFs are organised with page breaks between the sections. This allows you to prune the reports to only present selected data - maybe you want to reduce information overload, or you're not commissioned to deal with a certain category of issue, or you're not actually dealing with or presenting certain types of info; maybe all you want to focus on is AI Search Visibility, etc.

5. When to Scan

OptimAI is intended as an optimisation discipline, not just a one-off audit (though obviously you can do that if you like). The real compounding value comes from running scans on a cadence that matches the work you're doing. Run scans too often and you burn credits on noise; run them too rarely and you miss regressions. Here's a few suggestions on when and how often to scan as a working starting point.

5.1 Cadence by scenario

ScenarioRecommended cadence
Just published a new pageScan on publish to capture the baseline
Made batch of edits to an existing pageRefresh 2 weeks after deploy (Google needs time to recrawl) to capture changes (but make sure to have that baseline first to compare to!)
Actively optimising a high-priority pageAs and when you've committed a bunch of changes during the active sprint - SERP will lag of course
Steady-state contentMonthly if you want to keep an eye on things, or don't bother
Quarterly board reviewBatch-refresh every priority page so you can regenerate the executive summary with current data
Before a client presentationFresh scan of every referenced page, then PDF export
After a Google algorithm updateRe-scan all priority pages to see what moved and how
Competitor researchWhen you want to see what they are doing well or how you're outperforming

5.2 AI Visibility is non-deterministic

AI engines don't always cite the same sources on the same query. A single "Not Found" on Perplexity for your keyword doesn't mean you're invisible - it could just be this run's roll of the dice. For AI Visibility specifically:

  • Run at least 2-3 scans over a 1-2 week window for a reliable read
  • Watch the Citation Timeline sparkline - trends matter more than single data points
  • Don't re-scan mid-edit (you'll confuse cause and effect)
  • For Mentioned → Cited work, fix the structure and refresh 2 weeks later so Google has time to recrawl and AI models have fresh content to pull
Tip: Budget an extra monthly scan on your 3-5 most-important pages specifically for AI Visibility tracking. It's cheap insurance - you'll spot citation drops before they turn into traffic drops, and you'll have the timeline data to reassure stakeholders that short-term dips are noise.

5.3 Chase movement, not scores

The absolute score matters less than the trend. A page moving from 47 → 68 over three months is winning faster than a page that hits 85 and stays flat. The History panel is the view to watch for this - it's the only way to tell whether your edits are doing anything.

Tip: The best weekly habit is Monday morning batch-refresh of your priority pages, then 20 minutes reviewing the Needs Work and New badges in the Action Plan. Keeps you in touch with what changed without becoming a full-time measurement job.
Tip: Don't re-scan after every tiny edit. A well-scoped edit session should make at least 3-5 related changes, then you can refresh once to see the aggregate impact. Scanning after each individual change burns credits and the score movement is often within the noise floor.

5.4 What good looks like

Follow the signals OptimAI provides - we've carefully designed the tool to readily surface the highest-priority issues to the top. For most sites that have had at least some attention paid to SEO, when you first run a scan you'll probably end up with scores in the 45-65 range. This is normal, and in most cases it'll be because you haven't even begun to implement AI Search optimisation techniques (which OptimAI gives 40% of total score weight to, as that is the future). Don't panic.

As a general guide, you want to focus on getting the overall category scores into the green. You don't need to obsess about getting every dimension perfect, let alone every individual metric. That level of detail is really there to highlight what, if anything, is bringing down your overall score and holding you back. If you're stuck in red for FCP but the overall Performance category is 72, don't sweat it.

Remember, follow the Dimension Heatmap - it shows you which specific dimensions have the biggest impact, and those are the ones to focus on first.

Our score thresholds are set at 40 for fair and 70 for good, which we believe are honestly realistic. If you're getting red scores below 40, you really will be impacted and should sort that out. Getting over 70 is genuinely challenging, but that's precisely the level where you start to rise above the competition. Pass 80 and you're exceptional - that's the page that earns the traffic and the AI citations. It absolutely is possible to get green across the board - work through the suggestions systematically and you'll see your scores climb.

FAQ

Product fundamentals

What does OptimAI actually measure?

Each scan audits a single URL + keyword pair against 99 ranking metrics, rolled up into 20 dimensions across 5 categories: AI Search Readiness (40% of total), Content (24%), On-Page (20%), Technical (10%), Performance (6%). The scoring is deterministic - no AI in the scoring itself - and every metric, dimension and weight is visible in the report.

Does OptimAI rank my page?

No. We audit your page against known ranking factors and AI-search readiness signals, then hand you a prioritised list of what to fix. Actual ranking happens on Google. Actual citation happens inside AI engines. We measure your readiness for both.

How long does a scan take?

Around 45-60 seconds. The report opens automatically when it's done. See section 1.3 for the first-time walkthrough.

Why is the same page scoring differently for different keywords?

Every scan is scoped to one URL + keyword pair. Different keywords surface different strengths and weaknesses on the same page - keyword usage, opportunity coverage, SERP rank and AI citation behaviour all change with the keyword. Scanning the same page against multiple keywords is a legitimate and useful pattern (each one is a separate dashboard row) - covered in detail in section 2.7.

Why is my score low when the page seems fine?

Most often it's a keyword / page mismatch. A brand-hub page (e.g. /products/) won't score well for a commercial keyword like "best CRM software" because that intent belongs on a dedicated landing page. Before you fix the page, check you're scoring it against a keyword it was actually built to win. If the keyword is right, the report will tell you exactly which dimensions are dragging the score down. The Salesforce worked example in section 1.4 shows this pattern end-to-end.

How is this different from Semrush or Ahrefs?

Their Site Audit products do roughly the same crawl-hygiene job our Site Health tab does - broken links, redirects, canonicals, orphans. They give you breadth (lots of technical checks) but they stop at "here's a giant list of issues, you triage it." We go further: deep per-page content analysis, AI search readiness scoring, prioritised action plans, AI citation tracking. They hand you a spreadsheet. We hand you a to-do list.

AI Search

Does a high AI Search Readiness score guarantee AI engines will cite my page?

No. AI engines produce non-deterministic responses - their citations shift with every query, user and model update. A high score means your content is well-structured for extraction; a low score means you're effectively invisible. Think of it like dressing for a job interview: nailing the schema, entity naming and Q&A structure doesn't guarantee they cite you, but turning up in pyjamas guarantees they won't. Full detail on the readiness signals we measure is in section 3.2.

What's the difference between AI Search Readiness, Visibility and Opportunity?

Three distinct measurements, three different jobs:

  • Readiness - structural: how well your page is built for AI extraction (schema, Q&A, entities, freshness, etc.). 40% of your total score.
  • Visibility - live check: are AI engines actually citing your domain right now for this keyword?
  • Opportunity - gap analysis: reverse-engineers what AI engines are answering for this keyword and flags what's missing from your page.

Most other tools track only one of these (usually brand-mention monitoring). Combining all three is the OptimAI difference.

Why are my AI citations bouncing between scans?

AI engines are inherently non-deterministic - the same query won't always return the same citations. A single "Not Found" doesn't mean you're invisible; it could just be this run's roll of the dice. Watch the trend, not the single number. The Citation Timeline sparkline is the right view for this - patterns over weeks tell you whether you're moving the needle, single scans don't. See section 3.3 for the full visibility view, and section 5.2 on cadence specifically for AI Visibility tracking.

Pricing & credits

How much does a scan cost?

Standard tier: 12 credits for a new analysis, 8 for a refresh. Pro tier: 16 / 10. Partner subscribers get 1.5x credits for their money - effectively the same per-scan cost as Standard while keeping all Pro features. New accounts get 120 free credits (10 Standard analyses) with no card required. Full breakdown in section 4.1.

What counts as a refresh vs a new analysis?

Rerunning a scan on the same URL + keyword combination you've already audited is a refresh (reduced credit cost). A scan on a different URL, or the same URL with a different target keyword, is a new analysis (full cost).

Do credits expire?

No. Credit pack purchases never expire. Partner subscription credits also carry forward indefinitely if unused.

What happens if I cancel my Partner subscription?

You drop back to Pro tier instantly, keep all unused credits, keep access to all Pro features, and simply pay per-pack rates from then on. We don't wipe balances on downgrade. Full Partner subscription mechanics in section 4.2.

Can I get a refund if a scan fails?

Yes. Failed scans auto-refund credits. Batch scans handle this per-page - if one page in a batch fails, just that page's credit is refunded automatically.

Pro, Team & Branding

What's the difference between Standard, Pro, and Partner?

Standard and Pro are both free tiers - neither costs anything to be on, you just pay for credits when you need them. Standard gets you per-page scans. Pro unlocks everything else: site audit, batch scanning, white-label PDFs, summary reports, domain keyword intelligence and team sharing - upgrade is one click from the settings menu, no charge. Partner is a monthly subscription that gives you 1.5x credits for the same money plus early access to new features. Full breakdown in section 4.1.

Why is my gmail / outlook account blocked from team sharing?

Team sharing uses your email domain as the workspace boundary. Generic domains anyone can sign up for (gmail.com, outlook.com, yahoo.com, hotmail.com) can't share - we'd have no way to know if you're actually colleagues. You need a custom business domain running on Google Workspace or Microsoft 365 (which is 99.99% of company email setups). You don't need to set up a separate gmail / outlook account to use OptimAI - and you shouldn't if you want sharing. Full team-sharing mechanics in section 4.9.

Who controls our brand profile?

The first user on your domain to enable and save the brand settings becomes the brand owner. Other team members on the same domain see the active branding on their PDFs but can't edit the settings - this prevents accidental edits to client-facing output. Transfer ownership via support if needed. Full branding configuration in section 4.10.

Can I share scans with people outside my company?

Yes - via PDF download. Every report exports to a clean print-friendly PDF (white-label branded if you've configured a brand profile). Team sharing inside the app is domain-locked by design (can't be cross-organisation), but the exported deliverable is yours to send anywhere.

What's the maximum site I can audit?

500 pages per discovery crawl - that's how many pages Add Domain will surface in the page tree. From there, batch scanning handles up to 100 selected pages per batch (so to scan all 500 you'd run 5 batches; the more common workflow is to pick the high-value pages and skip the rest). For larger sites, scope the discovery to a sub-path (e.g. example.com/products) and run separate audits per section. Site Audit walkthrough in section 4.4.

Edge cases & limitations

Why are some Technical or Performance metrics missing from my scan?

Some sites (typically large enterprise domains with aggressive bot protection / WAF) block external page-speed audits at the network layer. When that happens we can't measure those dimensions, so we assign a neutral score of 50 (the exact midpoint) so it doesn't pull your overall score up or down. The metric breakdown for those dimensions won't render. The same fallback applies to login-gated pages and pages that hide their content behind heavy JavaScript - both are detection-resistant by design. See section 3.7 and section 3.8 for the affected categories.

Can I scan a page that requires login?

No. OptimAI fetches pages anonymously, so anything behind a login wall isn't visible to us - or to Google, or to AI engines either, which is the bigger point. If your most important content is gated, that itself is the SEO problem.

Can I scan JavaScript-heavy or single-page-app sites?

Best-effort. We fetch the rendered HTML of the page, but pure client-side-rendered content (React / Vue apps with no server-side rendering) may not be fully parsed. Affected dimensions get the same neutral fallback as login-gated content. If your site is SPA-only, the fix is server-side rendering or static generation - not just for us, but for every search engine and AI crawler.

Does region affect my scan results?

Yes. Search volumes, keyword rankings, SERP composition and competitor data all reflect the region your account is set to. We auto-detect region from your IP, but you can override it from the settings menu. Set region before you scan - existing scans stay tied to the region they were run under. Region picker covered in section 2.6.

Do you support other languages?

English only at launch. More languages coming post-launch based on demand.

Data & privacy

Is my scan data shared with anyone?

Only with teammates on the same email domain, and only if team sharing is enabled on both sides (see section 4.9). We never sell or share scan data with third parties.

Do you keep my scan history forever?

Yes - lifetime history. We don't delete, archive or expire scans. The History panel and dashboard show every scan you've ever run. Storage is negligible and the data feeds the trend views that make optimisation work meaningful over time.

Can I delete my account or scan history?

Account deactivation is self-serve from the Account modal - your scan history and credit balance are preserved if you want to reactivate later. Full deletion of historical scan data is available on request via the contact form below.

Workflow & cadence

How often should I rescan a page?

Refresh after making edits to verify the fix landed - 2 weeks after deploy if you want SERP / ranking signals to update too (Google needs time to recrawl). Otherwise monthly is plenty for steady-state content. The full cadence guide is in Section 5.

Can I scan competitor pages?

Yes - it's a legitimate competitive-intelligence move. Same per-page model. Scan a competitor's top-ranking page against your target keyword to see exactly how they're winning structurally; the gap between their score and yours is your roadmap.

Does the Action Plan persist between scans?

Yes. Suggestions update progressively as you fix them - resolved items are tracked, items that come back show a "Needs Work" badge, brand-new items get a "New" badge. The list is a live to-do board across the lifetime of the page, not a fixed snapshot.

Can I export my data?

Yes. Every individual report exports as PDF (with white-label branding on Pro). Pro accounts can also export the site audit page tree to Google Sheets for keyword strategy planning, and download zipped batch PDFs for the entire selection in one click.

Contact Support

Got a question, a bug to report, or a feature request? Send us a note. We reply from a real inbox, usually within a working day.