Agency How-To Guide

The Complete AI Search Audit Template: How to Run AEO Audits for Any Client

A step-by-step AI search audit template for agencies and in-house teams. Covers baseline citation measurement, content structure analysis, schema markup review, competitor benchmarking, and the prioritised action list that turns audit findings into improved AI visibility.

Devanshu
9 min read
Featured image for The Complete AI Search Audit Template: How to Run AEO Audits for Any Client

An AI search audit answers one question: why is this site being cited by AI engines at the rate it is, and what would change to improve that rate? Without a structured audit, AEO optimisation becomes guesswork - adding schema markup here, rewriting a heading there, without a clear picture of what is actually limiting citations.

This is the audit template used for AI visibility engagements. It covers every dimension that affects AI citation rates, produces a prioritised finding list, and is repeatable across clients in any industry. Work through the sections in order - each builds on the previous.

Before You Start: Establish Scope

Define these before running the audit:

  • Target AI engines: Which of the four (ChatGPT, Gemini, Perplexity, Claude) matter most for this client's audience? For B2B clients, all four matter. For consumer brands, prioritise ChatGPT and Perplexity. Run the full audit across all four.

  • Core keyword set: 10-20 queries representing how the client's potential customers research their category in AI engines. Include category research queries, competitor comparison queries, and problem-solution queries.

  • Key competitors: 3-5 direct competitors to benchmark against. Competitor AI visibility data contextualises the client's performance.

  • Audit period: Baseline citation rates are measured over 2-4 weeks to account for AI response variability. Immediate spot-checks are useful; statistical rates require repeated measurement.

Section 1: Baseline Citation Rate Measurement

This is the foundation of the audit. Without a baseline, nothing else has context.

1.1 Current citation rate by engine

For each of the 10-20 target queries, record whether the client brand is cited in the AI engine's response. Run each query at minimum 3 times across different sessions to account for response variability. Calculate citation rate as: (number of queries with citation / total queries measured) x 100.

Record in a table: Query | ChatGPT rate | Gemini rate | Perplexity rate | Claude rate | Notes

1.2 Citation type breakdown

For queries where the client is cited, record how they appear:

  • Direct recommendation: AI names the client brand as a recommended tool or service provider

  • Source citation: AI cites client content as a source but does not recommend the brand directly

  • Comparison mention: AI mentions the client in a comparison context ("Brand X versus Brand Y")

  • Negative mention: Client is cited in a context that does not serve the brand (rare but worth flagging)

1.3 Competitor citation rates

Run the same queries for each of the 3-5 competitor brands. Record citation rates by engine. This produces the competitive visibility map - where the client sits relative to competitors on each AI engine and each query type.

Key question: For which queries are competitors being recommended while the client is absent? These are the highest-priority optimisation targets.

1.4 AI Overview presence (Google)

Check Google AI Overviews for the core keyword set. Record: which queries trigger an AI Overview, whether the client is cited in the Overview, and whether competitor content is cited. Use an incognito browser to minimise personalisation effects.

Section 2: Content Structure Analysis

AI engines cite content that directly answers questions. This section assesses whether the client's content is structured to be citable.

2.1 Heading structure audit

For each key page (product pages, service pages, and top-performing blog content):

  • Do H2 and H3 headings match the question format of target queries? ("What is X" rather than "About X")

  • Does each H2 section deliver a direct answer within the first 1-2 sentences beneath the heading?

  • Are headings specific enough to be citable? ("How to improve AI citation rates" is citable; "Tips for Better Results" is not)

2.2 Direct-answer paragraph audit

AI engines pull answer snippets. Check whether the content includes:

  • Definition paragraphs that state what something is in the first sentence

  • Process paragraphs that state how many steps or what the outcome is upfront

  • Comparison content that explicitly compares options (rather than describing each option separately)

  • Statistic and data claims with sources cited (AI engines favour content with cited evidence)

2.3 Conversational query coverage

Map each of the 10-20 target queries to existing content. For each query:

  • Is there a page or section that directly answers this query?

  • Does that page use the query's exact phrasing or close variants in headings?

  • Is the answer complete within the page, or does it require the user to navigate elsewhere?

Queries with no matching content are content creation opportunities. Queries with weak matches are content improvement opportunities.

2.4 Content freshness check

AI engines favour current content. Check:

  • Last-modified dates on key pages

  • Whether statistical claims, product features, or market data are current (not 2022-era)

  • Whether pages include explicit publication or update dates visible to crawlers

complete-ai-search-audit-template-run-aeo-audits-clients

Section 3: Schema Markup Review

Schema markup is the highest-impact technical lever for AI citations. This section is non-negotiable.

3.1 Existing schema inventory

Use Google's Rich Results Test or a schema validator to check which schema types are currently implemented across key pages. Record: page URL | schema types present | validation errors.

3.2 FAQPage schema audit

FAQPage schema is the single most impactful schema type for AI citation rates. Check:

  • Is FAQPage schema present on product pages, service pages, and key blog posts?

  • Do the FAQ questions in schema match the actual conversational queries users ask in AI engines?

  • Are FAQ answers complete and direct (answerable without visiting the page)?

  • Are FAQ questions genuinely distinct, or are they variations of the same question?

3.3 Article and Organization schema

Check for:

  • Article schema on all blog content (including datePublished, dateModified, author, publisher)

  • Organization schema at the site level (name, URL, logo, description, sameAs social profiles)

  • Person schema for named authors with expertise claims (relevant for EEAT signals)

  • Product or Service schema where applicable

3.4 Schema gap prioritisation

Rate schema gaps by impact: FAQPage gaps on high-traffic pages = Critical; Article schema gaps = High; Organization schema missing = High; other schema types = Medium.

Section 4: Technical Accessibility Audit

AI crawlers cannot cite content they cannot access. This section checks whether technical barriers are limiting citations.

4.1 AI crawler access

Review robots.txt for disallow rules that might block known AI crawlers. Check for explicit blocks of:

  • GPTBot (OpenAI)

  • PerplexityBot

  • ClaudeBot (Anthropic)

  • Google-Extended

  • Bingbot (used by some AI systems)

Many sites have broad disallow rules or wildcard blocks that inadvertently exclude these crawlers. Ensure the crawlers you want to allow are explicitly permitted.

4.2 JavaScript rendering

Check whether key content on product and service pages is rendered in JavaScript. Content that requires JavaScript execution to appear is less reliably indexed by AI crawlers than server-rendered or static HTML content. Use the View Source test: if the content does not appear in raw HTML source, it is JavaScript-rendered.

4.3 Page speed and Core Web Vitals

AI engine citation rates correlate with content quality signals, and page speed is one of them. Check Largest Contentful Paint, Cumulative Layout Shift, and Interaction to Next Paint for key pages using PageSpeed Insights. Flag pages below threshold (LCP above 4 seconds, CLS above 0.25) for remediation.

4.4 LLMs.txt presence

Check whether the site has an llms.txt file at the root URL. If not, flag as a low-effort, medium-priority recommendation. If yes, verify the file is current and includes key pages with accurate descriptions.

Section 5: Third-Party Signal Audit

AI engines draw on the broader web - review sites, news mentions, industry directories - not just the client's own content. This section assesses the off-site signal landscape.

5.1 Review platform presence

Check the client's presence and review volume on platforms that AI engines draw from: G2, Capterra, Trustpilot, Google Reviews, Clutch (for agencies), TripAdvisor (for hospitality), Healthgrades (for healthcare). Low review volume on major platforms is a citation rate suppressor.

5.2 Press and publication mentions

Search for brand mentions in industry publications, news sources, and recognised directories. AI engines weight mentions from authoritative sources more heavily. Assess: volume of mentions, quality of publication sources, recency.

5.3 Wikipedia and knowledge graph presence

Check whether the brand, founding team members, or key products have Wikipedia pages or Google Knowledge Panel entries. Established knowledge graph entries significantly increase AI citation likelihood for branded queries.

Section 6: Prioritised Action List

The audit findings feed into a prioritised action list. Use this priority framework:

Critical (fix within 2 weeks):

  • AI crawler blocked in robots.txt

  • Zero FAQPage schema on key commercial pages

  • Key content not accessible (JavaScript rendering blocking crawl)

High (fix within 30 days):

  • Organization schema missing

  • Key query types with no matching content

  • Article schema missing on blog posts

  • Competitor citation gap on high-intent queries

Medium (address within 60 days):

  • Heading structure not aligned with conversational query format

  • Content freshness issues on key pages

  • Low review platform presence

  • LLMs.txt missing

Low (ongoing improvement):

  • Content depth expansion on partially-covered topics

  • Press and publication mention building

  • Statistical data updates across existing content

Reporting the Audit Findings

A well-structured audit report includes: the baseline citation rate table with competitive context, a finding summary organised by section and priority, the prioritised action list with owner and timeline for each action, and the measurement plan - how and when citation rate progress will be tracked against the baseline.

The measurement plan is essential. Without pre-agreed measurement cadence, there is no way to demonstrate that the optimisation actions are working. Establish the baseline in the audit, define the next measurement point (typically 60 days post-implementation), and agree on what constitutes meaningful improvement.

For running the baseline measurement and ongoing tracking, AI Rank Lab's citation tracking across ChatGPT, Gemini, Perplexity, and Claude provides the automated measurement infrastructure that makes repeatable audits practical. Start with a free baseline measurement before running the rest of the audit sections.

Frequently Asked Questions

What is an AI search audit?
An AI search audit (also called an AEO audit) measures how visible a website or brand is in AI engine responses from ChatGPT, Gemini, Perplexity, and Claude, and identifies the technical, content, and structural factors limiting those citation rates. The audit establishes a baseline citation rate across target queries, benchmarks against competitors, and produces a prioritised action list of improvements ranked by expected impact on citation rate.
How long does an AI search audit take?
A full AI search audit typically takes 2-4 weeks from initiation to final report. The baseline citation rate measurement phase requires 2-4 weeks to gather statistically meaningful data (since AI responses are variable and need to be queried multiple times per keyword). The content, schema, and technical analysis phases can run in parallel with measurement and typically require 8-15 hours of work depending on site size. Report preparation adds another 4-6 hours.
What is the most important part of an AEO audit?
The baseline citation rate measurement is the most important element because it contextualises all other findings. Without knowing which queries are generating citations and which are not, it is impossible to prioritise technical or content fixes correctly. Within the technical analysis, FAQPage schema coverage is typically the highest-impact finding - the absence of FAQPage schema on key pages is the most common reason businesses have low citation rates despite strong content quality.
How do you measure AI citation rates for an audit?
AI citation rates are measured by querying each target keyword across each AI engine (ChatGPT, Gemini, Perplexity, Claude) multiple times and recording the percentage of responses that include a citation to the client's brand or content. Single-query spot-checks are unreliable because AI responses are generative and variable. Statistical citation rates require querying each keyword at minimum 3-5 times per engine. A dedicated AI visibility tool like AI Rank Lab automates this measurement at scale and runs weekly tracking to build trend data over time.
How often should an AI search audit be run?
A full AI search audit should be run at the start of an AEO engagement and annually thereafter. Between full audits, monthly citation rate monitoring (tracking the keyword set against the established baseline) identifies whether optimisations are working and flags any competitive changes. Quarterly schema and content reviews ensure that technical implementations remain current and new content follows AEO best practices. The audit establishes the baseline; ongoing monitoring tracks progress against it.
Free Consultation

Get a Free AI Ranking Consultation

Want to improve your brand's visibility in AI search engines like ChatGPT, Gemini, and Perplexity? Fill out the form and our experts will create a personalized strategy for you.

This form is protected by reCAPTCHA. Your data is handled securely and we'll never spam you.

Written by

Devanshu

AI Search Optimization Expert

Enjoyed this article?

Subscribe to our newsletter and get the latest AI search optimization insights delivered to your inbox.

No spam, unsubscribe at any time. We respect your privacy.