Outline Technologies — SEO, AEO & GEO Agency
Back to Blog
Technical SEO2026 Checklist

Technical SEO Checklist 2026: 50-Point Audit for AI-Era Search

The complete technical SEO checklist for 2026. Covers Core Web Vitals (INP replaced FID in March 2024), AI crawler accessibility, schema markup, crawl budget, canonicals, and every technical signal that matters for both Google and AI search visibility.

Ahmed Shanti 18 min readMay 13, 2026
Checklist Sections
Core Web Vitals (2026 metrics)Crawlability and indexabilityAI crawler accessURL structure and canonicalsStructured data and schemaSite architecture and linksPage speed and performanceMobile and HTTPSSitemap and robots.txtMonitoring and maintenance

How to Use This Checklist

This is a working audit checklist. Each item is a binary check: it either passes or it needs fixing. Run through the full list on a new site or when preparing for a technical audit. For ongoing maintenance, focus on the monitoring section.

Two important notes before you start: First, INP replaced FID as a Core Web Vital in March 2024. If your audit tools still reference FID, they're outdated. Second, AI crawler accessibility is now a first-class technical concern — many sites inadvertently block the crawlers that determine AI citation eligibility.

Core Web Vitals 2026 Thresholds

LCP
≤ 2.5s
Largest Contentful Paint
INP
≤ 200ms
Interaction to Next Paint
CLS
≤ 0.1
Cumulative Layout Shift

INP replaced FID (First Input Delay) as a ranking signal on March 12, 2024. INP measures the full latency of interactions, not just the first one. FID is deprecated and no longer a Core Web Vital.

Section 1: Core Web Vitals

LCP ≤ 2.5 seconds on mobile (measured in CrUX field data)Critical
INP ≤ 200ms across key user interaction pagesCritical
CLS ≤ 0.1 — no significant layout shifts during page loadCritical
LCP element identified (use PageSpeed Insights to confirm which element)
LCP image is preloaded via <link rel='preload'> for above-fold images
No render-blocking resources delaying LCP (check Lighthouse)
JavaScript execution time under 2 seconds for INP compliance
CrUX data shows 75th percentile in Good range (not just lab test)Critical

Section 2: Crawlability and Indexability

robots.txt accessible at /robots.txt and returning 200 statusCritical
No unintended Disallow rules blocking key pages or sectionsCritical
XML sitemap submitted in Google Search ConsoleCritical
Sitemap contains only indexable, canonical URLs (no noindex pages)Critical
Google Search Console shows no crawl errors for key pagesCritical
All key pages return 200 status (check for soft 404s)Critical
Pagination handled correctly (rel=canonical or no-index on pagination)
Orphan pages addressed — all important pages linked from sitemap or internal links

Section 3: AI Crawler Accessibility

This section is new in 2026 — and it directly determines whether AI systems like Perplexity, ChatGPT Browse, and Copilot can index and cite your content.

Known AI Crawlers (Allow by default)

GPTBot(OpenAI training)
OAI-SearchBot(ChatGPT Browse retrieval)
PerplexityBot(Perplexity indexing + retrieval)
ClaudeBot(Anthropic training)
Google-Extended(Google AI training)
Applebot-Extended(Apple AI training)
PerplexityBot allowed in robots.txt for all content you want citedCritical
GPTBot not blocked (if you want to appear in ChatGPT)Critical
OAI-SearchBot not blocked (controls ChatGPT Browse retrieval)Critical
No broad User-agent: * Disallow rules that accidentally catch AI botsCritical
Google-Extended not blocked (unless you have specific reasons)
Key content pages render server-side (not JS-only) so AI crawlers can extract textCritical
No login wall or cookie consent gate blocking content extraction
AI crawler access tested manually (fetch URL as Googlebot in GSC, check renders)

Section 4: URL Structure and Canonicals

Every page has one canonical URL — no duplicate content without canonical tagsCritical
self-referencing canonical on every page (rel=canonical points to itself)Critical
www vs non-www redirect in place and consistentCritical
HTTP redirects to HTTPS for all pagesCritical
Trailing slash policy consistent sitewide (always trailing slash OR never — not mixed)
URLs use lowercase letters only (no case sensitivity issues)
No URL parameters creating duplicate pages (handle with canonical or parameter exclusion in GSC)

Section 5: Structured Data and Schema

Organization schema with @id, name, url, logo, description, sameAs on every pageCritical
Article/BlogPosting schema on all blog posts (author, datePublished, dateModified)Critical
FAQ schema on pages with Q&A content
HowTo schema on process/step-by-step content
BreadcrumbList schema for site navigation hierarchy
Product schema on product pages (name, price, availability)
All schema validated in Google Rich Results Test with no errorsCritical
Schema.org validator shows no structural issues
Google Search Console rich results report shows no critical errorsCritical

Section 6: Page Speed and Performance

Time to First Byte (TTFB) under 600msCritical
Total page weight under 2MB on mobile
Images in WebP or AVIF format (not PNG/JPEG for modern browsers)
Images have explicit width and height attributes (prevents CLS)Critical
Critical CSS inlined; non-critical CSS deferred
JavaScript deferred or split into code bundles (no render-blocking scripts)
CDN configured for static assets
Caching headers set correctly (long cache for static assets)

Section 7: Mobile and HTTPS

Google Mobile-Friendly Test passes (developers.google.com/search/docs/crawling-indexing/mobile/mobile-sites-mobile-first-indexing)Critical
Viewport meta tag present on all pagesCritical
No horizontal scroll on mobile (320px–430px viewport)Critical
Touch targets (buttons, links) minimum 44x44px
SSL certificate valid and not expiring within 30 daysCritical
HTTPS enforced sitewide with proper redirectsCritical
HSTS header configured
No mixed content warnings (HTTP assets on HTTPS pages)Critical

Section 8: On-Page Technical Elements

Every page has a unique, descriptive title tag (50–60 characters)Critical
Every page has a unique meta description (150–160 characters)Critical
One H1 per page — matches or closely reflects the title tagCritical
Open Graph tags on all pages (og:title, og:description, og:image)
Twitter Card tags present
Hreflang tags correctly implemented for multilingual sites
No duplicate title tags across the site (check with Screaming Frog)Critical
Image alt text present and descriptive for all non-decorative images

Section 9: Internal Linking and Architecture

No broken internal links (404s for internal hrefs)Critical
Key pages reachable within 3 clicks from the homepage
Pillar pages link to relevant cluster content, and vice versa
No orphan pages — every page has at least one internal link pointing to it
Internal links use descriptive anchor text (not 'click here')
Redirects reviewed — no redirect chains longer than 2 hops

Section 10: Monitoring and Maintenance

Google Search Console connected and verifiedCritical
Bing Webmaster Tools connected and verified
Core Web Vitals monitored monthly in GSC → Core Web Vitals reportCritical
Uptime monitoring in place (alerts for site downtime)Critical
Screaming Frog or similar crawler scheduled monthly
GSC manual action and security issue reports reviewed weeklyCritical
IndexNow implemented for instant URL submission on new content
AI citation tracking configured (GA4 custom channel group for AI referrers)

Prioritization: Fix Critical Items First

Not every checklist item has equal impact. The items marked Critical — Core Web Vitals failures, blocked crawlers, missing canonicals, broken indexation — can directly cause ranking drops or prevent AI citation entirely. Fix those first. The non-critical items are improvements and optimizations that compound over time.

The Most Common Technical SEO Failures in 2026

INP failures from heavy JavaScript. Many sites still haven't addressed INP since it replaced FID in March 2024. INP failures are almost always caused by long JavaScript tasks blocking the main thread. Profile with Chrome DevTools Performance tab to find the culprit.
Accidentally blocking AI crawlers. A common pattern: security-conscious developers add broad robots.txt rules to block scrapers, which also block PerplexityBot and OAI-SearchBot. This removes the site from AI citation eligibility entirely.
JavaScript-only content. AI crawlers don't execute JavaScript. Content that only renders after JS execution is invisible to AI crawlers, regardless of your robots.txt permissions.
Duplicate content without canonical control. URL parameters, session IDs, and sorting variations create hundreds of near-duplicate pages. Crawl budget is wasted, and Google may index the wrong version.
Missing or stale sitemap. Sites that add new content without updating their sitemap can take weeks to get new pages indexed. Implement programmatic sitemap generation that auto-updates with new content.
AS
Written by Ahmed Shanti
Co-Founder, Outline Technologies

Ahmed has run technical SEO audits for hundreds of sites. He updates this checklist quarterly as search engine requirements evolve — the AI crawler section was added in Q1 2025 when it became clear that AI indexation was distinct from traditional SEO crawlability.