Google's March 2026 core update changed the rules. Core Web Vitals are now evaluated holistically across entire sites, not on a per-page basis - a handful of slow-loading templates or high-CLS ad layouts can suppress rankings for your entire domain.
Early data from Ahrefs, Semrush, and independent tracking tools shows that affected sites experienced traffic declines of 20-35%, with some domains losing over 50% on their worst-hit sections.
Meanwhile, Googlebot's file size limit dropped from 15MB to 2MB. HTTP request headers count toward that limit. External resources like CSS and JavaScript get their own separate byte counters. When Googlebot hits 2MB, it stops fetching and passes the truncated content to indexing as if it were the complete file. If your most important structured data sits at the bottom of a bloated HTML document, Google may never see it. Technical SEO is no longer the thing you fix once and forget. It's never finished. Algorithms evolve, sites grow, and new issues emerge constantly. The websites that maintain their rankings long-term are those that treat technical SEO as an ongoing discipline - not a one-time project. This checklist reflects what actually matters right now - grounded in the latest documentation changes, algorithm updates, and practitioner-tested fixes.
Crawlability: Making Sure Google Can Reach Your Pages
Every technical SEO audit starts here. Crawling and indexing are the first steps search engines take to understand your website. If these foundations fail, your content will not appear in search results - regardless of how valuable it is.
Robots.txt as a Governance Document
Your robots.txt file is the first thing any crawler reads. It tells crawlers which pages to access and which to ignore. Mistakes here - like accidentally blocking CSS files or key landing pages - can have severe consequences.
But in 2026, robots.txt carries weight beyond Googlebot. It serves as a governance document that dictates which AI companies can access your proprietary data for training versus which can access it for real-time search retrieval. It is helpful to distinguish between Training Bots (which scrape your content to train models like GPT-5) and Retrieval Bots (which fetch your content to answer user questions in real-time).
Here's what to check:
- Audit your robots.txt with Google Search Console's tester. Confirm you aren't blocking CSS, JS, or critical landing pages.
- Add explicit rules for AI crawlers.
Block GPTBot and ChatGPT's training data never learns about your new products. Block ChatGPT-User, and you disappear from live AI searches entirely. Know the difference. - Consider emerging standards. Two emerging standards let you express granular preferences about how AI systems use your content without blocking crawlers entirely. This is crucial for AI Search Optimization: you want to be indexed for search but may want to opt-out of model training.
- Remember the fundamental limitation.
Robots.txt only prevents crawling, not indexing. If other pages link to a URL blocked via robots.txt, Google can still index it - just without knowing the content. The result is a search result without a snippet. For real indexing control, you need the noindex tag.
A common misconception involves Google-Extended. Google offers Google-Extended as a training-specific opt-out, but it covers model training only, not AI Overviews inference. Blocking Google-Extended signals you do not want your content used in training; it does not prevent your content from appearing in AI-generated search summaries. This is the most common misconception in the governance space. Many site owners have added Google-Extended disallow directives believing they have opted out of AI use. They have opted out of training data collection. They have not opted out of inference-time retrieval that powers AI Overviews.
XML Sitemaps and URL Discovery
Your sitemap is the roadmap you hand to crawlers. An XML sitemap acts as a roadmap, guiding search engines to your most important pages. Ensure your sitemap includes only canonical, indexable URLs - exclude 404 pages, redirect targets, and noindex pages.
Only include indexable, canonical URLs - no pages with noindex, no duplicates, no redirect targets. The lastmod date should only be updated when there are actual content changes, because Google learns whether you're "lying" and then completely ignores your lastmod data. This is a detail most practitioners miss. If your CMS auto-updates lastmod timestamps on every deploy, Google eventually distrusts your sitemap's freshness signals entirely. For active sites - blogs publishing daily, e-commerce stores rotating inventory - static sitemaps are no longer enough. For blogs and ecommerce, update sitemaps daily or in real time.
Crawl Budget: Who Gets It and Why It Matters
If your site has more than 10,000 URLs or generates new content faster than Google indexes it, crawl budget optimization is essential. Smaller sites can skip this section. Everyone else should pay close attention.
Common 2025–2026 challenges for Google include faceted navigation (often ~50% of crawling issues), action parameters, and irrelevant URL parameters. This is the single biggest crawl budget drain on e-commerce sites. E-commerce is where crawl budget optimization delivers its biggest returns. Faceted navigation is the main culprit - tens of thousands of low-value filter URL combinations quietly absorbing crawl activity while core categories and products wait in the queue.
The fix sequence: 1. Pull server logs to see where Googlebot actually spends its time. Server log analysis provides the most comprehensive view of how search engines interact with your website. Unlike Google Search Console data, which shows only a subset of crawling activity, server logs capture every single request made to your site.
-
Block zero-demand parameter URLs via robots.txt - internal search results, sorting parameters, tracking parameters, and cart pages. 3. Collapse redirect chains to a single hop. Collapse multi-hop redirects to a single step.
-
Return 410 status codes for permanently deleted content instead of soft 404s. 5. Monitor GSC Crawl Stats weekly to track whether changes improve crawl distribution.
For sites under 50,000 URLs, Screaming Frog Log File Analyser is the easiest entry point. Above 100,000 URLs, dedicated platforms like Botify, JetOctopus, and OnCrawl handle the scale better.
And don't forget the AI crawler load. Beyond Googlebot, you now have AI crawlers competing for your server resources. Bots like GPTBot, ClaudeBot, and PerplexityBot are requesting pages from websites at scale. Nearly 21% of the top 1,000 websites already have specific rules for AI crawlers in their robots.txt files. These bots consume bandwidth and compete with Googlebot for server capacity.
The 2MB Crawl Limit: HTML Hygiene Is Non-Negotiable
This is the technical change that caught many teams off guard. Googlebot currently fetches up to 2MB for any individual URL (excluding PDFs). This means it crawls only the first 2MB of a resource, including the HTTP header. For PDF files, the limit is 64MB.
When Googlebot crawls a page, it focuses only on the portion it successfully fetches. If your HTML content goes beyond the 2MB limit, everything after that point is completely ignored. The fetched portion is then treated as the full page and sent to Google's indexing systems.
The 2025 Web Almanac reports a median mobile homepage size of 2,362 KB. That means the median page is already pushing against this limit before you add inline JavaScript or bloated schema markup. What to do: - Keep your HTML lean: move heavy CSS and JavaScript to external files.
-
Important elements such as meta tags, titles, canonical links, and structured data should be positioned higher in the HTML. This ensures they are included within the portion Google actually processes.
-
Use the testing tool released in February 2026 that allows SEO professionals to simulate the limit on their own pages.
-
Audit pages with Screaming Frog's page size filter - any HTML document approaching 1.5MB needs intervention.
Indexing: From Crawled to Searchable
Getting crawled is step one. Getting indexed is where the real competition begins. Emphasis on quality means getting indexed is itself a stronger quality signal than before.
Status Codes and Error Handling
The most urgent shift is the December 2025 Rendering Update from Google. The search engine explicitly clarified that pages returning non-200 HTTP status codes (like 404 or 5xx errors) may be excluded from the rendering pipeline entirely. This means if your site relies on client-side JavaScript to display user-friendly error messages or "Recommended Products" on a 404 page, Googlebot may never see that content.
Use the right status codes intentionally:
- 200 for valid, indexable pages
- 301 for permanent redirects (one hop, not chains)
- 410 for intentionally removed content (signals permanence more strongly than 404)
- 503 for temporary maintenance (tells Googlebot to come back later)
Canonical Tags and Duplicate Content
Set correct canonical tags on all templates. Point variants and parameters back to the preferred URL. Use self-referencing canonicals on unique pages to prevent signal splitting.
Common pitfalls: your staging site leaking into the index, www and non-www variants both accessible, HTTP and HTTPS both resolving, trailing slashes creating duplicate URLs. Each of these doubles your crawlable URL count while splitting authority.
IndexNow: Push-Based Indexing for Non-Google Engines
Google does NOT support IndexNow. But Bing, Yandex, Naver, and Seznam do - and with 5+ billion daily submissions, instant indexing is critical for AI search discovery in 2026.
Over 80 million websites now actively use IndexNow for instant content discovery. Bing reports that 22% of all clicked URLs in search results originated from IndexNow submissions. Given that ChatGPT search relies partly on Bing's index, implementing IndexNow gives you indirect visibility in AI-powered search.
Most modern CDNs (Cloudflare, Akamai) have a "one-click" IndexNow integration. There's no reason not to enable it. For Google specifically, Google maintains its own crawling infrastructure and has a separate Google Indexing API limited to job postings and live streaming content. For Google indexing, traditional methods like XML sitemaps and Google Search Console remain necessary.
Core Web Vitals: The Site-Wide Performance Bar
The three metrics haven't changed. Largest Contentful Paint (LCP) measures loading performance, Interaction to Next Paint (INP) evaluates responsiveness, and Cumulative Layout Shift (CLS) tracks visual stability. What changed is how Google uses them.
Holistic Scoring Changes Everything
Until this update, Google evaluated Core Web Vitals on a per-URL basis. Each page was assessed individually for LCP, INP, and CLS. A page that met all three thresholds received the ranking benefit regardless of how the rest of the domain performed. This created a common optimization strategy: fix your top 50 landing pages and ignore everything else. That strategy no longer works.
Under holistic CWV scoring, Google aggregates performance data across your entire domain to produce a site-level assessment. What changed is how Google uses the data - from individual page assessment to aggregate site-level evaluation.
This means your neglected blog archive, those old landing pages with unoptimized hero images, or that heavyweight checkout process - they all drag down your best-performing pages.
INP: The Metric Most Sites Still Fail
The biggest shift in Core Web Vitals happened on March 12, 2024, when Google officially retired FID and replaced it with INP. As we move through 2026, INP is now fully embedded as a ranking signal - and many sites are still struggling to meet its thresholds.
INP looks at the full interaction journey, including processing and the next visual update. That difference sounds technical, but users feel it instantly. Your INP score should be 200 milliseconds or less. Above 500ms is considered poor. Common INP offenders: heavy JavaScript frameworks, unoptimized third-party scripts (especially analytics and consent banners), complex DOM structures, and filter-heavy listing pages. Core Web Vitals optimization in 2026 is less about chasing a pretty Lighthouse screenshot and more about making real interactions smoother. "The page looks fast, so why is INP still poor?" Because visual load and interactive responsiveness are not the same thing.
LCP and CLS Quick Wins
For LCP, the target stays at 2.5 seconds - though many practitioners now aim for sub-2 seconds to stay competitive. TTFB under 200ms is the gold standard in 2026. Achieve this through edge computing, multi-layer caching, and lightweight server-rendered HTML. For CLS, the target is 0.1 or less. Set explicit width and height attributes on every image and video element. Preload fonts. Reserve space for ad units before they render.
Google Search Console updates Core Web Vitals data based on 28 days of real user data. After implementing optimizations, expect to see metric improvements within 4-6 weeks. Ranking improvements may take longer. Most sites see measurable ranking changes within 2-3 months of sustained good scores.
JavaScript Rendering and the Two-Audience Problem
Googlebot's JavaScript rendering capability has improved substantially. Google processes JavaScript-heavy web pages in three phases: crawling, rendering, and indexing. In March 2026, Google even removed its longstanding "Design for Accessibility" JavaScript warning because Googlebot has rendered JavaScript using an evergreen Chromium engine since 2019 and processes JavaScript reliably.
But here's what the top-ranking articles routinely miss: Googlebot and AI crawlers are fundamentally different.
Googlebot has been developed over the years to be much better at handling JavaScript than the newer LLM bots. However, it's really important to understand that the LLM bots are not trying to crawl and render the web in the same way as Googlebot. Don't assume that they will ever try to mimic Googlebot's behavior.
Unlike Googlebot, most AI crawlers do not render JavaScript at all. They fetch the raw HTML, extract whatever text is available, and use that to generate answers. If your content only exists after JS execution, these crawlers see an empty page and your site won't appear in AI-generated responses.
This creates a two-audience problem. Your rendering strategy must satisfy both:
- For Googlebot: Ensure the DOM loads fully on first request. Use the URL Inspection Tool in Google Search Console to verify what Googlebot actually renders.
- For AI crawlers:
Check if your page loads all the pertinent information in the DOM on the first load. For the LLM bots, to be very sure the content is available to them, check your static HTML.
The practical solution: SSR, SSG, ISR, and hybrid server-first rendering are usually better long-term answers than relying on a permanent bot-only rendering layer. Server-side rendering ensures both Googlebot and AI crawlers see complete content on the first request - no rendering queue, no delays, no empty shells.
Structured Data: Entity Signals for Search and AI
Structured data helps search engines understand what your content is about, not just what it says. In 2026, this matters for traditional search and AI search alike.
What to Implement (and What to Skip)
In November 2025, Google announced it would deprecate support for seven structured data types starting January 2026. Google is phasing out support for lesser-used schema types like Practice Problem, Dataset, and Sitelinks Search Box. Don't panic about deprecations - core schema types like Product, Article, Organization, Person, and Review remain fully supported and prioritized.
Focus your schema strategy on two objectives: 1. Rich result eligibility. 31 schema types retain active rich result support as of March 2026. The types with the strongest performance are those tied to specific user intent signals - product availability, event timing, recipe details, and local business information.
2. Entity disambiguation for AI. The single highest-leverage schema implementation available in 2026 is entity markup that identifies your organization as a known, verified entity in Google's Knowledge Graph. This implementation is often absent from SEO schema strategies because it does not produce visible SERP features, but its impact on AI Mode citation and Knowledge Panel accuracy is substantial.
Implementation Best Practices
-
JSON-LD is Google's preferred method and the one most SEO professionals use today. It's clean, flexible, and doesn't interfere with your visible page content.
-
Schema must accurately reflect what is visible on the page - not exaggerate or invent details.
-
The knowsAbout property is the second most impactful entity markup change post-March 2026. Specifying the topics your organization genuinely has expertise in creates a topical authority signal that AI Mode uses when selecting sources. An organization schema that declares knowsAbout SEO, content marketing, and analytics is more likely to be cited for queries in those domains.
-
Validate with Google's Rich Results Test and the Schema Markup Validator. Test in Search Console after deployment.
JSON-LD delivered in the document head remains the recommended delivery method. Given the 2MB crawl limit, placement matters more than ever.
Site Architecture and Internal Linking
A logical, well-structured site architecture ensures that both users and bots can find the most important content quickly. In 2026, this is more critical than ever. AI-powered search relies on clear signals of context and hierarchy, and messy site structures can prevent your most valuable pages from being surfaced in generative results.
Keep critical pages within three clicks of the homepage. Pages that are nested too deep (users need more than 6 clicks to reach them) will receive less link equity from your homepage, which may affect their rankings.
Internal linking serves three functions: it distributes PageRank, establishes topical relationships between pages, and guides crawlers toward your priorities. In the worst-case scenario, pages that you want to rank won't have any internal links. These pages are called orphan pages. Web crawlers have limited ability to access those pages, and there is no link equity flowing to them.
Your audit should flag:
- Orphan pages - crawl your site with Screaming Frog and cross-reference against your sitemap
- Deep pages - anything beyond click depth 4 should be reconsidered
- Broken internal links -
69.32% of sites have pages with no inbound links
- Redirect chains within internal links - update the link target, don't just add another redirect
Mobile-First, Multi-Device, and Security Basics
Mobile-first indexing is no longer new; it's the standard. But in 2026, the conversation has shifted from "does your site work on mobile?" to "is your site optimized for all device experiences?"
Don't just test with browser emulators - use real devices. A page that looks perfect on Chrome DevTools might break on Safari or a mid-range Android. Regular multi-device QA is now a must-have part of every technical SEO audit.
Security is foundational, not optional:
- HTTPS across every URL, with proper 301 redirects from HTTP
- Auto-renewing SSL certificates to prevent outages
- No mixed content (secure pages loading insecure resources)
- Security headers configured (HSTS, X-Content-Type-Options, X-Frame-Options)
Only 47% of sites reach Google's "good" thresholds in 2026. That's a staggering gap - and it means genuine competitive advantage awaits anyone who closes it.
The Audit Cadence That Actually Works
One-time audits create temporary fixes. Sustained visibility requires rhythm.
Perform a full sitewide technical SEO audit quarterly. Conduct a content audit every six to twelve months. Perform a backlink audit quarterly or whenever traffic dips unexpectedly.
Monthly: Check GSC for crawl errors, 404s, and Core Web Vitals regressions. Review your top 20 keyword positions. Monitor server response times. Quarterly: Run a full Screaming Frog crawl. Analyze server logs for crawl waste. Validate structured data across templates. Test rendering on real mobile devices. Review robots.txt against the latest AI crawler user-agents. After every major site change: Migration, redesign, CMS update, new JavaScript framework - each one warrants a full technical check within 48 hours.
An audit is not a one-time thing you complete and forget. It's an ongoing practice of monitoring, analyzing, and improving your site's technical health.
--- Technical SEO in 2026 operates on a simple but unforgiving principle: it is no longer about "tricking" a crawler; it is about engineering a transparent relationship with the machines that power discovery. The site that loads fast, renders completely, communicates its structure through schema, and governs its crawler access intentionally - that site wins in traditional search and gets cited by AI systems. The gap between sites that treat technical SEO as infrastructure and those that treat it as an afterthought will only widen. Google's message is clear: in 2026, good rankings require original content, demonstrated expertise, and a technically flawless site. This checklist gives you the sequence. The execution is yours.
Ready to optimize for the AI era?
Get a free AEO audit and discover how your brand shows up in AI-powered search.
Get Your Free Audit