• Skip to primary navigation
  • Skip to main content

WebStuff

  • Home
  • Web Dev
  • SEO
  • AI Consulting
  • Brand Management
  • Publishing
  • About
  • Contact

How Google Crawling and Rendering Works in 2026

April 5, 2026 by Joe Davis

How the web is crawled

Google has provided updated insight into how its crawling and rendering systems function, with new details shared by Gary Illyes. The explanation focuses on how Googlebot operates, how much data it processes, and how pages are rendered and indexed.

More information on Google’s crawlers and user agents is available from the official Google documentation here:  https://developers.google.com/crawling/docs/crawlers-fetchers/overview-google-crawlers

Googlebot Is Not a Single Crawler

  • Googlebot is not one single crawler.
  • Google operates multiple crawlers, each designed for different purposes.
  • These crawlers use different user agents and are documented publicly.
  • Referring to “Googlebot” as one entity is no longer fully accurate.

Crawl Size Limits (Critical Technical Details)

Google enforces strict byte limits on how much of a page or resource it will process:

  • HTML pages:
    • Googlebot fetches up to 2MB per URL
    • This includes HTTP headers + HTML content
  • PDF files:
    • Limit is 64MB
  • Other file types (default):
    • Limit is 15MB
  • Images and videos:
    • Limits vary depending on the specific Google product using them

What Happens When a Page Exceeds 2MB

If a page is larger than 2MB, Google does not reject it—but it does not process it fully either.

Step-by-step behavior:

  • Partial fetching:
    • Googlebot stops downloading the page exactly at the 2MB limit
  • Processing the cutoff:
    • Only the first 2MB is sent to:
      • Indexing systems
      • Web Rendering Service (WRS)
  • Ignored content:
    • Any content beyond 2MB:
      • Is not fetched
      • Is not rendered
      • Is not indexed

How Resources Are Handled

  • External resources referenced in HTML (like CSS and JavaScript):
    • Are fetched separately
    • Have their own individual byte limits
    • Do not count toward the 2MB HTML limit
  • Exceptions:
    • Images, videos, fonts, and some uncommon file types may not be fetched by the renderer

How Google Renders Pages (WRS)

Google uses the Web Rendering Service (WRS) to process pages after crawling.

What WRS does:

  • Executes JavaScript (like a modern browser)
  • Processes CSS
  • Handles XHR (AJAX) requests
  • Determines the final visual and textual state of the page

Important constraints:

  • Each fetched resource (JS, CSS, etc.) is also subject to the same 2MB limit
  • WRS:
    • Does not request images or videos
    • Focuses on understanding content and structure

Key SEO Implications

1. HTML Size Matters

  • Only the first 2MB of HTML is considered
  • Anything beyond that is effectively invisible to Google

2. Content Placement Is Critical

  • Important elements must appear early in the HTML:
    • <title>
    • Meta tags
    • Canonical tags
    • <link> elements
    • Structured data
  • If these appear after 2MB:
    • Google will never see them

3. External Files Are Safer

  • Moving CSS and JavaScript to external files:
    • Prevents bloating HTML
    • Allows Google to fetch them independently

4. Rendering Still Has Limits

  • Even external JS/CSS files:
    • Must stay within their own 2MB limit
  • Heavy scripts can still cause issues

5. Server Performance Affects Crawling

  • If your server is slow:
    • Google will reduce crawl frequency
  • Google automatically backs off to avoid overloading servers

Best Practices from Google

  • Keep HTML lean
    • Avoid embedding large scripts or styles directly in HTML
  • Prioritize important content early
    • Place critical SEO elements near the top of the document
  • Use external resources
    • Separate CSS and JavaScript from HTML
  • Monitor server logs
    • Ensure fast response times
    • Identify crawl issues

The Bottom Line is…

Google’s crawling system in 2026 is highly structured and constrained by strict byte limits. The most important takeaway is that Google only processes the first 2MB of your HTML, and anything beyond that is ignored entirely.

This makes:

  • Page structure
  • Content ordering
  • File optimization

…more important than ever for SEO and indexation.

Watch the Search Off the Record podcast for more details:

Filed Under: Crawling, SEO

Copyright © 1995 - 2026 All Rights Reserved WebStuff ® | Privacy Policy | Text Converter

  • Home
  • Web Dev
  • SEO
  • AI Consulting
  • Brand Management
  • Publishing
  • Text Converter
  • Articles
  • About