Google Crawl Limit Checker
Googlebot stops parsing HTML after 2MB. Check if your page exceeds this limit and identify what's consuming the most space.
About This Tool
Googlebot has a technical limit when crawling web pages: it stops parsing HTML after 2MB (2,097,152 bytes). This means if your page's HTML exceeds this limit, parts of your content won't be indexed by Google, potentially affecting your SEO performance.
How It Works
- Fetches your page's HTML as Googlebot sees it (before JavaScript execution)
- Calculates the exact byte size using UTF-8 encoding
- Analyzes what's consuming space: inline scripts, styles, base64 images, SVG graphics, etc.
- Identifies the top 10 largest HTML elements with specific CSS selectors
- Generates actionable recommendations ranked by priority
- Shows external resources (scripts, CSS, images) that DON'T count toward the limit
What Counts Toward the 2MB Limit?
Counts Toward Limit
- • Inline JavaScript (script tags without src)
- • Inline CSS (style tags and style attributes)
- • Base64-encoded images in HTML/CSS
- • Inline SVG graphics
- • All text content and HTML structure
- • HTML comments and whitespace
Does NOT Count
- • External JavaScript files (script src="...")
- • External CSS files (link rel="stylesheet")
- • External images (img src="...")
- • External fonts, videos, and other media
- • Content loaded via AJAX/fetch
- • Content rendered by JavaScript
Pro Tip: The best way to stay under the limit is to externalize large inline scripts and styles into separate .js and .css files. This also improves caching and page load performance!
Frequently Asked Questions
Google introduced this limit to prevent extremely large HTML files from consuming excessive resources during crawling. Pages larger than 2MB are rare and usually indicate optimization opportunities. Googlebot will still index the first 2MB, but content after that point may not be indexed.
Googlebot will stop parsing your HTML after 2MB, which means important content, links, or structured data after that point won't be seen or indexed. This can significantly impact your SEO. The tool identifies what's consuming space so you can optimize accordingly.
No! External JavaScript files, CSS stylesheets, and images loaded via URLs do NOT count toward the 2MB HTML limit. Only the raw HTML content itself counts. This is why externalizing inline scripts and styles is an effective optimization strategy.
Very accurate! The tool fetches your page exactly as Googlebot sees it and calculates the raw HTML size using UTF-8 encoding (the same method Google uses). It analyzes the HTML before JavaScript execution, matching Google's initial crawl behavior.
Follow the recommendations provided by the tool: externalize inline JavaScript and CSS, convert base64 images to files, optimize or remove large SVGs, remove HTML comments, and minify your HTML. These optimizations also improve page load performance!
Yes! After analyzing your page, you can export all results to CSV format including HTML size, breakdown by category, top space consumers, and recommendations. This is useful for tracking progress over time or sharing with your development team.
reCAPTCHA helps prevent automated abuse and ensures the tool is used by real users. Fetching and analyzing web pages requires server resources, so this protection maintains service quality for everyone.