Updated: December 2025 | Reading time: 10 minutes
Crawl errors silently kill your rankings. Every broken link, server timeout, or blocked resource is a missed opportunity for Google to discover and index your content. The frustrating part? Most site owners don’t know they have a problem until their traffic drops.
I regularly audit client sites and find hundreds of crawl errors that have been quietly accumulating for months. The good news is they’re usually straightforward to fix once you know where to look.
This guide covers what crawl errors actually are, how to find them, and how to fix the most common issues—plus some automation tips to catch problems before they hurt your rankings.
Quick wins: Check these 3 things right now
- Open Google Search Console → Indexing → Pages → look for anything in red
- Check your
/robots.txtfile isn’t accidentally blocking important pages - Click “Why pages aren’t indexed” and note any 404s or server errors
If you find issues, keep reading for how to fix them.
What are crawl errors?
Crawl errors happen when search engine bots (like Googlebot) try to access your pages but can’t. The bot hits a dead end, records the failure, and moves on. If this keeps happening, Google may stop trying to crawl those URLs altogether.
There are two main types:
- Site-level errors — affect your entire site (DNS failures, server outages, robots.txt problems)
- URL-level errors — affect specific pages (404s, redirect loops, blocked resources)
Site-level errors are more urgent because they can stop Google crawling anything on your domain. URL-level errors are more common and tend to accumulate over time as content gets moved, deleted, or restructured.
Why crawl errors matter for SEO
Every site has a crawl budget—roughly how many pages Google will crawl in a given period. When Googlebot wastes time on broken links and error pages, it has less capacity for your actual content.
The impact compounds:
- Wasted crawl budget — bots spend time on dead ends instead of valuable pages
- Broken internal linking — link equity stops flowing through your site properly
- Poor user experience — visitors hitting 404s are likely to leave
- Delayed indexing — new content takes longer to appear in search results
- Ranking signals — Google interprets persistent errors as poor site maintenance
If you’re investing in SEO services, fixing crawl errors should be one of the first things on the list. It’s foundational work that makes everything else more effective.
How to find crawl errors
Google Search Console (free, essential)
Your primary tool.
Go to Indexing → Pages to see:
- Errors — critical issues blocking indexing
- Valid with warnings — indexed but with potential problems
- Valid — successfully indexed
- Excluded — intentionally not indexed (check these aren’t mistakes)
Also check Settings → Crawl stats for host availability issues, DNS problems, and robots.txt fetch failures.
Crawling tools
- Screaming Frog — industry standard, free up to 500 URLs
- Sitebulb — excellent visualisations and prioritised recommendations
- Ahrefs Site Audit — good if you already have a subscription
Run a full crawl monthly at minimum. For larger or frequently updated sites, weekly is better.
Common crawl errors and how to fix them
1. 404 errors (page not found)
What it means: The URL doesn’t exist. Either the page was deleted, moved without a redirect, or the link was wrong in the first place.
How to fix:
- If the content moved → set up a
301redirect to the new URL - If the content was deleted → redirect to the most relevant alternative page
- If it was never a real page → update or remove the broken links pointing to it
- Remove the URL from your XML sitemap if it’s listed there
Note: Not all 404s are bad. If you deliberately deleted a page and nothing links to it, a 404 is the correct response. Only fix 404s that have internal links pointing to them or that appear in your sitemap.
2. Soft 404s
What it means: The page returns a 200 (OK) status code, but Google thinks it looks like an error page—thin content, “page not found” text, or essentially empty.
How to fix:
- If the page should exist → add substantial, unique content
- If the page shouldn’t exist → return a proper
404or410status code - Check for template issues that might be serving empty pages
3. Server errors (5xx)
What it means: Your server failed to respond properly. Could be temporary (overload, timeout) or persistent (misconfiguration, hosting issues).
How to fix:
- Check your server logs to identify the cause
- Contact your hosting provider if errors persist
- Consider upgrading hosting if you’re hitting resource limits
- Implement caching and use a CDN to reduce server load
- Monitor uptime with tools like UptimeRobot or Pingdom
If you’re experiencing persistent server issues, it might be worth looking at your web hosting setup.
4. Redirect errors
What it means: Redirect chains (A → B → C → D), redirect loops (A → B → A), or redirects to broken pages.
How to fix:
- Update redirect chains to go directly to the final destination
- Fix loops by identifying and breaking the circular reference
- Use
301(permanent) redirects, not302(temporary), unless genuinely temporary - Use the Redirect Path Chrome extension to trace redirect chains
5. Blocked by robots.txt
What it means: Your robots.txt file is telling Google not to crawl certain URLs—sometimes unintentionally.
How to fix:
- Review your
/robots.txtfile for overly broadDisallowrules - Make sure you’re not blocking CSS, JS, or images that Google needs to render pages
- Use Search Console’s
robots.txttester to check specific URLs - Be precise with paths—watch for case sensitivity and trailing slashes
6. DNS errors
What it means: Google couldn’t resolve your domain name to an IP address.
How to fix:
- Check your domain’s DNS settings with your registrar
- Verify nameservers are correctly configured
- Ensure your DNS provider has good uptime (consider Cloudflare if not)
7. Crawled but not indexed
What it means: Google found the page but decided not to add it to the index. Often a content quality signal.
How to fix:
- Improve content quality and uniqueness
- Add internal links from other relevant pages
- Check for duplicate content issues
- Ensure the page has a clear purpose and isn’t thin
- Build some external links if the page is genuinely valuable
XML sitemap hygiene
Your sitemap should only contain URLs that:
- Return a
200status code - Are the canonical version (not duplicates)
- Aren’t blocked by
robots.txtornoindex - You actually want indexed
If you’re using Yoast or Rank Math, they handle this automatically. But still check periodically—plugins aren’t perfect, and site changes can introduce errors.
Submit your sitemap in both Google Search Console and Bing Webmaster Tools. Only ping search engines when the sitemap actually changes.
Mobile-first indexing considerations
Google primarily uses the mobile version of your site for indexing. This means:
- Mobile and desktop content should match—don’t hide content on mobile
- Internal links must work on mobile, not just desktop
- Structured data should be present on mobile pages
- Mobile page speed affects crawl efficiency
Test with Search Console’s mobile usability report and Google’s Mobile-Friendly Test.
A well-designed, responsive website handles most of these issues automatically.
Automating crawl error monitoring
Manual checks are fine for small sites, but if you’re managing multiple sites or a large site with frequent changes, automation saves time and catches issues faster.
Here’s a simple workflow I use with n8n:
- Pull crawl stats and index coverage via the Search Console API daily
- Compare against previous day to detect new errors
- Classify by severity (site-level errors = urgent, URL errors = weekly review)
- Send Slack alerts for critical issues, weekly email summary for everything else
- Log to a spreadsheet for trend tracking
You could also use Screaming Frog’s scheduling feature or Ahrefs’ automated site audits if you prefer not to build custom workflows. For businesses looking at more advanced setups, AI automation can take this even further.
Tools summary
- Google Search Console — essential, free, source of truth for how Google sees your site
- Screaming Frog / Sitebulb — detailed crawl analysis, find issues GSC doesn’t show
- Ahrefs / Semrush — broader SEO context, backlink analysis, competitor comparison
- Cloudflare / Bunny.net — CDN, performance, security, and better uptime
- n8n / Make — automation for monitoring and alerts
- UptimeRobot / Pingdom — server uptime monitoring
Maintenance schedule
Crawl health isn’t a one-time fix. Build these checks into your routine:
- Weekly: Quick check of Search Console for new errors
- Monthly: Full site crawl with Screaming Frog or Sitebulb
- Quarterly: Deep audit including redirect chains, orphan pages, and crawl budget analysis
- After any major site change: Immediate crawl to catch issues early
The bottom line
Crawl errors are one of the most fixable SEO problems. Unlike content quality or backlinks, which take time to improve, most crawl errors can be resolved in an afternoon with the right tools.
Start with Search Console. Fix the critical errors first—server issues and site-level problems. Then work through the URL-level errors systematically. Set up some basic monitoring so you catch new issues quickly.
Small, consistent fixes compound. A site with clean crawl health gets indexed faster, ranks more reliably, and gives you one less thing to worry about.
Need help fixing crawl errors on your site? Get in touch for a technical SEO audit and we’ll identify exactly what needs fixing.
