Google’s Search Advocate, John Mueller, provided insights on diagnosing widespread crawling issues.
This advice was given in response to a disruption reported by Adrian Schmidt on LinkedIn. Google’s crawler halted access to several of his domains at once.
Despite the interruption, Schmidt noted that live tests via Search Console continued to function without error messages.
Investigations showed no increase in 5xx errors or issues with robots.txt requests.
What could the problem be?
Mueller’s Response
Addressing the situation, Mueller suggested that shared infrastructure might be the cause:
"If it shared across a bunch of domains and focuses on something like crawling, it’s probably an issue with a shared piece of infrastructure. If it’s already recovering, at least it’s not urgent anymore and you have a bit of time to poke at recent changes / infrastructure logs."
Infrastructure Investigation
All affected sites used Cloudflare as their CDN, which raised some questions.
When asked about debugging, Mueller recommended checking Search Console data to determine if DNS or failed requests were the problem.
Mueller stated:
"The crawl stats in Search Console will also show a bit more, perhaps help decide between say DNS vs requests failing."
He also highlighted that timing was a key clue:
"If it’s all at exactly the same time, it wouldn’t be robots.txt, and probably not DNS."
Impact on Search Results
Regarding concerns about search visibility, Mueller reassured this type of disruption wouldn’t cause any problems:
"If this is from today, and it just lasted a few hours, I wouldn’t expect any visible issues in search."
Why This Matters
When Googlebot suddenly stops crawling across numerous sites simultaneously, identifying the root cause can be challenging.
While temporary crawling pauses might not immediately impact search rankings, they can disrupt Google’s ability to discover and index new content.
The incident highlights a vulnerability organizations might face without realizing it, especially those relying on shared infrastructure.
How This Can Help You
If Googlebot stops crawling your sites:
- Check if the problem affects multiple sites at once
- Look at your shared infrastructure first
- Use Search Console data to narrow down the cause
- Don’t rule out DNS just because regular traffic looks fine
- Keep an eye on your logs
For anyone running multiple sites behind a CDN, ensure you:
- Have good logging set up
- Watch your crawl rates
- Know who to contact when issues arise
- Monitor your infrastructure provider
Featured Image: PeopleImages.com – Yuri A/Shutterstock