It seems that since mid January, Netlify is treating requests from the Google Search index bot differently, breaking sites in severe ways; my website has completely fallen out from the Google Index!
This graph from the Google Search console shows when the problem started appearing; around that time nothing about my site has changed:
For example for the main page of my site, https://www.joachim-breitner.de/blog, it would report “error due to redirections”. But there is no redirect happening!
The helpful people at the Google Search Community solved the puzzle: It seems that Netlify is behaving differently if the user agent is Googlebot Smartphone
! In particular, it seems to be ignoring the edge functions and instead applies some other, contradicting URL normalization, which of course completely breaks the site. Compare
- https://view.hugo-decoded.be/?scheme=https&url=www.joachim-breitner.de%2Fblog&ua=Googlebot+Smartphone&ref=None
- https://view.hugo-decoded.be/?scheme=https&url=www.joachim-breitner.de%2Fblog
A very likely related, and also somewhat serious issue is that the _headers
file seems to be ignored for that user agent, compare
- https://view.hugo-decoded.be/?scheme=https&url=alt2.joachim-breitner.de%2F&ua=Googlebot+Smartphone&ref=None
- https://view.hugo-decoded.be/?scheme=https&url=alt2.joachim-breitner.de%2F&ref=None
where I am using _headers
to set x-robots-tag: noindex
, but if that setting isn’t effective precisely when Google is requesting the page, then that has a certain irony to it.
This very strongly points towards a problem within Netlify, and a severe one at that, as it can cause sites to becomes unfindable on Google. Is there a swift fix possible? Or would Netlify users somehow have to live with that?