We got an error on a new site:
and Google was saying it couldn’t access the site. Privacy settings were turned to allow indexing.
When we unchecked the box “Add sitemap URL to the virtual robots.txt file” it started working. Can you please explain? Should we always uncheck that box? Why is it checked by default if it blocks the indexing?
Thanks for your help in understanding this!
- The topic ‘Google not crawling’ is closed to new replies.