[Plugin: Google XML Sitemaps] Robots.txt File blocking Google from Crawling
Every since updating the plugin Google is not able to crawl my site, urls are blocked. All of my post pages seem to be blocked. Post pages that once appeared in Google Search are now replaced with home page (Google is able to send the home page due to pagination – picking up some of the words from within the post because of the link to the home page).
I would really appreciate some assistance because I do not know how to begin to fix this. I have noticed others have had the same issue when searching in Google forums and the problems seems to stem from this section of the plugin settings:
Add sitemap URL to the virtual robots.txt file.
The virtual robots.txt generated by WordPress is used. A real robots.txt file must NOT exist in the blog directory!
I’m not sure if the backend settings need to be adjusted or what needs to be done to correct this issue but I am concerned if I don’t get it fixed sooner than later I may lose my ranking. Others have stated once they fixed this issue everything returned to normal within a couple of weeks.
- The topic ‘[Plugin: Google XML Sitemaps] Robots.txt File blocking Google from Crawling’ is closed to new replies.