Support » Plugin: Sitemap » After submitting the sitemap- Robots.txt not functioning properly

  • Resolved rrachel11985

    (@rrachel11985)


    Google’s crawlers are being dined access by the robots.txt file.
    This is the warning from google master tools:
    “Url blocked by robots.txt.”
    “Sitemap contains urls which are blocked by robots.txt”

    Here is the file:
    http://www.splendidstoneandtile.com/robots.txt

    The search engines allowed to crawl from the reading settings.
    We have tried to produce new robots.txt file but it’s still not working.
    The blocked URL’s are important to the site and includes the home page.

    We also tried different ‘disallow’ commands
    and we also tested the sitemap in different programs.

    https://wordpress.org/plugins/sitemap/

Viewing 1 replies (of 1 total)
Viewing 1 replies (of 1 total)
  • The topic ‘After submitting the sitemap- Robots.txt not functioning properly’ is closed to new replies.