Robots.txt error
-
Hi. I received an error notification with robots.txt
How can I exclude this?# BEGIN W3TC ROBOTS
User-agent: * Disallow: /wp-content/cache/
# END W3TC ROBOTSAt the moment, I have manually changed the file robots.txt, but when clearing the cache in the plugin, this entry appears again.
The page I need help with: [log in to see the link]
-
See https://wordpress.org/support/topic/robots-txt-cache-directory/ for a quick fix (basically
chmod 444 robots.txt).Hello @dmitresku @galfom
Thank you for reaching out and I am sorry about the issue you are experiencing.
This issue is reported numerous times and we already gave a Github issue reported for this.
The temporary solution is either to revert back to the previous version of W3TC or to change the permission for the robots.txt file.
Please refer to the existing topic for this issue where the Github issue is posted
https://wordpress.org/support/topic/w3tc-conflict-with-aioseo/Thanks!
Hi, i am using Yoast, but see error in robots:
# BEGIN W3TC ROBOTS
User-agent: *
Disallow: /wp-content/cache/
# END W3TC ROBOTSLast update (2.1.7) W3 Total Cache have a bug. Version 2.1.6 works correctly!
How do we change permission for the robots.txt.file or revert the plugin to a previous version? I am not a coder. There’s needs to be a solution for this type of users as well.
Hello @chaosfranklin
The best thing to do is to delete the robots.txt file, create your own and set the permissions to 444.
Alternatively, you can install the WP rollback plugin and revert to the previous version of the plugin.
The files that were changed are:
Generic_Environment.php
Util_Rule.php
w3-total-cache-api.php
So you can download the previous version from here and upload the files that were changed.
Once again thank you for your patience and we are working on a fix for thisI don’t know how to delete the robots.txt.file. Instructions would be more helpful.
Hello @chaosfranklin
The robots.txt file should be generated at the root of your website public_html/yourwebsite.com/
You can cancel or FTP to your server and delete the file manually.
I hope this helps!I’ve tried the WP rollback plugin because I am not a coder. I’m a usual user of WordPress. It did not resolve the problem. I’m still getting errors.
Furthermore, my hosting package does provide me access to the root of my website. I have a package only for WordPress.
Hello @chaosfranklin
Thank you for the info.
Even on shared hosting, you should have the access to your server for accounts, files, installing WordPress, etc.I am not sure where you hosted your website but you should have access to your website files.
Please reach out to your hosting support for more details about how to access it.
Thanks!This is WordPress hosting from a web host. WordPress is pre-installed. I do not have access to WordPress’s files. None of your advice in any conversation about this issue has been helpful.
Please provide a solution for people and casual users, who are not coders. This glitch is hurting my business.
Игнорируете мой вопрос на английском, тогда напишу по-русски! Бажное правило “Disallow: /wp-content/cache/”, которое автоматически добавляется в роботс блокирует доступ Гуглу к минифицированным данным кэша, после чего в Консоли Гугла пачками появляются ошибки по мобильной оптимизации.
Во всех ваших рекомендациях указано, что такая проблема появляется только у сайтов, использующих All in One Seo, но я пользуюсь Yoast! Откатил плагин к старой версии и всё работает. Ваше последнее обновление содержит баг, который и создаёт вышеуказанные проблемы.
Hello @abogachev
Thank you for the information. Yes, you are correct. The problem is in the robots.txt file and we are working on a fix https://github.com/W3EDGE/w3-total-cache/issues/420
@chaosfranklin I’ve never heard of a hosting provider that does not allow users to access their files.
Please let me know the URL of your hosting provider or drop us a note with the details via the plugin in Performance>Support.Thanks!
Hello @abogachev @chaosfranklin @dmitresku
We have released a patch in version 2.1.8. We do apologize for any inconvenience. Please update and do let us know if there are any issues. We will be happy to assist you.
Thank you.The update has not resolved this issue on my end. Google Search Console continues to generate errors and failed validations. I’ve updated the plugin and cleaned the site’s cache.
The topic ‘Robots.txt error’ is closed to new replies.