Introduction

Robots.txt is an important file for your website that tells search engine crawlers which pages can and cannot be accessed. When a page is “indexed but blocked by robots txt”, it means that while the page has been indexed by search engines, it is not being crawled or indexed due to the instructions in the robots.txt file. This issue needs to be fixed in order to ensure that all pages on your website are accessible to search engine crawlers.

Check the Robots.txt File for Mistakes
Check the Robots.txt File for Mistakes

Check the Robots.txt File for Mistakes

The first step in fixing the issue of indexed but blocked by robots txt is to check the robots.txt file for any mistakes. This can be done by accessing the robots.txt file directly from the root folder of your website. Once you have opened the file, you can look for any mistakes or errors that may be causing the page to be blocked. Common mistakes to look out for include incorrect syntax, disallowing access to pages that should be allowed, and allowing access to pages that should be blocked.

Make Sure the Pages are Indexable

Once you have checked the robots.txt file for any mistakes, the next step is to make sure that the pages in question are indexable. This can be done by ensuring that the page is properly linked to other pages on the website, and that there are no broken links or redirects that could be preventing the page from being indexed. Additionally, you should also make sure that the page is not being blocked by any other settings or configurations.

Use a Meta Robots Tag
Use a Meta Robots Tag

Use a Meta Robots Tag

Another way to fix indexed but blocked by robots txt issues is to use a meta robots tag. A meta robots tag is an HTML element that provides instructions to search engine crawlers about how to crawl and index a page. The meta robots tag can be used to override the instructions in the robots.txt file, allowing the page to be indexed and crawled.

Submit a URL Removal Request in Google Search Console
Submit a URL Removal Request in Google Search Console

Submit a URL Removal Request in Google Search Console

If the issue persists after making changes to the robots.txt file and using a meta robots tag, then you should submit a URL removal request in Google Search Console. This will allow you to remove the page from the Google search index, thereby resolving the issue of indexed but blocked by robots txt.

Create an XML Sitemap to Help Search Engines Access Blocked Content

Finally, another way to fix indexed but blocked by robots txt issues is to create an XML sitemap. An XML sitemap is an XML file that contains information about the pages on your website, including the URLs and the frequency with which they are updated. By creating an XML sitemap, you can help search engine crawlers find and access the pages on your website, even if they are blocked by robots.txt.

Conclusion

In conclusion, indexed but blocked by robots txt is an issue that can cause pages on your website to be inaccessible to search engine crawlers. In order to fix this issue, you should check the robots.txt file for any mistakes, make sure that the pages in question are indexable, use a meta robots tag, submit a URL removal request in Google Search Console, and create an XML sitemap. Following these steps will help ensure that all pages on your website are accessible to search engine crawlers.

(Note: Is this article not meeting your expectations? Do you have knowledge or insights to share? Unlock new opportunities and expand your reach by joining our authors team. Click Registration to join us and share your expertise with our readers.)

By Happy Sharer

Hi, I'm Happy Sharer and I love sharing interesting and useful knowledge with others. I have a passion for learning and enjoy explaining complex concepts in a simple way.

Leave a Reply

Your email address will not be published. Required fields are marked *