Ever wondered why a perfectly good page on your Shopify store isn't showing up in search results? It could be because of your robots.txt file. This file informs search engines which pages on your site they should index and crawl, and which ones they should skip.
Have you ever received "Indexed, though blocked by robots.txt" notification from Google Search Console? Unsure about its significance or what steps to take next? No need to fret! In this blog, I will cover everything in detail you need to know about this issue and provide you with a step-by-step guide for resolving it.
Shopify robots.txt is a file that works as a guide for search engine robots, commonly referred to as "crawlers" or "spiders," regarding which pages or sections of a website they should refrain from crawling or indexing.
In the event that your robots.txt file contains the line "Disallow: /checkout," it means that the robot will avoid crawling or indexing any pages located within the checkout section of the website.
Shopify is the platform that automates the creation of a robots.txt file. Store owners can further customize in the Shopify admin to not to crawl or index specific pages. This customization is crucial for optimizing your overall store's performance.
To access the robots.txt file for your Shopify store, you just need to add "/robots.txt" to the end of your store's URL. Alternatively, you can access it through your Shopify admin. Just log in, then navigate to Online Store > Preferences > Search engine listing preview > Edit robots.txt
The default robots.txt file provided by Shopify automatically blocks Google from indexing irrelevant pages or those that might create duplicate content. This naturally benefits your shopify store's SEO.
The default configuration of Shopify's robots.txt file contains the following instructions:
Furthermore, the robots.txt file may contain the line "Allow: /" and "Sitemap: https://mystore.com/sitemap.xml https://mystore.com/sitemap.xml", directing the robots to the store's sitemap, providing a comprehensive list of all website URLs for indexing purposes.
Limiting the indexing of lower-value or redundant pages enables search engines to guide users to more relevant content, resulting in increased targeted traffic.
In this issue, certain pages intended to be excluded from indexing may still be indexed by Google, despite being blocked in your store's robots.txt file.
The following are the reasons causing this issue:
To resolve the "Indexed, though blocked by robots.txt" problem, it's essential to verify that the file effectively blocks the pages you wish to prevent from being indexed. You can either inspect the syntax of your robots.txt file or use the Google Search Console to identify which pages are currently being blocked.
Here are steps to resolve this issue with the help of Google Search Console
Remember, just one error in editing your robots.txt file can have a notable effect on how your web pages are indexed and your SEO rankings. Don’t do it yourself. We advise you to take an expert’s help for editing your robots.txt file as it requires technical knowledge
Essentially, any page holding valuable information, that you desire both search engines and customers to access, should not be blocked. If you notice any pages being blocked by robots.txt, it's advisable to eliminate them from the robots.txt file and provide Google with a fresh sitemap for reindexing them.
Pages such as login pages, shopping carts, and checkout pages, and pages that contain duplicate and low-quality content should be blocked by robots.txt.
The "Shopify Indexed Though Blocked by Robots.txt" issue can be daunting, but thankfully, it's solvable! By understanding the reasons behind it and following the steps outlined in this guide, you can ensure your desired pages are properly indexed and discoverable in search results.