Disallowed Pages Indexed by Google

Yes, there can be pages that are disallowed in the robots.txt file but are still indexed by Google. The robots.txt file is used to instruct search engine crawlers on which pages or sections of a website should not be crawled. However, it does not prevent those pages from being indexed if they are linked to from other websites or if Google has already indexed them before the disallow rule was added. To ensure that a page is not indexed, you should use the tag within the HTML of the page itself.

For Shopify, you can edit the robots.txt.liquid file in your theme to disallow specific pages. For example:

Disallow: /page-you-want-to-block

For WooCommerce on WordPress, you can edit the robots.txt file directly or use an SEO plugin like Yoast SEO to manage these settings. In Yoast SEO, you can go to SEO > Tools > File editor to edit the robots.txt file.

Remember, to fully prevent a page from being indexed, use the tag in the HTML of the page:

In Shopify, you can add this tag by editing the theme’s liquid files. In WooCommerce, you can add this tag by editing the page template files or using an SEO plugin to manage meta tags.