Mobile app version of vmapp.org
Login or Join
Reiling115

: Should I use nofollow on internal links that don't need to be crawled? We have a product navigation engine that allow our customers to filter products based on attributes (size, color, etc.).

@Reiling115

Posted in: #CanonicalUrl #Google #Nofollow #Seo #WebCrawlers

We have a product navigation engine that allow our customers to filter products based on attributes (size, color, etc.). We are implementing an upgrade that allows them to use "multi-select" refinements. Basically they can say "show me all the red, blue, and green products". We want the search engines to crawl just the first page without any refinements. We've set the canonical of the page with refinments on it to the one without any refinements.

From a system resource standpoint I'm not sure I want google crawling all of those refinements links. The links that those refinements generate can be exponentially different based on the the order of the refinements, how many have been selected, etc. It just seems like a waste for google to try and index those pages.

Should I put a no-follow tag on those refinement URL's so that google doesn't even crawl them?

10.03% popularity Vote Up Vote Down


Login to follow query

More posts by @Reiling115

3 Comments

Sorted by latest first Latest Oldest Best

 

@Lengel546

Since you mentioned the pages can be "exponentially different" you might want to consider using the "follow, noindex" meta tag, the different classifiers can provide Google with additional information about the product so it is potentially beneficial to let them know about the additional content even if you don't want it indexed (this can help for long tail queries).

10% popularity Vote Up Vote Down


 

@Kevin317

If you're sure Google is reaching the all of your products, and it seems like you are since you're using canonical URLs, then it's a good idea to prevent Google from crawling those pages since all it will do is eat up your bandwidth and server resources.

Using nofollow on any internal links is a good start. I'd also consider using the robots meta tag or x-robots-tag HTTP header to tell them not to index it, too. That way you're sure even if they find those URLs through external linking or otherwise that they will definitely not index those page. Blocking them via robots.txt is also a good thing to since you can use wildcards to catch the pages with query strings, etc.

10% popularity Vote Up Vote Down


 

@Bryan171

Nofollow doesn't stop Google from crawling pages, it stops any 'link juice' being passed to those pages.

Robots.txt is what you want to prevent google from crawling pages.
www.robotstxt.org/robotstxt.html

10% popularity Vote Up Vote Down


Back to top | Use Dark Theme