: Duplicate content caused by HTTPS pages I have a landing page that can be called by HTTP and also by HTTPS. The canonical URL links to the HTTP version. In some comments I have read that
I have a landing page that can be called by HTTP and also by HTTPS.
The canonical URL links to the HTTP version.
In some comments I have read that the HTTPS version is never read by a Googlebot.
Is this true?
Some people say, you waste Google crawl credits because Google needs to crawl the same content again. Is this true?
Do I really have a problem with duplicate content?
More posts by @Murray432
1 Comments
Sorted by latest first Latest Oldest Best
Googlebot will crawl HTTPS links. Source: Will Googlebot crawl HTTPS URLs that are encrypted by SSL?
Googlebot does have a limited crawl budget for every site. The number of pages on a site that Google is willing to crawl is determined by the amount of PageRank in the site and how fast your server is. Googlebot does so much crawling that very few sites have more pages than Googlebot is willing to crawl. Unless you have hundreds of thousands of pages, very little Pagerank, and a slow server; it shouldn't be an issue for you. Google says:
Our goal is to crawl as many pages from your site as we can on each visit without overwhelming your server's bandwidth.
The fact that you use a canonical meta tag means that you will not have any duplicate content problems. Your canonical meta tag will let Google safely identify the duplicate content and tell it which URL to show in the SERPs. Even without the canonical, Googlebot does a fairly good job identifying and correctly handling common causes of duplicate content such as this. Google rarely penalizes for internally duplicated content. For more information see: What is duplicate content and how can I avoid being penalized for it on my site?
Terms of Use Create Support ticket Your support tickets Stock Market News! © vmapp.org2024 All Rights reserved.