Mobile app version of vmapp.org
Login or Join
BetL925

: How can I stop Google from indexing "pretty links" external redirects from my WordPress site? When I search site:[example].com in Google for my blog, the majority of the pages that are being

@BetL925

Posted in: #Google #Indexing #Links #SearchEngines #WebCrawlers

When I search site:[example].com in Google for my blog, the majority of the pages that are being indexed are "pretty links" to external sites.

I created these links using the Pretty Links plugin so that I can make links to external sites look nice and clean, and also so I can track how many people click through to those sites from my blog.

But many of them were created before I understood anything about follow/no follow links. As I've learned more, I realised this isn't ideal and have gone and changed each of these links to "no follow" , but this hasn't stopped then from being indexed by Google.

How can I stop these links from being index?

10.02% popularity Vote Up Vote Down


Login to follow query

More posts by @BetL925

2 Comments

Sorted by latest first Latest Oldest Best

 

@Welton855

Add

<meta name="robots" content="noindex,nofollow"/>
<meta name="googlebot" content="noindex,nofollow,noarchive,unavailable_after:[date]"/>


at the head of pages you want to exclude from Google search, where [date] should be a date in the past to tell Googlebot that the page should no longer be available at present.

10% popularity Vote Up Vote Down


 

@Ogunnowo487

Google doesn't index links, it indexes pages.

It seems that the pages that these pretty links point to like example.org/clickout/5 don't have a proper robots setting or http status code to let Google know that you don't want these pages in the index.

Have you tried the following?


Add a meta robots noindex to the intermediate clickout pages? This will stop Google from indexing the intermediate clickout pages.
Switch to non-pretty external links that are tracked using Javascript click events (for instance using Google Analytics?)
Use a tool like httpstatus.io/ to check the http status codes that the intermediate page is returning. As Stephen noted, if these are 301 it's unlikely that these clickout pages will get indexed themselves.


[Edit] Removed suggestion to block the intermediate URLs in robots.txt, added suggestion to check http status code.

10% popularity Vote Up Vote Down


Back to top | Use Dark Theme