Mobile app version of vmapp.org
Login or Join
LarsenBagley505

: How to tell search engines to lower the robots.txt scanning rate on obsolete subdomains? After putting my site into one domain instead of multiple subdomains, I checked my logs and noticed that

@LarsenBagley505

Posted in: #CrawlRate #Google #RobotsTxt #WebCrawlers

After putting my site into one domain instead of multiple subdomains, I checked my logs and noticed that search engines (particularly google) are scanning robots.txt several times a day for each subdomain. I only left the subdomains active so that people who accessed the old URLs can be automatically redirected to the proper URLs.

I have removed my subdomains (except for the one active one I'm using) from all webmaster tools accounts and I make zero references to them on my site.

Is there a way to minimize the amount of times search engines scan robots.txt on my obsolete subdomains?

Ideally, I'd like to have the search engines scan the robots.txt found on each obsolete domain only one time, and then scan all the necessary URLs on the subdomain and that's it, not scan robots.txt every few seconds.

10.02% popularity Vote Up Vote Down


Login to follow query

More posts by @LarsenBagley505

2 Comments

Sorted by latest first Latest Oldest Best

 

@Odierno851

May be Google check your robots.txt frequently, because you have moved your site content to new location. There is no any way to control robots.txt crawling rate.

But, If your robots.txt did not blocked anything like any directory, then don't use robots.txt at all. For blocking other bots, you can use httacess. I have seen most of sites, don't use robots.txt properly. They specify useragent and allow them to crawl, like this,

User-agent: Googlebot
Allow: /
User-agent: Mediapartners-Google
Disallow:


This type of robots.txt is junk, it just waste space and bandwidth.

So Use Robots.txt only, if you really want to block such a thing, For small type of things, I recommended to use X-Robots header tag.

10% popularity Vote Up Vote Down


 

@Bryan171

You should keep the subdomains setup with redirects to your top level domain - that way old traffic will land on a functional website and Google will re-crawl over time and realised all the traffic has been directed elsewhere

Add this to .htaccess and modify it to suite your sites Url

RewriteEngine on
RewriteCond %{HTTP_HOST} ^m.example.com$
RewriteRule ^ example.com/m%{REQUEST_URI} [L,P]


Read more here: stackoverflow.com/questions/1456956/redirect-subdomain-into-top-level-domains-folder

10% popularity Vote Up Vote Down


Back to top | Use Dark Theme