: Google continues to index Word documents and PDF files that were blocked by robots.txt and have been removed from the server I've been having some issues removing some content from Google's search
I've been having some issues removing some content from Google's search results. I have some word documents and pdf files in public directories that I have since password protected. When you try to access any of the documents you are prompted for a user name and password. The documents have also been completely removed from the server.
I did have a disallow on the directories in my robots.txt but this didn't seem to do the trick either.
I've also been adding the urls to the temporary removal tool but they just expire and appear back in the results.
Any suggestions on how to completely remove the content that used to be in these directories from Googles search results? The urls currently return a 401. Would a 410 be more effective?
More posts by @Shelley277
1 Comments
Sorted by latest first Latest Oldest Best
Google have two tools to remove content from search.
www.google.com/webmasters/tools/url-removal
This will remove the content for temporary(90 days), and it may index the content again as you said in your question.
And this is new tool, I don't know when they launch it, but it work's really great for me.
www.google.com/webmasters/tools/removals
This tool have only two rules, if the content return 4xx error, simply remove it from search results ASAP, no matter you own the site or not.
It's damn good, I successfully remove search results from github site as well, which return 404 error.
Terms of Use Create Support ticket Your support tickets Stock Market News! © vmapp.org2024 All Rights reserved.