Approaches Utilized to Avoid Google Indexing
Have you at any time desired to stop Google from indexing a particular URL on your website web site and displaying it in their research engine final results internet pages (SERPs)? If you regulate world-wide-web internet sites extensive ample, a day will most likely arrive when you need to know how to do this.
The 3 strategies most typically employed to reduce the indexing of a URL by Google are as follows:
Utilizing the rel=”nofollow” attribute on all anchor aspects utilised to connection to the site to stop the one-way links from currently being followed by the crawler.
Employing a disallow directive in the site’s robots.txt file to avert the web site from becoming crawled and indexed.
Applying the meta robots tag with the information=”noindex” attribute to prevent the website page from currently being indexed.
Even though the variances in the three strategies surface to be delicate at to start with look, the success can range substantially based on which technique you decide on.
Applying rel=”nofollow” to protect against Google indexing
Several inexperienced webmasters attempt to prevent Google from indexing a individual URL by working with the rel=”nofollow” attribute on HTML anchor factors. They increase the attribute to each individual anchor ingredient on their web page utilized to website link to that URL.
Together with a rel=”nofollow” attribute on a connection prevents Google’s crawler from following the backlink which, in convert, stops them from discovering, crawling, and indexing the goal page. While this method may get the job done as a short-phrase solution, it is not a viable very long-term resolution.
The flaw with this method is that it assumes all inbound backlinks to the URL will contain a rel=”nofollow” attribute. The webmaster, nevertheless, has no way to avert other internet internet sites from linking to the URL with a adopted website link. So the likelihood that the URL will eventually get crawled and indexed making use of this method is very large.
Utilizing robots.txt to reduce Google indexing
Yet another typical strategy used to reduce the indexing of a URL by Google is to use the robots.txt file. A disallow directive can be extra to the robots.txt file for the URL in issue. Google’s crawler will honor the directive which will reduce the web site from remaining crawled and indexed. In some conditions, having said that, the URL can however look in the SERPs.
Occasionally Google will show a URL in their SERPs even though they have never indexed the contents of that web site. If ample world wide web web sites backlink to the URL then Google can typically infer the topic of the site from the link text of people inbound hyperlinks. As a outcome they will clearly show the URL in the SERPs for related lookups. When making use of a disallow directive in the robots.txt file will reduce Google from crawling and indexing a URL, it does not assure that the URL will hardly ever show up in the SERPs.
Making use of the meta robots tag to avoid Google indexing
If you will need to reduce Google from indexing a URL when also avoiding that URL from staying exhibited in the SERPs then the most effective approach is to use a meta robots tag with a content=”noindex” attribute inside the head element of the web webpage. Of google index download , for Google to essentially see this meta robots tag they want to first be in a position to discover and crawl the web page, so do not block the URL with robots.txt. When Google crawls the page and discovers the meta robots noindex tag, they will flag the URL so that it will by no means be shown in the SERPs. This is the most successful way to avoid Google from indexing a URL and exhibiting it in their search success.