Wednesday, September 7, 2011

Robot.txt and its Use


Every website owner wants that the search engines frequently visit their website and index their content but often there are some cases when indexing parts of your online content is not what you want or you don’t want search engine bots crawl your website or any page. If you have private, sensitive and confidential data and information on your website that you do not want the world to see, you will also have a preference that search engines do not index these pages. Moreover, if you want to save some bandwidth by excluding images, style sheets and JavaScript from indexing, you also need a way to tell spiders to repel from these objects.


To out come from this problem Robot Metatags is an option but it has also some limitations. By Robots Metatags you can tell search engines which files and folders on your website to avoid is with the use of the Robots Metatag but the drawback is that the all search engines not read Robot metatags, the Robots matatag can simply go unnoticed by some search engine. The complete solution of this problem is Robot.txt.


Robot.txt is known as Robot Exclusion Standard, it is also termed as the Robots Exclusion Protocol or robots.txt protocol. Robot.txt is a convention to put off cooperating web spiders and other web robots from crawling and indexing complete or any specified part of a website which is in public viewable. Robot.txt has different codes to let search engine know what page or content not to be crawled. One can choose different codes for different requirement. 

Robots.txt is simply a text file (not an html file) you put on your site to tell search robots which pages you would like them not to visit. Robots.txt is by no means mandatory for search engines but generally search engines obey what they are asked not to do. Robot.txt is a better way to tell search engines about your will is to use a robots.txt file.




Click on SEO Services to get the expert services

No comments:

Post a Comment