PDA

View Full Version : What is robots.txt



rehanamalik
11-21-2017, 12:35 AM
hello friends,


What is robots.txt?

Medventa
11-21-2017, 10:34 PM
The robots exclusion standard, also known as the robots exclusion protocol or simply robots.txt, is a standard used by websites to communicate with web crawlers and other web robots. The standard specifies how to inform the web robot about which areas of the website should not be processed or scanned.

alwaysprompt
11-22-2017, 02:33 AM
The robots exclusion standard, also known as the robots exclusion protocol or simply robots.txt, is a standard used by websites to communicate with web crawlers and other web robots. The standard specifies how to inform the web robot about which areas of the website should not be processed or scanned. Robots are often used by search engines to categorize websites. Not all robots cooperate with the standard; email harvesters, spambots, malware, and robots that scan for security vulnerabilities may even start with the portions of the website where they have been told to stay out. The standard is different from but can be used in conjunction with, Sitemaps, a robot inclusion standard for websites.

damponting44
11-23-2017, 12:40 AM
It is great when search engines frequently visit your site and index your content but often there are cases when indexing parts of your online content is not what you want. For instance, if you have two versions of a page , you'd rather have the printing version excluded from crawling, otherwise you risk being imposed a duplicate content penalty.

nancy07
11-23-2017, 01:32 AM
A "robots.txt" file is a text file placed on your server which contains a list of robots and "disallows" for those robots.
Using a robots.txt file is easy, but does require access to your server's root location. For instance, if your site is located at:
http://abc.com/mysite/index.html
you will need to be able to create a file located here:
http://abc.com/robots.txt
If you cannot access your server's root location you will not be able to use a robots.txt file to exclude pages from your index.

fayeseom
11-23-2017, 11:31 PM
Robots.txt is a text file webmasters create to instruct web robots typically search engine robot to crawl pages on their website.

velsonsocial
11-28-2017, 01:42 AM
The robots exclusion protocol (REP), or robots.txt is a text file webmasters make to instruct robots (usually search engine robots) how to crawl and index pages on their website

virginoilseom
11-28-2017, 08:25 PM
robots.txt, is a standard utilized by sites to speak with web crawlers and other web robots. The standard determines how to educate the web robot about which regions of the site ought not be prepared or filtered.

simon56
12-06-2017, 12:04 AM
Robots.txt is a text file that lists webpages which contain instructions for search engines robots. The file lists webpages that are allowed and disallowed from search engine crawling.