The robots.txt file is used by websites to communicate with web crawlers and robots about how to crawl and index website pages. It was proposed in 1994 as a standard for webmasters to control how search engine robots access their sites. Examples show how to block all crawlers, block specific crawlers from folders or pages, exclude parts of the site from crawling, and set crawl delays. Important rules note using meta robots tags for restricting indexing, that malicious crawlers may ignore robots.txt, and specifics of the protocol syntax.