robots.txt Online Generator - Customize Search Engine Crawler Access Rules

Generate robots.txt, output only explicitly configured crawlers

Default - All robots:
Crawl delay:
Sitemap: (leave empty for none)
Normal search robots: Google
googlebot
Baidu
baiduspider
MSN Search
msnbot
Yahoo
yahoo-slurp
Ask/Teoma
teoma
Cuil
twiceler
GigaBlast
gigabot
Scrub The Web
scrubby
DMOZ Checker
robozilla
Nutch
nutch
Alexa/Wayback
ia_archiver
Naver
naverbot, yeti
Special search robots: Google Image
googlebot-image
Google Mobile
googlebot-mobile
Yahoo MM
yahoo-mmcrawler
MSN PicSearch
psbot
SingingFish
asterias
Yahoo Blogs
yahoo-blogs/v3.9
Restricted directories: Paths are relative, but each must start with: "/"

What is robots.txt file

  • robots.txt (all lowercase) is a text file placed in the root directory of a website to tell search engine crawlers which pages can be crawled and which cannot...
  • Because URLs are case-sensitive in some systems, the filename of robots.txt should be uniformly lowercase...
  • If you want to define the behavior of search engine crawlers when accessing subdirectories separately...
  • The robots.txt protocol is not a specification, but a convention...
  • Robots protocol is a widely accepted moral code in the international internet community...

robots.txt file content

  • Whether search engine spiders are accessible or crawlable.
  • Accessibility of search engine spiders to directories or files.
  • Definition of website sitemap path.
  • Crawl delay limit for search engine spiders.

About robots.txt file generator

  • Set the data to be configured through the web interface, click the generate button to get the content of robots.txt in the text input box below.
  • Now, create a blank text file, name it "robots.txt", then copy and paste the above content into "robots.txt".
  • Place "robots.txt" in the root directory of your website, access robots.txt to ensure it is accessible to visitors (such as search engines).