25'ten fazla konu seçemezsiniz Konular bir harf veya rakamla başlamalı, kısa çizgiler ('-') içerebilir ve en fazla 35 karakter uzunluğunda olabilir.

robots.txt 1.2 KiB

1234567891011121314151617181920212223242526272829303132333435363738
  1. # $Id$
  2. # Do not crawl CVS and .svn directories (they are 403 Forbidden anyway)
  3. User-agent: *
  4. Disallow: CVS
  5. Disallow: .svn
  6. # Prevent excessive search engine hits
  7. Disallow: /cgi-bin/trac.cgi
  8. Disallow: /log
  9. # Don’t crawl git repos
  10. Disallow: /git/*.git/*
  11. Disallow: /git/*.git.broken/*
  12. # "This robot collects content from the Internet for the sole purpose of
  13. # helping educational institutions prevent plagiarism. [...] we compare
  14. # student papers against the content we find on the Internet to see if we
  15. # can find similarities." (http://www.turnitin.com/robot/crawlerinfo.html)
  16. # --> fuck off.
  17. User-Agent: TurnitinBot
  18. Disallow: /
  19. # "NameProtect engages in crawling activity in search of a wide range of
  20. # brand and other intellectual property violations that may be of interest
  21. # to our clients." (http://www.nameprotect.com/botinfo.html)
  22. # --> fuck off.
  23. User-Agent: NPBot
  24. Disallow: /
  25. # "iThenticate® is a new service we have developed to combat the piracy
  26. # of intellectual property and ensure the originality of written work for
  27. # publishers, non-profit agencies, corporations, and newspapers."
  28. # (http://www.slysearch.com/)
  29. # --> fuck off.
  30. User-Agent: SlySearch
  31. Disallow: /