A list of bad bots and nasty spiders..........

Hi malware fighters,

There are well behaved bots like Googlebot, and the bot that works DrWeb’s av linkscanner plug-in or finjan secure browsing, but there are bad bots and spiders as well, look here:
http://www.kloth.net/internet/badbots.php

polonus

P.S. to block : http://www.tech-faq.com/bad-robots.shtml
How:
Preventing Web Site Downloading Using robots.txt

The first step is to disallow the downloading programs in your robots.txt file. To do this, you will need to define which bad robots you wish to disallow.

Disallowing bad programs in robots.txt does not prevent all web site downloading, because many bad programs simply ignore the contents of robots.txt and do what they want to do.
Preventing Web Site Downloading Using User Agent Blocking in httpd.conf

Another method is to exclude the downloading programs user agent in httpd.conf.

Add every agent you wish to exclude to httpd.conf:

SetEnvIfNoCase User-Agent ^Httrack keep_away
SetEnvIfNoCase User-Agent ^Offline Explorer keep_away
SetEnvIfNoCase User-Agent ^psbot keep_away
SetEnvIfNoCase User-Agent ^Teleport keep_away
SetEnvIfNoCase User-Agent ^WebCopier keep_away
SetEnvIfNoCase User-Agent ^WebReaper keep_away
SetEnvIfNoCase User-Agent ^Webstripper keep_away


Order Allow,Deny
Allow from all
Deny from env=keep_away

User agent blocking also does not prevent all web site downloading, because the user can delete his user agent or spoof it to appear to be Internet Explorer or another common browser.

pol