Author Topic: A list of bad bots and nasty spiders..........  (Read 2445 times)

0 Members and 1 Guest are viewing this topic.

Offline polonus

  • Avast Überevangelist
  • Probably Bot
  • *****
  • Posts: 33913
  • malware fighter
A list of bad bots and nasty spiders..........
« on: June 17, 2008, 12:36:05 AM »
Hi malware fighters,

There are well behaved bots like Googlebot, and the bot that works DrWeb's av linkscanner plug-in or finjan secure browsing, but there are bad bots and spiders as well, look here:
http://www.kloth.net/internet/badbots.php

polonus

P.S. to block : http://www.tech-faq.com/bad-robots.shtml
How:
Preventing Web Site Downloading Using robots.txt

The first step is to disallow the downloading programs in your robots.txt file. To do this, you will need to define which bad robots you wish to disallow.

Disallowing bad programs in robots.txt does not prevent all web site downloading, because many bad programs simply ignore the contents of robots.txt and do what they want to do.
Preventing Web Site Downloading Using User Agent Blocking in httpd.conf

Another method is to exclude the downloading programs user agent in httpd.conf.

Add every agent you wish to exclude to httpd.conf:


    SetEnvIfNoCase User-Agent ^Httrack keep_away
    SetEnvIfNoCase User-Agent ^Offline Explorer keep_away
    SetEnvIfNoCase User-Agent ^psbot keep_away
    SetEnvIfNoCase User-Agent ^Teleport keep_away
    SetEnvIfNoCase User-Agent ^WebCopier keep_away
    SetEnvIfNoCase User-Agent ^WebReaper keep_away
    SetEnvIfNoCase User-Agent ^Webstripper keep_away


    Order Allow,Deny
    Allow from all
    Deny from env=keep_away

User agent blocking also does not prevent all web site downloading, because the user can delete his user agent or spoof it to appear to be Internet Explorer or another common browser.

pol
« Last Edit: June 17, 2008, 12:40:20 AM by polonus »
Cybersecurity is more of an attitude than anything else. Avast Evangelists.

Use NoScript, a limited user account and a virtual machine and be safe(r)!