Re: Banning Bad Bots by UA

by jim <jim(at)newglobal.net>

 Date:  Sat, 23 Jun 2001 08:34:39 -0500
 To:  "Rich Bowen" <rbowen(at)rcbowen.com>,
hwg-servers(at)hwg.org
 References:  newglobal omnitec newglobal2
  todo: View Thread, Original
Thanks Rich, that was very helpful. Would you happen to know if the syntax 
would allow for banning multiple user agents in the form of..

SetEnvIf User-Agent Wget, Emailsiphon, ExtractorPro GoAway

or is an entire set of directives needed for each UA?

I have tried several combinations but it seems to work only with one UA at 
a time which is indeed helpful, but banning a "list" of these critters 
would be most excellent.

Thanks again,

Jim
At 07:07 AM 6/23/01 -0400, Rich Bowen wrote:

>On Sat, 16 Jun 2001 18:55:43 -0500, jim said:
>
> > Hi folks,
> >
> >  I was intending to use .htaccess to ban bots based on User Agent being 
> used
> >  by less than the nicest of Netizens using Spiders and Bots that do not
> >  follow exclusion protocol. I'm running Apache servers but the build is to
> >  support Frontpage 2000 and has problems such that mod_rewrite won't 
> compile
> >  in without going through a lot of pain. Any ideas that won't make me 
> go the
> >  SSL route?
>
>Sorry, I missed this note earlier.
>
>See http://apachetoday.com/news_story.php3?ltsn=2000-11-13-003-01-SC-LF-SW
>which talks about denying access based on a variety of different things. There
>are two things you want to do. You need to explicitly deny by host, if you 
>know
>for sure which hosts are spidering you. And you need to deny by user agent,
>which is discussed later in the article, when I discuss "SetEnvIf".
>
>Hope this helps.
>
>--
>Rich Bowen - rbowen(at)rcbowen.com
>http://www.rcbowen.com/kenya/

HWG: hwg-servers mailing list archives, maintained by Webmasters @ IWA