On Thu, Jan 03, 2013 at 06:24:54AM +0000, Andy Smith wrote:
> Hello,
>
> What's the list's preferred techniques for preventing spidering of a
> web application (in this case Mediawiki) by misguided web robots?
>
> robots.txt already in place, but they ignore that of course.
>
> Ideally Apache-based.
>
> I don't particularly care if they are still able to download the
> content or not, I just don't want them taking up every single
> process slot thus impacting non-abusive 'real' web clients. So a
> rate-limiting solution would be acceptable.
http://dominia.org/djao/limitipconn2.html
I implemented this at DreamHost 5 or 6 years ago to great effect and as
far as I know it's still in place. It was implemented for a different
reason, but it worked quite well.
There does seem to be a race condition with regard to a high rate of
incoming connections from a given IP which will cause them to get
dropped slightly lower than the normal threshhold, but I just set our
limit to 20 and it stopped the baddies and I think we got maybe
3 complaints about it ever.
-Jeremy