[wicked] Horde :: Wicked ::Patches to solve problems with web robots

Chuck Hagenbuch chuck at horde.org
Tue Dec 14 19:15:14 PST 2004


Quoting kracker <thekracker at gmail.com>:

> Since web robots crawl through a web page and hit every link on a page
> in the order they are presented, this creates a problem for the wicked
> because it's unlock link comes before the history link and the history
> page contains links to revert to a previous revision without
> confirmation, so robots who crawl through a site can easily /
> unknowingly revert large chunks of your content until your wicked
> installation in total disarray.

This is a clumsy way of doing it; you should just Browser::isRobot() to 
disallow
robots to make changes.

> I started this email after I wrote and tested a wicked patch to lock a
> wiki page immediately after a page is saved (after an edit, see below
> ) so that the wiki pages are by default always locked and must be
> unlocked to edit to reduce the chance of the wiki being unlocked and
> then reverted (only unlocked pages can be reverted .... )

That's unnecessary and bad for general wiki useability.

> Below is are two patches, the first is for the auto_lock feature, the
> second is to not process wiki pages for the IP address of the
> fastsearch.net robot (which is nice because it kills the app if the
> robot tries to use it but lets other robots which are not so
> destructive continue to troll the wiki (i kinda like google cache :) )

Again, use Browser::isRobot().

-chuck

-- 
"But she goes not abroad in search of monsters to destroy." - John 
Quincy Adams


More information about the wicked mailing list