Palula...

On Wednesday 21 December 2005 06:17, Palula Brasil wrote:
> I created a file with a some strings I don't want my clients to access.
> Very nice it works fine, but it is blocking some sites with string I
> don't want it to block... So I created another acl with permitted
> strings ok? So the thing goes like this...
>
> acl bad_strings url_regex "path_to_file/file"
> acl good_strings url_regex "path_to_file/file"
>
> Denial:
>
> http_access allow good_strings
> http_access deny bad_strings
>
> But the problem is that I blocked the word "anal" on the bad strings
> file and I have the word "canal" (means channel in portuguese) in the
> good_strings file. But now, the word anal can be searched/accessed etc.
> How can I overcome this...

Your syntactical solution would be:

http_access deny bad_strings !good_strings

However blocking by keywords has proven to be very inefficient. It takes a 
user with an IQ of a three year old child to circumvent this "security".
Take the google cache, all the anonymizing proxies, web anonymizers etc.
You can't block "bad content" by using URL keywords decently. Rather - 
depending on the seriousness of blocking - try SquidGuard or consider 
throwing money at a commercial product.

 Christoph
-- 
~
~
".signature" [Modified] 2 lines --100%--                2,41         All

Reply via email to