Re: blocking site downloaders

2009-10-17 Thread James Holmes
Session variables are always stored in the server's memory. J2EE session cookies simply expire in the client when the browser is closed, rather than only at the time set on the session timeout. Bots will often ignore either style of session, as they often don't keep track of cookies. mxAjax /

blocking site downloaders

2009-10-16 Thread Michael Muller
Hey all, Every once in a while I'll notice in my logs that someone comes to one of my sites and hits thousands of pages in a short span and then leaves. This annoys me for a few reasons: (a) It's an unecessary tax on my server (and we all hate taxes) (b) It artificially inflates my page

Re: blocking site downloaders

2009-10-16 Thread Azadi Saryev
some thoughts on the issue: 1) it may very well be a spider from a chinese search engine indexing your pages... i'll leave it up to you to decide if you want none of 1 billion chinese people to be able to find your site... in case it's a search spider, you can try adding appropriate robots.txt

Re: blocking site downloaders

2009-10-16 Thread Mike Chabot
Is it a search engine robot? If so you could block it using robots.txt. What is the Web browser they are using? Do you have any customers in China? If not, you can block the entire country at the firewall level. -Mike Chabot On Fri, Oct 16, 2009 at 10:12 AM, Michael Muller

Re: blocking site downloaders

2009-10-16 Thread Claude Schneegans
(c) What the hell are they doing? Some will search for copyrighted images and, if they find some, they will send you some lawyer letter with a bill of 1500$ or so for using their image. (ie: Getty images,...) Most of them are simply looking for email addresses they could collect and sell to

Re: blocking site downloaders

2009-10-16 Thread Claude Schneegans
in case it's a search spider, you can try adding appropriate robots.txt robots.txt file may be useful for well behaved robots like Google or Yahoo, but bad bot disguise themselves like Explorer or Mozilla and they do not comply to robots.txt directives. Most of them never read it, and some of

Re: blocking site downloaders

2009-10-16 Thread Azadi Saryev
On 17/10/2009 09:08, Claude Schneegans wrote: robots.txt file may be useful for well behaved robots like Google or Yahoo, but bad bot disguise themselves like Explorer or Mozilla and they do not comply to robots.txt directives. that's exactly why i also said of course, not all robots