When Google asks to not have special treatment for their bot, they are
referring to content more than anything. Regarding the session id being
coded in the URL, see the Technical guidelines section of Google's
Webmaster Guidelines -
http://www.google.com/support/webmasters/bin/answer.py?answer=35769#desi
gn 

It specifically recommends "allow(ing) search bots to crawl your sites
without session IDs or arguments that track their path through the
site."

-----Original Message-----
From: Johan Compagner [mailto:[EMAIL PROTECTED] 
Sent: Thursday, April 03, 2008 7:35 AM
To: [email protected]
Subject: Re: Removing the jsessionid for SEO

isnt google always saying that you shouldn't alter behavior of your site
depending of it is there bot or not?

On Thu, Apr 3, 2008 at 1:00 PM, Artur W. <[EMAIL PROTECTED]> wrote:

>
> Hi!
>
>
> igor.vaynberg wrote:
> >
> > also by doing what you have done users with cookies disabled wont be
> > able to use your site...
> >
>
> In my opinion session id is a problem. Google index the same page
again
> and
> again.
>
> About the users without cookies we can do like this:
>
>
>        static class Unbuffered extends WebResponse {
>
>                 private static final String[] botAgents = {
"onetszukaj",
> "googlebot",
> "appie", "architext",
>                        "jeeves", "bjaaland", "ferret", "gulliver",
> "harvest", "htdig",
>                        "linkwalker", "lycos_", "moget",
"muscatferret",
> "myweb", "nomad",
> "scooter",
>                        "yahoo!\\sslurp\\schina", "slurp", "weblayers",
> "antibot", "bruinbot",
> "digout4u",
>                        "echo!", "ia_archiver", "jennybot", "mercator",
> "netcraft", "msnbot",
> "petersnews",
>                        "unlost_web_crawler", "voila", "webbase",
> "webcollage", "cfetch",
> "zyborg",
>                        "wisenutbot", "robot", "crawl", "spider" }; /*
and
> so on... */
>
>                public Unbuffered(final HttpServletResponse res) {
>            super(res);
>         }
>
>        @Override
>        public CharSequence encodeURL(final CharSequence url) {
>             return isAgent() ? url : super.encodeURL(url);
>        }
>
>                private static boolean isAgent() {
>
>                        String agent =
>
>
((WebRequest)RequestCycle.get().getRequest()).getHttpServletRequest().ge
tHeader("User-Agent");
>
>                        for(String bot : botAgents) {
>                                if (agent.toLowerCase().indexOf(bot) !=
-1)
> {
>                                        return true;
>                                }
>                        }
>
>                        return false;
>                }
>    }
>
>
> I didn't test this code but I do similar thing in my old application
in
> Spring and it works.
>
> Take care,
> Artur
>
>
> --
> View this message in context:
>
http://www.nabble.com/Removing-the-jsessionid-for-SEO-tp16464534p1646739
6.html
> Sent from the Wicket - User mailing list archive at Nabble.com.
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: [EMAIL PROTECTED]
> For additional commands, e-mail: [EMAIL PROTECTED]
>
>

______________

The information contained in this message is proprietary and/or confidential. 
If you are not the 
intended recipient, please: (i) delete the message and all copies; (ii) do not 
disclose, 
distribute or use the message in any manner; and (iii) notify the sender 
immediately. In addition, 
please be aware that any message addressed to our domain is subject to 
archiving and review by 
persons other than the intended recipient. Thank you.
_____________

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to