On Thursday 30 May 2002 19:42, James Holden wrote:
> I do it a different way - using IP's can be a bit of a bummer since they
> might change but often as not the hostnames dont.
> I detect spiders using the HTTP_USER_AGENT which identifies them either as
> say "kitty once hourly", "GoogleBot" or "Lycos" or some such - most of the
> decent spiders use the user_agent var to identify themselves and you can
> display alternate info dependent on which spider it is.
Do note that some (most?) crawlers don't take kindly to this practice of
delivering content based on the HTTP_USER_AGENT. If you're found out you'll
most likely be removed from the search engine in question.
Jason Wong -> Gremlins Associates -> www.gremlins.com.hk
Open Source Software Systems Integrators
* Web Design & Hosting * Internet & Intranet Applications Development *
On the whole, I'd rather be in Philadelphia.
-- W.C. Fields' epitaph
PHP General Mailing List (http://www.php.net/)
To unsubscribe, visit: http://www.php.net/unsub.php