Jon Jensen wrote:
It's a way for web sites to tell bots/spiders/crawlers how to behave (e.g. what they can and cannot view), though it is of course entirely up to the bot to comply.

http://www.robotstxt.org/wc/norobots.html
We find that there are a ton of bots that do the exact opposite of what you tell them to. I'd use .htaccess to deny all traffic in directories and files that no one needs to access via http, to password protect admin / private directories, and to just deny bots in directories you simply don't want indexed.

Robots.txt files are great though for telling google which pictures it can and can't use :-)

-V

_______________________________________________

UPHPU mailing list
[email protected]
http://uphpu.org/mailman/listinfo/uphpu
IRC: #uphpu on irc.freenode.net

Reply via email to