> I think this one is a keeper, unfortunately. If someone comes up with a 
> better way to keep search engines from being evil, I'm willing to work 
> on it, but I don't see any way to do a POST request from a textual link.

I'm not sure about this, but what about making a pseudo directory for the
buttons.

That is, instead of the URL of the history page being
http://mysite/myswiki/2.history

it could by

http://mysite/myswiki/foo/2.history

My understanding of how a webcrawler should work is that before it reads
any files in directory foo, it will first try to get foo/robots.txt.  If
this file exists, it wouldn't index that directory.

So, for pages that you don't want to be indexed, you make their URL's
pointers to "virtual directories."

Eg:

http://mysite/myswiki/1  remains the same, but
http://mysite/myswiki/1.edit becomes
http://mysite/myswiki/virtual/1.edit

and then virtual/robots.txt would return some content to avoid those links
being chased.

I don't know if that would work, though.  Any thoughts?

-Eric


Reply via email to