-e robots=off
Jon W. Backstrom wrote:
> Dear Gnu Developers,
>
> We just ran into a situation where we had to "spider" a site of our
> own on a outsourced service because the company was going out of
> business. Because wget respects the robots.txt file, however, we
> could not get an archive made until we had the outsourced company
> delete their robots.txt file on the server in the last 2 days of
> service. This might not have had such a happy ending, however.
>
> Would you consider an "--ignore-robotfile" option perhaps, or would
> that be too abusive? I know I can always edit the source and make my
> own, but I wondered if this was something that WGET might want to do
> in a release version or is the potential for abuse too great?
>
> Thank you!
>
> Jon Backstrom
> [EMAIL PROTECTED]