Micah Cowan wrote:

> If you mean that you want Wget to find any file that matches that
> wildcard, well no: Wget can do that for FTP, which supports directory
> listings; it can't do that for HTTP, which has no means for listing
> files in a "directory" (unless it has been extended, for example with
> WebDAV, to do so).

Seems to me that is a big "unless" because we've all seen lots of websites
that have http directory listings. Apache will do it out of the box (and by
default) if there is no index.htm[l] file in the directory.

Perhaps we could have a feature to grab all or some of the files in a HTTP
directory listing. Maybe something like this could be made to work:

wget http://www.exelana.com/images/mc*.gif

Perhaps we would need an option such as --http-directory (the first thing
that came to mind, but not necessarily the most intuitive name for the
option) to explicitly tell wget how it is expected to behave. Or perhaps it
can just try stripping the filename when doing an http request and wildcards
are specified.

At any rate (with or without the command line option), wget would retrieve
http://www.exelana.com/images/ and then retrieve any links where the target
matches mc*.gif.

If wget is going to explicitly support http directory listings, it probably
needs to be intelligent enough to ignore the sorting options. In the case of
Apache, that would be things like <A HREF="?N=D">Name</A>.

Anyone have any idea how many different http directory listing formats are
out there?

Tony

Reply via email to