On Mon, May 11, 2009 at 3:07 PM, Platonides <[email protected]> wrote:
> Robert Rohde wrote:
>> Since people are doing dump redesign right now, might I suggest that
>> providing better integration / information on Commons-hosted images
>> would actually be useful.  As far as I know the current system has no
>> way to distinguish between Commons images and missing images except by
>> downloading the Commons dump files.  That can be frustrating since the
>> Commons dumps are larger (and hence more trouble to work with) than
>> all but a handful of other wikis.
>>
>> -Robert Rohde
>
> You only need the image.sql dump from commons to determine if the image
> exists there (it will also include other useful and not-so-useful data
> like filetype, image size, metadata...).
> http://download.wikimedia.org/commonswiki/20090510/

That wouldn't get you file descriptions or copyright status, etc.  If
your goal is something like mirroring a wiki, you really need access
to page descriptions as well.

At present, the main solution is to copy all of Commons, which is
overkill for many applications.  It would be nice if the dump
generator had a way of parsing out only the relevant Commons content.

-Robert Rohde

_______________________________________________
Wikitech-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Reply via email to