[Wikitech-l] Collection extension

2009-04-10 Thread Bence Damokos
Hi,
Is there a planned deployment date for the Collection (PediaPress) extension
on Wikimedia wikis that currently don't have them? The blog post about it
[1] mentioned March if everything went well.

Best regards,
Bence


[1]
http://blog.wikimedia.org/2009/02/20/wiki-to-print-feature-activated-in-six-more-wikipedia-languages/
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


[Wikitech-l] Dealing with Large Files when attempting a wikipedia database download.

2009-04-10 Thread Jameson Scanlon
Does anyone on the wikitech mailing list happen to know whether it
would be possible for some of the larger wikipedia database downloads
(which are, say, 16GB or so in size) to be split into parts so that
they can be downloaded.  For whatever reason, whenever I have
attempted to download the ~14GB files (say, from
http://static.wikipedia.org/downloads/2008-06/en/ ), I have found that
only 2GB (presumably, the first 2GB) of what I have sought to download
has actually been downloaded.  Is there anyway around this?  Could
anyone possibly suggest what possible reasons there might be for this
difficulty in downloading the material?

Thanks. .

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Dealing with Large Files when attempting a wikipedia database download.

2009-04-10 Thread David Gerard
2009/4/10 Jameson Scanlon jameson.scan...@googlemail.com:

 Does anyone on the wikitech mailing list happen to know whether it
 would be possible for some of the larger wikipedia database downloads
 (which are, say, 16GB or so in size) to be split into parts so that
 they can be downloaded.  For whatever reason, whenever I have
 attempted to download the ~14GB files (say, from
 http://static.wikipedia.org/downloads/2008-06/en/ ), I have found that
 only 2GB (presumably, the first 2GB) of what I have sought to download
 has actually been downloaded.  Is there anyway around this?  Could
 anyone possibly suggest what possible reasons there might be for this
 difficulty in downloading the material?


Downloading to a filesystem that only does maximum 2GB files?


- d.

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Dealing with Large Files when attempting a wikipedia database download.

2009-04-10 Thread Daniel Kinzler
David Gerard schrieb:
 2009/4/10 Jameson Scanlon jameson.scan...@googlemail.com:
 
 Does anyone on the wikitech mailing list happen to know whether it
 would be possible for some of the larger wikipedia database downloads
 (which are, say, 16GB or so in size) to be split into parts so that
 they can be downloaded.  For whatever reason, whenever I have
 attempted to download the ~14GB files (say, from
 http://static.wikipedia.org/downloads/2008-06/en/ ), I have found that
 only 2GB (presumably, the first 2GB) of what I have sought to download
 has actually been downloaded.  Is there anyway around this?  Could
 anyone possibly suggest what possible reasons there might be for this
 difficulty in downloading the material?
 
 
 Downloading to a filesystem that only does maximum 2GB files?
 

Also, several http clients don't like files over 2GB - this is because the large
number of bytes in the Length field causes an integer overflow (2GB is the 31
bit limit). wget likes to die with a segmentation fault on those. I found that
curl works.

But of course, the file system also has to support very large files, as Gerard 
said.

Finally: yes, it would be nive to have such dumps available in pieces of perhaps
1GB in size.

-- daniel

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Dealing with Large Files when attempting a wikipedia database download.

2009-04-10 Thread Finne Boonen
http://en.wikipedia.org/wiki/Wikipedia_database has some information
on how to deal with the large files

henna

On Fri, Apr 10, 2009 at 21:43, Daniel Kinzler dan...@brightbyte.de wrote:
 David Gerard schrieb:
 2009/4/10 Jameson Scanlon jameson.scan...@googlemail.com:

 Does anyone on the wikitech mailing list happen to know whether it
 would be possible for some of the larger wikipedia database downloads
 (which are, say, 16GB or so in size) to be split into parts so that
 they can be downloaded.  For whatever reason, whenever I have
 attempted to download the ~14GB files (say, from
 http://static.wikipedia.org/downloads/2008-06/en/ ), I have found that
 only 2GB (presumably, the first 2GB) of what I have sought to download
 has actually been downloaded.  Is there anyway around this?  Could
 anyone possibly suggest what possible reasons there might be for this
 difficulty in downloading the material?


 Downloading to a filesystem that only does maximum 2GB files?


 Also, several http clients don't like files over 2GB - this is because the 
 large
 number of bytes in the Length field causes an integer overflow (2GB is the 31
 bit limit). wget likes to die with a segmentation fault on those. I found that
 curl works.

 But of course, the file system also has to support very large files, as 
 Gerard said.

 Finally: yes, it would be nive to have such dumps available in pieces of 
 perhaps
 1GB in size.

 -- daniel

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l





-- 
Maybe you knew early on that your track went from point A to B, but
unlike you I wasn't given a map at birth! Alyssa, Chasing Amy

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Dealing with Large Files when attempting a wikipedia database download.

2009-04-10 Thread Bilal Abdul Kader
I have downloaded the history dump file (~150 GB) using Firefox on XP and
using wget on Ubuntu and it works fine. I have downloaded it using a
download manager on Vista and it is fine also.

A more probable reason is the file system limitations.

bilal


On Fri, Apr 10, 2009 at 3:49 PM, Finne Boonen hen...@gmail.com wrote:

 http://en.wikipedia.org/wiki/Wikipedia_database has some information
 on how to deal with the large files

 henna

 On Fri, Apr 10, 2009 at 21:43, Daniel Kinzler dan...@brightbyte.de
 wrote:
  David Gerard schrieb:
  2009/4/10 Jameson Scanlon jameson.scan...@googlemail.com:
 
  Does anyone on the wikitech mailing list happen to know whether it
  would be possible for some of the larger wikipedia database downloads
  (which are, say, 16GB or so in size) to be split into parts so that
  they can be downloaded.  For whatever reason, whenever I have
  attempted to download the ~14GB files (say, from
  http://static.wikipedia.org/downloads/2008-06/en/ ), I have found that
  only 2GB (presumably, the first 2GB) of what I have sought to download
  has actually been downloaded.  Is there anyway around this?  Could
  anyone possibly suggest what possible reasons there might be for this
  difficulty in downloading the material?
 
 
  Downloading to a filesystem that only does maximum 2GB files?
 
 
  Also, several http clients don't like files over 2GB - this is because
 the large
  number of bytes in the Length field causes an integer overflow (2GB is
 the 31
  bit limit). wget likes to die with a segmentation fault on those. I found
 that
  curl works.
 
  But of course, the file system also has to support very large files, as
 Gerard said.
 
  Finally: yes, it would be nive to have such dumps available in pieces of
 perhaps
  1GB in size.
 
  -- daniel
 
  ___
  Wikitech-l mailing list
  Wikitech-l@lists.wikimedia.org
  https://lists.wikimedia.org/mailman/listinfo/wikitech-l
 

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Dealing with Large Files when attempting a wikipedia database download.

2009-04-10 Thread Brian
I'm pretty sure it's impossible to encourage people to include relevant
information in their OPs.
You don't suppose you could have at least told us your operating system,
whether you are running 32 or 64 bits?

Are you on linux with no large file support?

On Fri, Apr 10, 2009 at 12:21 PM, Jameson Scanlon 
jameson.scan...@googlemail.com wrote:

 Does anyone on the wikitech mailing list happen to know whether it
 would be possible for some of the larger wikipedia database downloads
 (which are, say, 16GB or so in size) to be split into parts so that
 they can be downloaded.  For whatever reason, whenever I have
 attempted to download the ~14GB files (say, from
 http://static.wikipedia.org/downloads/2008-06/en/ ), I have found that
 only 2GB (presumably, the first 2GB) of what I have sought to download
 has actually been downloaded.  Is there anyway around this?  Could
 anyone possibly suggest what possible reasons there might be for this
 difficulty in downloading the material?

 Thanks. .

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l