Anyone?

Thanks.

--- Rahul Joshi <[EMAIL PROTECTED]> wrote:

> Hello!
> 
> I have Wget version 1.10.2 running on PC (Cygwin on
> Windows XP; 768 MB RAM; 1150 MB virtual memory). I
> was
> downloading files from some URLs when one of the
> files
> was unexpectedly large and I got the following:  
> 
> Command and subsequent Error: 
> 
> =============
> $ wget --execute=input=c:/test.txt
> --execute=dir_prefix=c:/data
> 15:33:57
>
URL:http://goldenspud.com/webrog/archives/category/geek-stuff/python/
> 
> [335607029] -> "c:/data/index.html" [1]
> wget: realloc: Failed to allocate 536870912 bytes;
> memory exhausted.
> =============
> 
> The above URL is the only entry in the input file:
> test.txt. An index.html file of size 327,720 KB was
> downloaded in "c:/data" directory. In another
> attempt,
> the size of the same file downloaded was 420,609 KB
> but with the same error.
> 
> Firstly, this file size is less than the 2 GB limit;
> so ideally there should not be any problem. Should I
> change any settings?
> 
> Secondly, actually I do not want to download this
> large file. But as far as I know there is no way to
> skip files above some size. Can you suggest me any 
> quick change to the code to skip large files?
> 
> Thirdly, if I have many URL entries in the input
> file,
> after this URL and the error, the wget process ends.
> Is there a way to "catch" this error and continue?
> 
> Any help is appreciated!
> 
> Thanks,
> Rahul.



                
__________________________________ 
Yahoo! FareChase: Search multiple travel sites in one click.
http://farechase.yahoo.com

Reply via email to