Hi!

Platonides wrote:
> seth wrote:
> 
>>I wrote a perl script, which works on some HTML content of some 
>>wikipedia-webpages. Some of those pages are >300kB and perls LWP-mirror 
>>hangs up.

I was wrong. LWP-mirror did not hang up, but the content was not fully 
loaded, because of caching. After I purged the site manually, everything 
was ok.

>>1. Is there a better/faster way to get the HTML content of e.g. 
>>http://meta.wikimedia.org/wiki/Spam_blacklist/Log
>>than
>>   my $ua = LWP::UserAgent->new;
>>   $ua->mirror($url, $filename);
>>?
> 
> To get the content of wikipedia pages you should be using WikiProxy 
> http://meta.wikimedia.org/wiki/User:Duesentrieb/WikiProxy

Does this tool purge automatically? Is there any manual for that tool?

bye
seth

_______________________________________________
Toolserver-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/toolserver-l

Reply via email to