Hi! Platonides wrote: > seth wrote: > >>I wrote a perl script, which works on some HTML content of some >>wikipedia-webpages. Some of those pages are >300kB and perls LWP-mirror >>hangs up.
I was wrong. LWP-mirror did not hang up, but the content was not fully loaded, because of caching. After I purged the site manually, everything was ok. >>1. Is there a better/faster way to get the HTML content of e.g. >>http://meta.wikimedia.org/wiki/Spam_blacklist/Log >>than >> my $ua = LWP::UserAgent->new; >> $ua->mirror($url, $filename); >>? > > To get the content of wikipedia pages you should be using WikiProxy > http://meta.wikimedia.org/wiki/User:Duesentrieb/WikiProxy Does this tool purge automatically? Is there any manual for that tool? bye seth _______________________________________________ Toolserver-l mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/toolserver-l
