Yaron Koren wrote:
> Hi,
> 
> My extension External Data uses Http::get() to retrieve the contents of a
> URL. Unfortunately, it turns out that this function fails on some URLs,
> returning an HTTP return code of "0". Here's an example of a URL for which
> it doesn't work:
> 
> http://mouse.brain-map.org/GeneExpression/Hippocampal+region/1.xml
> 
> Http::get() uses PHP's cURL library when possible (and on my server, it's
> possible), and I assume that's what's causing the problem: when I replace
> the call to Http::get() with a call to file_get_contents(), it works fine.
> Does anyone know what the exact problem is? (It might be related to the fact
> that the page is in XML format, although other XML pages work.) Should I
> just use file_get_contents() instead?
> 
> Thanks,
> Yaron

Try increasing $wgHTTPTimeout
It's a big file, so it's probably timeouting before fully transferring
it, as it happened here. file_get_contents doesn't have a timeout,
unless specifically set with a context.


_______________________________________________
MediaWiki-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l

Reply via email to