Hi, I have a script to download some large binary files from a web site. Everything works ok initially but for the bigger files (~100Mb) I am getting a timeout. I am guessing that this is an issue with the web server I am connecting to but I was wondering if there was something that I can change in my script to fix or get around this problem? Or is there a better way to do this? (I have pasted the script below) The error is get is "500 read timeout" Thanks for any help Adam ################################################## #! perl -w use strict; use LWP::UserAgent; my @files = ('Malaria110699_10PA.tar.gz', 'PfSpzS4_10PA.tar.gz'); for(my $i=0; $i < @files; $i++) { print "downloading ".$files[$i]."\t"; my $ua = new LWP::UserAgent; my $request = new HTTP::Request('GET', "<URL>"); my $response = $ua->request($request, "$files[$i]"); if($response->is_error()) { print "\n".$response->status_line."\n" } else{print "done\n"} } ##################################################