Probably because if the upload server tried to fail a large upload, the 
client would just retry the file.

In SMTP, there is a similar bug with many versions of Microsoft Outlook. 
  If the server tries to reject a message after the DATA statement 
(which is allowed) Outlook will immediately resend it.

Taking the message makes Outlook stop sending, and it could be the same 
here.

-- Lynn

Mark Pottorff wrote:
> While everyone seems to now be interested in such things... is there any 
> reason why files that exceed the maximum upload size are pulled in their 
> entirety through the pipe, and sent to null??
> 
> I mean had this been an ACTUAL DOS... where someone spoofs legitimate looking 
> file info, with 3GB file sizes...
> 
> See /sched/file_upload_handler.cpp
> line 339
> 339            if (!config.ignore_upload_certificates) { 
> 340                 if (nbytes > file_info.max_nbytes) { 
> 341                     sprintf(buf, 
> 342                         "file size (%d KB) exceeds limit (%d KB)", 
> 343                         (int)(nbytes/1024), 
> (int)(file_info.max_nbytes/1024) 
> 344                     ); 
> 345                     copy_socket_to_null(in); 
> 346                     return return_error(ERR_PERMANENT, buf); 
> 347                 } 
> 
> Why read all the data?? Can't the response just be sent (as if the hacker 
> cares about a response), and the socket closed?
> 
> * * * If any projects are interested in implementing the front-end, off 
> network, buffer uploads (or downloads) tiered sort of scheme, please let me 
> know. * * *
> 
> In SETI's case, just having a server receiving uploads during the weekly 
> backup window would help keep work flowing more smoothly and avoid a weekly 
> congestion period (perhaps that is done already). It will also keep the 
> non-permenant connection users happy. Whenever they choose go online, there 
> will be an upload server available.
> 
> In the case where multiple upload URLs exist on tasks, how are file sizes 
> determined accurately? Wouldn't all of the upload servers have to be polled 
> to see if they are the one that actually received the file? Or, I mean, at 
> least poll through the list of servers until the file is found? Perhaps the 
> assumption is that all servers are sharing the same network storage system? 
> That doesn't seem very robust, nor flexible.
> 
> Same question applies to an interrupted upload. Won't it have to be continued 
> on the same server that it got started with? I've not seen code that appears 
> to support this. Are multiple upload URLs even supported? Or does it always 
> have to be done behind a single URL? Perhaps it is handled on the client 
> side. What if upload to server 1 fails, upload to server 2 gets started and 
> is then interrupted, and then server 2 goes down before the rest of the file 
> is received? How to continue upload or recover from this state?
> 
> 
> Running Microsoft's "System Idle Process" will never help cure cancer,
>  AIDS nor Alzheimer's. But running rose...@home just might!
> http://boinc.bakerlab.org/rosetta/
> 
> 
>       
> _______________________________________________
> boinc_dev mailing list
> [email protected]
> http://lists.ssl.berkeley.edu/mailman/listinfo/boinc_dev
> To unsubscribe, visit the above URL and
> (near bottom of page) enter your email address.
> 
_______________________________________________
boinc_dev mailing list
[email protected]
http://lists.ssl.berkeley.edu/mailman/listinfo/boinc_dev
To unsubscribe, visit the above URL and
(near bottom of page) enter your email address.

Reply via email to