Pierre Quentel added the comment: @martin.panter
For HTTP/1.0, since chunked transfer is not supported, and storage in a temporary file is also not an option, I see 2 possible solutions : - give up compressing big files - it would be a pity, compression is actually made for them... - compress the file 2 times : a first time just to compute the content length, without storing or sending anything, and a second time to send the gzipped data after all headers have been sent If there is a 3rd solution I'd be glad to know ; otherwise I prefer the second one, in spite of the waste of CPU. ---------- _______________________________________ Python tracker <rep...@bugs.python.org> <http://bugs.python.org/issue30576> _______________________________________ _______________________________________________ Python-bugs-list mailing list Unsubscribe: https://mail.python.org/mailman/options/python-bugs-list/archive%40mail-archive.com