On Sat, Apr 28, 2001 at 06:48:58PM -0500, Scott G. Miller wrote: > > > > One solution would be to have a minimum compression threshold. Files > > under this threshold would be uncompressed, and files larger than this > > threshold would be compressed. This would result in there be space > > and bandwidth savings for big files, but not slowdowns from > > compressing small files. > > But thats likely to mean that no files would be compressed, as nearly all > decently sized files are already compressed with some scheme (jpeg, mpeg, > etc). Me think that you guys just like coding too much. > > Scott This is a special case - text or HTML of a reasonable size that should be compressed to save space and hence increase its likelihood of survival on the network. It would not be sensible for it to occur automatically, unless if only with particular MIME types, but decoding support in freenet_request and fproxy might be worthwhile. Medium sized HTML pages are quite compressible, as evidenced by the W3C performance paper. And it's not just about bandwidth. And sticking stuff in ZIPs is not a good answer for pages which people would like to refer to and browse individually. The size bias means that it will signigicantly improve short term survival of pages.
--
