About a factor of 2 on a small, optimized index. Gzipping took 20 seconds, so it isn't free.
$ cd index-copy $ du -sk 134336 . $ gzip * $ du -sk 62084 . wunder On 10/30/08 8:20 AM, "Otis Gospodnetic" <[EMAIL PROTECTED]> wrote: > Yeah. I'm just not sure how much benefit in terms of data transfer this will > save. Has anyone tested this to see if this is even worth it? > > > Otis > -- > Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch > > > > ----- Original Message ---- >> From: Erik Hatcher <[EMAIL PROTECTED]> >> To: solr-user@lucene.apache.org >> Sent: Thursday, October 30, 2008 9:54:28 AM >> Subject: Re: replication handler - compression >> >> +1 - the GzipServletFilter is the way to go. >> >> Regarding request handlers reading HTTP headers, yeah,... this will improve, >> for >> sure. >> >> Erik >> >> On Oct 30, 2008, at 12:18 AM, Chris Hostetter wrote: >> >>> >>> : You are partially right. Instead of the HTTP header , we use a request >>> : parameter. (RequestHandlers cannot read HTP headers). If the param is >>> >>> hmmm, i'm with walter: we shouldn't invent new mechanisms for >>> clients to request compression over HTTP from servers. >>> >>> replicatoin is both special enough and important enough that if we had to >>> add special support to make that information available to the handler on >>> the master we could. >>> >>> but frankly i don't think that's neccessary: the logic to turn on >>> compression if the client requests it using "Accept-Encoding: gzip" is >>> generic enough that there is no reason for it to be in a handler. we >>> could easily put it in the SolrDispatchFilter, or even in a new >>> ServletFilte (i'm guessing iv'e seen about 74 different implementations of >>> a GzipServletFilter in the wild that could be used as is. >>> >>> then we'd have double wins: compression for replication, and compression >>> of all responses generated by Solr if hte client requests it. >>> >>> -Hoss >