[ 
https://issues.apache.org/jira/browse/LUCENE-1591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12698211#action_12698211
 ] 

Shai Erera commented on LUCENE-1591:
------------------------------------

That's the way I wrap FIS with BZIP:

{code}
      if (doBzipCompression) {
        // According to CBZip2InputStream's documentation, we should first
        // consume the first two file header chars ('B' and 'Z'), as well as 
        // wrap the underlying stream with a BufferedInputStream, since CBZip2IS
        // uses the read() method exclusively.
        fileIS = new BufferedInputStream(fileIS, READER_BUFFER_BYTES);
        fileIS.read(); fileIS.read();
        fileIS = new CBZip2InputStream(fileIS);
      }
{code}

bq. Is it possible your bunzipping code is messing up the XML?

I successfully read the file and compressed it with Java's GZIP classes, 
however I did not attempt to parse the XML itself. Did you run EnwikiDocMaker 
on the actual XML or the bz2 archive?
The 20070527 run should end soon (I hope - it reached 2.2M documents, so if it 
doesn't fail, I guess that bzip wrapping is very unlikely to affect the XML 
parsing.

bq. Shai, why did it take 9 hours to get to that exception? Is bunzip that 
slow? That seems crazy. 

I run the test on my TP 60, which is not a snail-of-a-machine, but definitely 
not a strong server. You can download the patch and the jar and try it out on 
your machine.
But yes, I did notice bzip is very slow compared to gzip, however it has better 
compression ration. I do want to measure the times though, to give more 
accurate numbers, but in order to do that I need to finish a successful run 
first.

bq. Can you run only your bunzip code and confirm it ...

I would have done that, but the output XML is 17GB, and doing it twice is not 
an option on my TP. That's why I wanted this bzip thing in the first place :)
I'll try to do that with the 20070527 version, which hopefully will be ~half 
the size ... 


> Enable bzip compression in benchmark
> ------------------------------------
>
>                 Key: LUCENE-1591
>                 URL: https://issues.apache.org/jira/browse/LUCENE-1591
>             Project: Lucene - Java
>          Issue Type: Improvement
>          Components: contrib/benchmark
>            Reporter: Shai Erera
>             Fix For: 2.9
>
>         Attachments: ant-1.7.1.jar, LUCENE-1591.patch
>
>
> bzip compression can aid the benchmark package by not requiring extracting 
> bzip files (such as enwiki) in order to index them. The plan is to add a 
> config parameter bzip.compression=true/false and in the relevant tasks either 
> decompress the input file or compress the output file using the bzip streams.
> It will add a dependency on ant.jar which contains two classes similar to 
> GZIPOutputStream and GZIPInputStream which compress/decompress files using 
> the bzip algorithm.
> bzip is known to be superior in its compression performance to the gzip 
> algorithm (~20% better compression), although it does the 
> compression/decompression a bit slower.
> I wil post a patch which adds this parameter and implement it in 
> LineDocMaker, EnwikiDocMaker and WriteLineDoc task. Maybe even add the 
> capability to DocMaker or some of the super classes, so it can be inherited 
> by all sub-classes.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: java-dev-h...@lucene.apache.org

Reply via email to