Greetings

are there any plans to implement compression of the variable length bites[] binary doc Values,
say in blocks of 16k like for stored values ?

my .cfs file goes from 2MB to like 400k when I zip it

Best regards,
Olivier



On 08/08/2015 02:32 PM, jamie wrote:
Greetings

Our app primarily uses Lucene for its intended purpose i.e. to search across large amounts of unstructured text. However, recently our requirement expanded to perform look-ups on specific documents in the index based on associated custom defined unique keys. For our purposes, a unique key is the string representation of a 128 bit murmur hash, stored in a Lucene field named uid. We are currently using the TermsFilter to lookup Documents in the Lucene index as follows:

List<Term> terms = new LinkedList<>();
            for (String id : ids) {
                terms.add(new Term("uid", id));
}
TermsFilter idFilter = new TermsFilter(terms);
... search logic...

At any time we may need to lookup say a couple of thousand documents. Our problem is one of performance. On very large indexes with 30 million records or more, the lookup can be excruciatingly slow. At this stage, its not practical for us to move the data over to fit for purpose database, nor change the uid field to a numeric type. I fully appreciate the fact that Lucene is not designed to be a database, however, is there anything we can do to improve the performance of these look-ups?

Much appreciate

Jamie




---------------------------------------------------------------------
To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org
For additional commands, e-mail: java-user-h...@lucene.apache.org

Reply via email to