If you make the compression external this is already done. In order to do what the poster requires, you still need to read and update fields without reading the entire document. You just do this at a binary field level, and do all of he compression/decompression externally.

I think putting the compression into Lucene needlessly complicates matters. All that is required is in place field updating, and binary field support.


On Aug 11, 2006, at 6:07 AM, Michael McCandless wrote:


I don't understand why the compressed fields are not just handled externally in the Document class - just add uncompress/compress methods. This way all Lucene needs to understand is binary fields, and you don't have any of these problems during merging or initial indexing.

The original poster of this issue (on java-user) raised another aspect of his use case: he needs to update documents that have large compressed fields.

Ideally, one could pull out a Document, change one of the other (not compressed) fields, then re-index it (and delete the original one), all without uncompresssing / recompressing the untouched compressed field. I guess this would require ability to mark a Field as compressed but also mark that it's already in compressed form (to avoid re-compressing it!).

It's of course always possible as a workaround do to all of this outside of Lucene (as this original poster has done), but, I think these use cases should be "in scope" for Lucene: if we are going to offer compressed fields we should try hard to make it efficient for basic use cases as well as for document updates?

Mike

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]



---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to