[
https://issues.apache.org/jira/browse/LUCENE-3069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13617982#comment-13617982
]
David Smiley commented on LUCENE-3069:
--------------------------------------
I'd love to see this come to pass. I've been thinking about what goes on a
layer beneath TermsEnum (i.e. how it is implemented) as I work on spatial
stuff. Geohash prefixes are a natural fit for FSTs; it should compress
ridiculously well. There is an approach to building a heatmap (spatial grid
faceting) that I'm thinking of that would do 2500 seek()'s for a 50x50 grid;
I'd like those seek's to be as fast as possible. I have another approach in
mind requiring a slightly different encoding, but it would do 2500 next()'s
which should be faster. Nonetheless; it's a lot -- ideally the terms dict
would be entirely memory resident.
> Lucene should have an entirely memory resident term dictionary
> --------------------------------------------------------------
>
> Key: LUCENE-3069
> URL: https://issues.apache.org/jira/browse/LUCENE-3069
> Project: Lucene - Core
> Issue Type: Improvement
> Components: core/index, core/search
> Affects Versions: 4.0-ALPHA
> Reporter: Simon Willnauer
> Assignee: Simon Willnauer
> Labels: gsoc2013
> Fix For: 4.3
>
>
> FST based TermDictionary has been a great improvement yet it still uses a
> delta codec file for scanning to terms. Some environments have enough memory
> available to keep the entire FST based term dict in memory. We should add a
> TermDictionary implementation that encodes all needed information for each
> term into the FST (custom fst.Output) and builds a FST from the entire term
> not just the delta.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]