[ 
https://issues.apache.org/jira/browse/SOLR-2155?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12925001#action_12925001
 ] 

David Smiley commented on SOLR-2155:
------------------------------------

bq. Using the canonical geohash gives facet values that can be copy&pasted with 
other software. Thinking again, this is a great feature. Would it be worth 
optimizing geohash with a Trie version? Trie fields (can be made to) show up 
correctly in facets.

The geohash usage is purely internal to the implementation; users don't see it 
when they use this field.  And even if they were exposed, they can be generated 
on-demand.  There's even javascript code I've seen to do this.  So I'm not 
married to using geohashes -- it's the underlying heirarchical/gridded nature 
of them that is key.  I'm not sure how a "trie version" of geohash is 
developed.  I already spoke of further refining the implementation to index the 
geohashes at each grid level and I think that is very similar to what trie does 
for numbers.

Thanks for the suggestion of using OpenStreetMaps to get locations; I'll look 
into that.  I want to put together a useful data set -- using real data as much 
as possible is good.  I'll need to synthesize a one-to-many document to points 
mapping, randomly, however.  And I'll need to come up with various random 
lat-lon box queries to perform.  I'd like to use Lucene's benchmark contrib 
module as a framework to develop the performance test.  I read about it in LIA2 
and it seems to fit the bill.

> Geospatial search using geohash prefixes
> ----------------------------------------
>
>                 Key: SOLR-2155
>                 URL: https://issues.apache.org/jira/browse/SOLR-2155
>             Project: Solr
>          Issue Type: Improvement
>            Reporter: David Smiley
>         Attachments: GeoHashPrefixFilter.patch
>
>
> There currently isn't a solution in Solr for doing geospatial filtering on 
> documents that have a variable number of points.  This scenario occurs when 
> there is location extraction (i.e. via a "gazateer") occurring on free text.  
> None, one, or many geospatial locations might be extracted from any given 
> document and users want to limit their search results to those occurring in a 
> user-specified area.
> I've implemented this by furthering the GeoHash based work in Lucene/Solr 
> with a geohash prefix based filter.  A geohash refers to a lat-lon box on the 
> earth.  Each successive character added further subdivides the box into a 4x8 
> (or 8x4 depending on the even/odd length of the geohash) grid.  The first 
> step in this scheme is figuring out which geohash grid squares cover the 
> user's search query.  I've added various extra methods to GeoHashUtils (and 
> added tests) to assist in this purpose.  The next step is an actual Lucene 
> Filter, GeoHashPrefixFilter, that uses these geohash prefixes in 
> TermsEnum.seek() to skip to relevant grid squares in the index.  Once a 
> matching geohash grid is found, the points therein are compared against the 
> user's query to see if it matches.  I created an abstraction GeoShape 
> extended by subclasses named PointDistance... and CartesianBox.... to support 
> different queried shapes so that the filter need not care about these details.
> This work was presented at LuceneRevolution in Boston on October 8th.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to