On 11/5/2013 3:36 PM, Scott Schneider wrote:
I'm wondering if filter queries are efficient enough for my use cases.  I have 
lots and lots of users in a big, multi-tenant, sharded index.  To run a search, 
I can use an fq on the user id and pass in the search terms.  Does this scale 
well with the # users?  I suppose that, since user id is indexed, generating 
the filter data (which is cached) will be fast.  And looking up search terms is 
fast, of course.  But if the search term is a common one that many users have 
in their documents, then Solr may have to perform an intersection between two 
large sets:  docs from all users with the search term and all of the current 
user's docs.

Also, how about auto-complete and searching with a trailing wildcard?  As I understand it, these work well in 
a single-tenant index because keywords are sorted in the index, so it's easy to get all the search terms that 
match "foo*".  In a multi-tenant index, all users' keywords are stored together.  So if Lucene were 
to look at all the keywords from "foo" to "foozzzzz" (I'm not sure if it actually does 
this), it would skip over a large majority of keywords that don't belong to this user.

From what I understand, there's not really a whole lot of difference between queries and filter queries when they are NOT cached, except that the main query and the filter queries are executed in parallel, which can save time.

When filter queries are found in the filterCache, it's a different story. They get applied *before* the main query, which means that the main query won't have to work as hard. The filterCache stores information about which documents in the entire index match the filter. By storing it as a bitset, the amount of space required is relatively low. Applying filterCache results is very efficient.

There are also advanced techniques, like assigning a cost to each filter and creating postfilters:

http://yonik.com/posts/advanced-filter-caching-in-solr/

Thanks,
Shawn

Reply via email to