Hi,

I have a sharded collection distributed over several solr nodes. Each solr
node hosts one shard and one replica of another shard. shards are huge (100
millions documents). Queries are using several filterQuery. filterCache for
this number of documents can use high amount of heap memory.

Is it a good idea to split shards by 2 or 4 in order to have shards with 50
or 25 millions documents ?
With a split by 4, a Solr node will host 8 replicas instead of 2, but with
smaller filterCache for each replica.

I don't expect to have better search performances, but I expect to have
faster warming and mainly less impacted heap memory by open searcher during
sofcommit.  For instance, instead of having one large filterCache warming
up once each minute, 4 smaller filterCaches will warm up not at the same
time (hopefully).

So shard splitting or not on the same Solr nodes ?

Thank you

Dominique

Reply via email to