This solution makes sense, but I still don't know if I can use solrCloud with this configuration or not.
On Fri, Mar 9, 2012 at 2:06 PM, Robert Stewart <bstewart...@gmail.com>wrote: > Split up index into say 100 cores, and then route each search to a > specific core by some mod operator on the user id: > > core_number = userid % num_cores > > core_name = "core"+core_number > > That way each index core is relatively small (maybe 100 million docs or > less). > > > On Mar 9, 2012, at 2:02 PM, Glen Newton wrote: > > > millions of cores will not work... > > ...yet. > > > > -glen > > > > On Fri, Mar 9, 2012 at 1:46 PM, Lan <dung....@gmail.com> wrote: > >> Solr has no limitation on the number of cores. It's limited by your > hardware, > >> inodes and how many files you could keep open. > >> > >> I think even if you went the Lucene route you would run into same > hardware > >> limits. > >> > >> -- > >> View this message in context: > http://lucene.472066.n3.nabble.com/Lucene-vs-Solr-design-decision-tp3813457p3813511.html > >> Sent from the Solr - User mailing list archive at Nabble.com. > > > > > > > > -- > > - > > http://zzzoot.blogspot.com/ > > - > > -- Alireza Salimi Java EE Developer