4GB for the operating system to use to buffer disk files. That is not a Solr setting.
wunder On 4/16/08 11:05 AM, "oleg_gnatovskiy" <[EMAIL PROTECTED]> wrote: > > Hello. I am having a similar problem as the OP. I see that you recommended > setting 4GB for the index, and 2 for Solr. How do I allocate memory for the > index? I was under the impression that Solr did not support a RAMIndex. > > > Walter Underwood wrote: >> >> Do it. 32-bit OS's went out of style five years ago in server-land. >> >> I would start with 8GB of RAM. 4GB for your index, 2 for Solr, 1 for >> the OS and 1 for other processes. That might be tight. 12GB would >> be a lot better. >> >> wunder >> >> On 4/16/08 7:50 AM, "Jonathan Ariel" <[EMAIL PROTECTED]> wrote: >> >>> In order to do that I have to change to a 64 bits OS so I can have more >>> than >>> 4 GB of RAM.Is there any way to see how long does it takes to Solr to >>> warmup >>> the searcher? >>> >>> On Wed, Apr 16, 2008 at 11:40 AM, Walter Underwood >>> <[EMAIL PROTECTED]> >>> wrote: >>> >>>> A commit every two minutes means that the Solr caches are flushed >>>> before they even start to stabilize. Two things to try: >>>> >>>> * commit less often, 5 minutes or 10 minutes >>>> * have enough RAM that your entire index can fit in OS file buffers >>>> >>>> wunder >>>> >>>> On 4/16/08 6:27 AM, "Jonathan Ariel" <[EMAIL PROTECTED]> wrote: >>>> >>>>> So I counted the number if distinct values that I have for each field >>>> that I >>>>> want a facet on. In total it's around 100,000. I tried with a >>>> filterCache >>>>> of 120,000 but it seems like too much because the server went down. I >>>> will >>>>> try with less, around 75,000 and let you know. >>>>> >>>>> How do you to partition the data to a static set and a dynamic set, and >>>> then >>>>> combining them at query time? Do you have a link to read about that? >>>>> >>>>> >>>>> >>>>> On Tue, Apr 15, 2008 at 7:21 PM, Mike Klaas <[EMAIL PROTECTED]> >>>> wrote: >>>>> >>>>>> On 15-Apr-08, at 5:38 AM, Jonathan Ariel wrote: >>>>>> >>>>>>> My index is 4GB on disk. My servers has 8 GB of RAM each (the OS is >>>>>>> 32 >>>>>>> bits). >>>>>>> It is optimized twice a day, it takes around 15 minutes to optimize. >>>>>>> The index is updated (commits) every two minutes. There are between >>>>>>> 10 >>>>>>> and >>>>>>> 100 inserts/updates every 2 minutes. >>>>>>> >>>>>> >>>>>> Caching could help--you should definitely start there. >>>>>> >>>>>> The commit every 2 minutes could end up being an unsurmountable >>>> problem. >>>>>> You may have to partition your data into a large, mostly static set >>>> and a >>>>>> small dynamic set, combining the results at query time. >>>>>> >>>>>> -Mike >>>>>> >>>> >>>> >> >> >>