I wonder.  What if, instead of 200 instances, you had one instance, but built a 
uniqueKey up out of whatever you have now plus whatever information currently 
segregates the instances.  Then this would be much more manageable.

In other words, what is different about each of the 200 instances?  Is the 
schema for each essentially the same, as I am guessing?

JRJ

-----Original Message-----
From: Sujatha Arun [mailto:suja.a...@gmail.com] 
Sent: Thursday, October 20, 2011 12:21 AM
To: solr-user@lucene.apache.org
Cc: Otis Gospodnetic
Subject: Re: OS Cache - Solr

Yes 200 Individual Solr Instances not solr cores.

We get an avg response time of below 1 sec.

The number of documents is  not many most of the isntances ,some of the
instnaces have about 5 lac documents on average.

Regards
Sujahta

On Thu, Oct 20, 2011 at 3:35 AM, Jaeger, Jay - DOT <jay.jae...@dot.wi.gov>wrote:

> 200 instances of what?  The Solr application with lucene, etc. per usual?
>  Solr cores? ???
>
> Either way, 200 seems to be very very very many: unusually so.  Why so
> many?
>
> If you have 200 instances of Solr in a 20 GB JVM, that would only be 100MB
> per Solr instance.
>
> If you have 200 instances of Solr all accessing the same physical disk, the
> results are not likely to be satisfactory - the disk head will go nuts
> trying to handle all of the requests.
>
> JRJ
>
> -----Original Message-----
> From: Sujatha Arun [mailto:suja.a...@gmail.com]
> Sent: Wednesday, October 19, 2011 12:25 AM
> To: solr-user@lucene.apache.org; Otis Gospodnetic
> Subject: Re: OS Cache - Solr
>
> Thanks ,Otis,
>
> This is our Solr Cache  Allocation.We have the same Cache allocation for
> all
> our *200+ instances* in the single Server.Is this too high?
>
> *Query Result Cache*:LRU Cache(maxSize=16384, initialSize=4096,
> autowarmCount=1024, )
>
> *Document Cache *:LRU Cache(maxSize=16384, initialSize=16384)
>
>
> *Filter Cache* LRU Cache(maxSize=16384, initialSize=4096,
> autowarmCount=4096, )
>
> Regards
> Sujatha
>
> On Wed, Oct 19, 2011 at 4:05 AM, Otis Gospodnetic <
> otis_gospodne...@yahoo.com> wrote:
>
> > Maybe your Solr Document cache is big and that's consuming a big part of
> > that JVM heap?
> > If you want to be able to run with a smaller heap, consider making your
> > caches smaller.
> >
> > Otis
> > ----
> > Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch
> > Lucene ecosystem search :: http://search-lucene.com/
> >
> >
> > >________________________________
> > >From: Sujatha Arun <suja.a...@gmail.com>
> > >To: solr-user@lucene.apache.org
> > >Sent: Tuesday, October 18, 2011 12:53 AM
> > >Subject: Re: OS Cache - Solr
> > >
> > >Hello Jan,
> > >
> > >Thanks for your response and  clarification.
> > >
> > >We are monitoring the JVM cache utilization and we are currently using
> > about
> > >18 GB of the 20 GB assigned to JVM. Out total index size being abt 14GB
> > >
> > >Regards
> > >Sujatha
> > >
> > >On Tue, Oct 18, 2011 at 1:19 AM, Jan Høydahl <jan....@cominvent.com>
> > wrote:
> > >
> > >> Hi Sujatha,
> > >>
> > >> Are you sure you need 20Gb for Tomcat? Have you profiled using
> JConsole
> > or
> > >> similar? Try with 15Gb and see how it goes. The reason why this is
> > >> beneficial is that you WANT your OS to have available memory for disk
> > >> caching. If you have 17Gb free after starting Solr, your OS will be
> able
> > to
> > >> cache all index files in memory and you get very high search
> > performance.
> > >> With your current settings, there is only 12Gb free for both caching
> the
> > >> index and for your MySql activities.  Chances are that when you backup
> > >> MySql, the cached part of your Solr index gets flushed from disk
> caches
> > and
> > >> need to be re-cached later.
> > >>
> > >> How to interpret memory stats vary between OSes, and seing 163Mb free
> > may
> > >> simply mean that your OS has used most RAM for various caches and
> > paging,
> > >> but will flush it once an application asks for more memory. Have you
> > seen
> > >> http://wiki.apache.org/solr/SolrPerformanceFactors ?
> > >>
> > >> You should also slim down your index maximally by setting stored=false
> > and
> > >> indexed=false wherever possible. I would also upgrade to a more
> current
> > Solr
> > >> version.
> > >>
> > >> --
> > >> Jan Høydahl, search solution architect
> > >> Cominvent AS - www.cominvent.com
> > >> Solr Training - www.solrtraining.com
> > >>
> > >> On 17. okt. 2011, at 19:51, Sujatha Arun wrote:
> > >>
> > >> > Hello
> > >> >
> > >> > I am trying to understand the  OS cache utilization of Solr .Our
> > server
> > >> has
> > >> > several solr instances on a server .The total combined Index size of
> > all
> > >> > instances is abt 14 Gb and the size of the maximum single Index is
> abt
> > >> 2.5
> > >> > GB .
> > >> >
> > >> > Our Server has Quad processor with 32 GB RAM .Out of which 20 GB has
> > been
> > >> > assigned to  JVM. We are running solr1.3  on tomcat 5.5 and Java 1.6
> > >> >
> > >> > Our current Statistics indicate that  solr uses 18-19 GB of 20 GB
> RAM
> > >> > assigned to JVM .However the  Free physical seems to remain constant
> > as
> > >> > below.
> > >> > Free physical memory = 163 Mb
> > >> > Total physical memory = 32,232 Mb,
> > >> >
> > >> > The server also serves as a backup server for Mysql where the
> > application
> > >> DB
> > >> > is backed up and restored .During this activity we see that lot of
> > >> queries
> > >> > that nearly take even 10+ minutes to execute .But other wise
> > >> > maximum query time is less than  1-2 secs
> > >> >
> > >> > The physical memory that is free seems to be constant . Why is this
> > >> constant
> > >> > and how this will be used between the  Mysql backup and solr while
> > >> > backup activity is  happening How much free physical memory should
> be
> > >> > available to OS given out stats.?
> > >> >
> > >> > Any pointers would be helpful.
> > >> >
> > >> > Regards
> > >> > Sujatha
> > >>
> > >>
> > >
> > >
> > >
> >
>

Reply via email to