Thanks a lot for the responses, after the optimize is complete and i have
some time to experiment ill throw some of these settings in place,

On Tue, Jul 25, 2017 at 4:39 PM, Walter Underwood <wun...@wunderwood.org>
wrote:

> I’ve never been fond of elaborate GC settings. I prefer to set a few
> things then let it run. I know someone who’s worked on garbage collectors
> for thirty years. I don’t second guess him.
>
> From watching GC performance under a load benchmark (CMS/ParNew) with Solr
> 4.x, I increased the new space. Individual requests make a lot of
> allocations that are garbage at the end of the request. All of those need
> to come from the new space. If new space is not big enough, they’ll be
> allocated from tenured space. I settled on an 8G heap with 2G of new space.
> These are the options (Java 7):
>
> export CATALINA_OPTS="$CATALINA_OPTS -d64"
> export CATALINA_OPTS="$CATALINA_OPTS -server"
> export CATALINA_OPTS="$CATALINA_OPTS -Xms8g"
> export CATALINA_OPTS="$CATALINA_OPTS -Xmx8g"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:NewSize=2g"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:+UseConcMarkSweepGC"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:+UseParNewGC"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:+ExplicitGCInvokesConcurrent"
> export CATALINA_OPTS="$CATALINA_OPTS -verbose:gc"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:+PrintGCDetails"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:+PrintGCTimeStamps"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:-TraceClassUnloading"
> export CATALINA_OPTS="$CATALINA_OPTS -Xloggc:$CATALINA_HOME/logs/gc.log"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:+HeapDumpOnOutOfMemoryError"
> export CATALINA_OPTS="$CATALINA_OPTS -XX:HeapDumpPath=$CATALINA_
> HOME/logs/“
>
> Tenured space will rise slowly, mostly because of cache ejections and
> background merges, I think. Cache ejections from an LRU cache will almost
> always be in tenured space.
>
> For Java 8 and Solr 6.5.1, we are running the G1 collector and very happy
> with it. We run the options recommended by Shawn Heisey on this list.
>
> SOLR_HEAP=8g
> # Use G1 GC  -- wunder 2017-01-23
> # Settings from https://wiki.apache.org/solr/ShawnHeisey
> GC_TUNE=" \
> -XX:+UseG1GC \
> -XX:+ParallelRefProcEnabled \
> -XX:G1HeapRegionSize=8m \
> -XX:MaxGCPauseMillis=200 \
> -XX:+UseLargePages \
> -XX:+AggressiveOpts \
> “
>
> Last week, I benchmarked the 4.x config handling 15,000 requests/minute
> with a 95th percentile response time of 30 ms, using production logs.
>
> wunder
> Walter Underwood
> wun...@wunderwood.org
> http://observer.wunderwood.org/  (my blog)
>
>
> > On Jul 25, 2017, at 1:24 PM, Markus Jelsma <markus.jel...@openindex.io>
> wrote:
> >
> > Upgrade to 6.x and get, in general, decent JVM settings. And decrease
> your heap, having it so extremely large is detrimental at best.
> >
> > Our shards can be 25 GB in size, but we run fine (apart from other
> problems recently discovered) with a 900 MB heap, so you probably have a
> lot of room to spare. Your max heap is over a 100 times larger than ours,
> your index just around 16 times. It should work with less.
> >
> > As a bonus, with a smaller heap, you can have much more index data in
> mapped memory.
> >
> > Regards,
> > Markus
> >
> > -----Original message-----
> >> From:David Hastings <hastings.recurs...@gmail.com>
> >> Sent: Tuesday 25th July 2017 22:15
> >> To: solr-user@lucene.apache.org
> >> Subject: Re: Optimize stalls at the same point
> >>
> >> it turned out that i think it was a large GC operation, as it has since
> >> resumed optimizing.  current java options are as follows for the
> indexing
> >> server (they are different for the search servers) if you have any
> >> suggestions as to changes I am more than happy to hear them, honestly
> they
> >> have just been passed down from one installation to the next ever since
> we
> >> used to use tomcat to host solr
> >> -server -Xss256k -d64 -Xmx100000m -Xms7000m-XX:NewRatio=3
> >> -XX:SurvivorRatio=4 -XX:TargetSurvivorRatio=90
> -XX:MaxTenuringThreshold=8
> >> -XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:ConcGCThreads=4
> >> -XX:ParallelGCThreads=8 -XX:+CMSScavengeBeforeRemark
> >> -XX:PretenureSizeThreshold=64m -XX:+UseCMSInitiatingOccupancyOnly
> >> -XX:CMSInitiatingOccupancyFraction=50 -XX:CMSMaxAbortablePrecleanTime=
> 6000
> >> -XX:+CMSParallelRemarkEnabled -XX:+ParallelRefProcEnabled -verbose:gc
> >> -XX:+PrintHeapAtGC -XX:+PrintGCDetails -XX:+PrintGCDateStamps
> >> -XX:+PrintGCTimeStamps -XX:+PrintTenuringDistribution
> >> -XX:+PrintGCApplicationStoppedTime
> >> -Xloggc:XXXXXX/solr-5.2.1/server/logs/solr_gc.log
> >>
> >> and for my live searchers i use:
> >> server Xss256k Xms50000m Xmx50000m XX:NewRatio=3 XX:SurvivorRatio=4
> >> XX:TargetSurvivorRatio=90 XX:MaxTenuringThreshold=8
> XX:+UseConcMarkSweepGC
> >> XX:+UseParNewGC XX:ConcGCThreads=4 XX:ParallelGCThreads=8
> >> XX:+CMSScavengeBeforeRemark XX:PretenureSizeThreshold=64m
> >> XX:+UseCMSInitiatingOccupancyOnly XX:CMSInitiatingOccupancyFraction=50
> >> XX:CMSMaxAbortablePrecleanTime=6000 XX:+CMSParallelRemarkEnabled
> >> XX:+ParallelRefProcEnabled verbose:gc XX:+PrintHeapAtGC
> XX:+PrintGCDetails
> >> XX:+PrintGCDateStamps XX:+PrintGCTimeStamps
> XX:+PrintTenuringDistribution
> >> XX:+PrintGCApplicationStoppedTime Xloggc:/SSD2TB01/solr
> >> 5.2.1/server/logs/solr_gc.log
> >>
> >>
> >>
> >> On Tue, Jul 25, 2017 at 4:02 PM, Walter Underwood <
> wun...@wunderwood.org>
> >> wrote:
> >>
> >>> Are you sure you need a 100GB heap? The stall could be a major GC.
> >>>
> >>> We run with an 8GB heap. We also run with Xmx equal to Xms, growing
> memory
> >>> to the max was really time-consuming after startup.
> >>>
> >>> What version of Java? What GC options?
> >>>
> >>> wunder
> >>> Walter Underwood
> >>> wun...@wunderwood.org
> >>> http://observer.wunderwood.org/  (my blog)
> >>>
> >>>
> >>>> On Jul 25, 2017, at 12:03 PM, David Hastings <
> >>> hastings.recurs...@gmail.com> wrote:
> >>>>
> >>>> I am trying to optimize a rather large index (417gb) because its
> sitting
> >>> at
> >>>> 28% deletions.  However when optimizing, it stops at exactly 492.24 GB
> >>>> every time.  When I restart solr it will fall back down to 417 gb, and
> >>>> again, if i send an optimize command, the exact same 492.24 GB and it
> >>> stops
> >>>> optimizing.  There is plenty of space on the drive, and im running it
> >>>> at -Xmx100000m -Xms7000m on a machine with 132gb of ram and 24
> cores.  I
> >>>> have never ran into this problem before but also never had the index
> get
> >>>> this large.  Any ideas?
> >>>> (solr 5.2 btw)
> >>>> thanks,
> >>>> -Dave
> >>>
> >>>
> >>
>
>

Reply via email to