Hi guys,
Using solr 3.6.1 and the following settings, I am trying to run without
optimizes. I used to optimize nightly, but sometimes the optimize took a very
long time to complete and slowed down our indexing. We are continuously
indexing our new or changed data all day and night. After a few days running
without an optimize, the index size has nearly doubled and maxdocs is nearly
twice the size of numdocs. I understand deletes should be expunged on merges,
but even after trying lots of different settings for our merge policy it seems
this growth is somewhat unbounded. I have tried sending an optimize with
numSegments = 2 which is a lot lighter weight then a regular optimize and that
does bring the number down but not by too much. Does anyone have any ideas for
better settings for my merge policy that would help? Here is my current index
snapshot too:
Location: /var/LucidWorks/lucidworks/solr/10000/data/index
Size: 25.05 GB (when the index is optimized it is around 15.5 GB)
searcherName : Searcher@6c3a3517 main
caching : true
numDocs : 16852155
maxDoc : 24512617
reader :
SolrIndexReader{this=6e3b4ec8,r=ReadOnlyDirectoryReader@6e3b4ec8,refCnt=1,segments=61}
<mergePolicy class="org.apache.lucene.index.TieredMergePolicy">
<int name="maxMergeAtOnce">35</int>
<int name="segmentsPerTier">35</int>
<int name="maxMergeAtOnceExplicit">105</int>
<double name="maxMergedSegmentMB">6144.0</double>
<double name="reclaimDeletesWeight">8.0</double>
</mergePolicy>
<mergeScheduler
class="org.apache.lucene.index.ConcurrentMergeScheduler">
<int name="maxMergeCount">20</int>
<int name="maxThreadCount">3</int>
</mergeScheduler>
Thanks,
Robert (Robi) Petersen
Senior Software Engineer
Search Department
(formerly Buy.com)
85 enterprise, suite 100
aliso viejo, ca 92656
tel 949.389.2000 x5465
fax 949.448.5415