Hi,
1.We Upgraded Lucene 4.6 to 8+, After upgrading we are facing issue with Lucene Index Creation. We are indexing in Multi-threading environment. When we create bulk indexes , Lucene Document is getting corrupted. (data is not getting updated correctly. Merging of different row data). 2. when we are trying to updateDocument method for single record. It is not reflecting in IndexReader until the count is 8. Once the count exceeds, than records are visible for IndexReader. (creating 8 segment files.) is there any alternative for reducing these segment file creation. 3. above two issues are resolved by forcemerge(1). But it is not feasible for our use case , because it takes 3X memory. We are creating indexes for huge data. 4. IndexWriter Config: analyzer=com.datanomic.director.casemanagement.indexing.AnalyzerFactory$MA ramBufferSizeMB=64.0 maxBufferedDocs=-1 mergedSegmentWarmer=null delPolicy=com.datanomic.director.casemanagement.indexing.engines.TimedDeletionPolicy commit=null openMode=CREATE_OR_APPEND similarity=org.apache.lucene.search.similarities.BM25Similarity mergeScheduler=ConcurrentMergeScheduler: maxThreadCount=-1, maxMergeCount=-1, ioThrottle=true codec=Lucene80 infoStream=org.apache.lucene.util.InfoStream$NoOutput mergePolicy=[TieredMergePolicy: maxMergeAtOnce=10, maxMergeAtOnceExplicit=30, maxMergedSegmentMB=5120.0, floorSegmentMB=2.0, forceMergeDeletesPctAllowed=10.0, segmentsPerTier=10.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.1, deletesPctAllowed=33.0 indexerThreadPool=org.apache.lucene.index.DocumentsWriterPerThreadPool@24348e05 readerPooling=true perThreadHardLimitMB=1945 useCompoundFile=false commitOnClose=true indexSort=null checkPendingFlushOnUpdate=true softDeletesField=null readerAttributes={} writer=org.apache.lucene.index.IndexWriter@23a84a99 Please suggest some ideas alternate of forceMerge, dealing with indexwriter.commit for multithreading, committing data while updating single record. Thanks, Jyothsna