I have 1TB so usual consider 120GB to be small; but then I'm not a Java person. I will set the java_opts as per below and see what happens....
Wayne. On Monday, 27 February 2017 16:57:49 UTC, Michael Hunger wrote: > > Do you have really that much RAM in your machine ? 120G usually doesn't > make sense. Most people run with 32G as large heap. > > That said. I asked and currently the numbers from the config are not used, > you have to do: > > export JAVA_OPTS=-Xmx24G -Xms24G > neo4j-admin ... > > > On Mon, Feb 27, 2017 at 8:32 AM, unrealadmin23 via Neo4j < > [email protected] <javascript:>> wrote: > >> >> I should have said, that the head sizes are the ones that I have set in >> neo4j.conf. >> >> Will these be used by check-consistency or do I need to supply them >> elsewhere ? >> >> Wayne. >> >> >> On Monday, 27 February 2017 07:27:33 UTC, [email protected] wrote: >>> >>> Michael, >>> >>> neo4j-admin check-consistency --database=test.db --verbose >>> >>> dbms.memory.heap.initial_size=120000m >>> dbms.memory.heap.max_size=120000m >>> >>> Wayne. >>> >>> >>> >>> On Monday, 27 February 2017 02:47:26 UTC, Michael Hunger wrote: >>>> >>>> How did you call the consistency checker? >>>> >>>> How much heap did you provide for it? >>>> >>>> Cheers, Michael >>>> >>>> >>>> On Sun, Feb 26, 2017 at 8:28 PM, unrealadmin23 via Neo4j < >>>> [email protected]> wrote: >>>> >>>>> The following o/p was obtained: >>>>> >>>>> . >>>>> . >>>>> . >>>>> >>>>> .................... 90% >>>>> 2017-02-26 00:03:16.883+0000 INFO [o.n.c.ConsistencyCheckService] === >>>>> Stage7_RS_Backward === >>>>> 2017-02-26 00:03:16.885+0000 INFO [o.n.c.ConsistencyCheckService] I/Os >>>>> RelationshipStore >>>>> Reads: 3374851294 >>>>> Random Reads: 2743390177 >>>>> ScatterIndex: 81 >>>>> >>>>> 2017-02-26 00:03:16.886+0000 INFO [o.n.c.ConsistencyCheckService] >>>>> Counts: >>>>> 10338005177 skipCheck >>>>> 1697668360 missCheck >>>>> 5621138678 checked >>>>> 10338005177 correctSkipCheck >>>>> 1688855306 skipBackup >>>>> 3951022795 overwrite >>>>> 2247865 noCacheSkip >>>>> 239346598 activeCache >>>>> 119509521 clearCache >>>>> 2429587416 relSourcePrevCheck >>>>> 995786837 relSourceNextCheck >>>>> 2058354842 relTargetPrevCheck >>>>> 137409583 relTargetNextCheck >>>>> 6917470274 forwardLinks >>>>> 7991190672 backLinks >>>>> 1052730774 nullLinks >>>>> 2017-02-26 00:03:16.887+0000 INFO [o.n.c.ConsistencyCheckService] >>>>> Memory[used:1.09 GB, free:1.07 GB, total:2.17 GB, max:26.67 GB] >>>>> 2017-02-26 00:03:16.887+0000 INFO [o.n.c.ConsistencyCheckService] >>>>> Done in 1h 36m 37s 219ms >>>>> .........2017-02-26 00:23:26.188+0000 INFO >>>>> [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp === >>>>> 2017-02-26 00:23:26.189+0000 INFO [o.n.c.ConsistencyCheckService] I/Os >>>>> NodeStore >>>>> Reads: 231527337 >>>>> Random Reads: 228593774 >>>>> ScatterIndex: 98 >>>>> RelationshipStore >>>>> Reads: 420334193 >>>>> Random Reads: 143404207 >>>>> ScatterIndex: 34 >>>>> RelationshipGroupStore >>>>> Reads: 409845841 >>>>> Random Reads: 105935972 >>>>> ScatterIndex: 25 >>>>> >>>>> 2017-02-26 00:23:26.189+0000 INFO [o.n.c.ConsistencyCheckService] >>>>> Counts: >>>>> 2017-02-26 00:23:26.190+0000 INFO [o.n.c.ConsistencyCheckService] >>>>> Memory[used:751.21 MB, free:1.29 GB, total:2.02 GB, max:26.67 GB] >>>>> 2017-02-26 00:23:26.191+0000 INFO [o.n.c.ConsistencyCheckService] >>>>> Done in 20m 9s 303ms >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-11" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:131) >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158) >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391) >>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at >>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>> at >>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>> at >>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>> at >>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>> at >>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>> at >>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-21" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnumFrame.<init>(OrdsSegmentTermsEnumFrame.java:52) >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.<init>(OrdsSegmentTermsEnum.java:84) >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsFieldReader.iterator(OrdsFieldReader.java:141) >>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:93) >>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>> at >>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>> at >>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>> at >>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>> at >>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>> at >>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-8" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:128) >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158) >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391) >>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at >>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>> at >>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>> at >>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>> at >>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>> at >>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>> at >>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-46" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.newOutput(FSTOrdsOutputs.java:225) >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.add(FSTOrdsOutputs.java:162) >>>>> at >>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:450) >>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at >>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>> at >>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>> at >>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>> at >>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>> at >>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>> at >>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>> at >>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>> at >>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-22" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-10" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-40" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-58" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-61" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> >>>>> >>>>> >>>>> >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-18" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-25" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-45" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-28" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-50" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-39" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-51" >>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>> >>>>> -- >>>>> You received this message because you are subscribed to the Google >>>>> Groups "Neo4j" group. >>>>> To unsubscribe from this group and stop receiving emails from it, send >>>>> an email to [email protected]. >>>>> For more options, visit https://groups.google.com/d/optout. >>>>> >>>> >>>> -- >> You received this message because you are subscribed to the Google Groups >> "Neo4j" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to [email protected] <javascript:>. >> For more options, visit https://groups.google.com/d/optout. >> > > -- You received this message because you are subscribed to the Google Groups "Neo4j" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/d/optout.
