Yes Also, in the 90% scan, what every Java memory parameter I use, htop shows the same memory foot print. Its as if the heap isn't being set as per the env parameters that you area asking me to set.
Wayne On Friday, 3 March 2017 07:51:48 UTC, Mattias Persson wrote: > > Querying Lucene, at the very least the way consistency checker uses it, > has a drawback that all matching documents will be read and kept in heap > before going through them. > > So let me ask you something about your data: are there certain property > values that are very common and also indexed? > > On Thursday, March 2, 2017 at 7:07:31 PM UTC+1, [email protected] > wrote: >> >> It appears not: >> >> $env >> JAVA_MEMORY_OPTS=-Xmx32G -Xms32G >> >> . >> . >> . >> >> >> .................... 90% >> 2017-03-01 23:24:55.705+0000 INFO [o.n.c.ConsistencyCheckService] === >> Stage7_RS_Backward === >> 2017-03-01 23:24:55.706+0000 INFO [o.n.c.ConsistencyCheckService] I/Os >> RelationshipStore >> Reads: 3373036269 >> Random Reads: 2732592348 >> ScatterIndex: 81 >> >> 2017-03-01 23:24:55.707+0000 INFO [o.n.c.ConsistencyCheckService] Counts: >> 10338061780 skipCheck >> 1697668359 missCheck >> 5621138678 checked >> 10338061780 correctSkipCheck >> 1688855306 skipBackup >> 3951022794 overwrite >> 2191262 noCacheSkip >> 239346600 activeCache >> 119509522 clearCache >> 2429587416 relSourcePrevCheck >> 995786837 relSourceNextCheck >> 2058354842 relTargetPrevCheck >> 137409583 relTargetNextCheck >> 6917470274 forwardLinks >> 7991190672 backLinks >> 1052730774 nullLinks >> 2017-03-01 23:24:55.708+0000 INFO [o.n.c.ConsistencyCheckService] >> Memory[used:404.70 MB, free:1.63 GB, total:2.03 GB, max:26.67 GB] >> 2017-03-01 23:24:55.708+0000 INFO [o.n.c.ConsistencyCheckService] Done >> in 1h 37m 39s 828ms >> .........2017-03-01 23:45:36.032+0000 INFO >> [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp === >> 2017-03-01 23:45:36.032+0000 INFO [o.n.c.ConsistencyCheckService] I/Os >> RelationshipGroupStore >> Reads: 410800979 >> Random Reads: 102164662 >> ScatterIndex: 24 >> NodeStore >> Reads: 229862945 >> Random Reads: 226895703 >> ScatterIndex: 98 >> RelationshipStore >> Reads: 423304043 >> Random Reads: 139746630 >> ScatterIndex: 33 >> >> 2017-03-01 23:45:36.032+0000 INFO [o.n.c.ConsistencyCheckService] Counts: >> 2017-03-01 23:45:36.033+0000 INFO [o.n.c.ConsistencyCheckService] >> Memory[used:661.75 MB, free:1.39 GB, total:2.03 GB, max:26.67 GB] >> 2017-03-01 23:45:36.034+0000 INFO [o.n.c.ConsistencyCheckService] Done >> in 20m 40s 326ms >> .Exception in thread "ParallelRecordScanner-Stage8_PS_Props-19" >> java.lang.OutOfMemoryError: GC overhead limit exceeded >> at org.apache.lucene.util.BytesRef.<init>(BytesRef.java:73) >> at >> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(FSTOrdsOutputs.java:181) >> at >> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(FSTOrdsOutputs.java:32) >> at org.apache.lucene.util.fst.Outputs.readFinalOutput(Outputs.java:77) >> at org.apache.lucene.util.fst.FST.readNextRealArc(FST.java:1094) >> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1262) >> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1186) >> at >> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:405) >> at org.apache.lucene.index.TermContext.build(TermContext.java:94) >> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >> at >> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >> at >> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >> at >> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >> >> >> I have also tried larger memory values. >> >> Wayne. >> >> >> On Wednesday, 1 March 2017 01:52:47 UTC, Michael Hunger wrote: >>> >>> Sorry I just learned that neo4j-admin uses a different variable >>> >>> "You can pass memory options to the JVM via the `JAVA_MEMORY_OPTS` >>> variable as a workaround though." >>> >>> >>> >>> Von meinem iPhone gesendet >>> >>> Am 28.02.2017 um 18:50 schrieb unrealadmin23 via Neo4j < >>> [email protected]>: >>> >>> Michael, >>> >>> After running the check_consistency command for 1 day with the above >>> parameters, it failed in exactly the same manner. >>> >>> $env | grep -i java >>> JAVA_OPTS=-Xmx32G -Xms32G >>> >>> Any other ideas ? >>> >>> Wayne >>> >>> >>> On Monday, 27 February 2017 16:57:49 UTC, Michael Hunger wrote: >>>> >>>> Do you have really that much RAM in your machine ? 120G usually doesn't >>>> make sense. Most people run with 32G as large heap. >>>> >>>> That said. I asked and currently the numbers from the config are not >>>> used, you have to do: >>>> >>>> export JAVA_OPTS=-Xmx24G -Xms24G >>>> neo4j-admin ... >>>> >>>> >>>> On Mon, Feb 27, 2017 at 8:32 AM, unrealadmin23 via Neo4j < >>>> [email protected]> wrote: >>>> >>>>> >>>>> I should have said, that the head sizes are the ones that I have set >>>>> in neo4j.conf. >>>>> >>>>> Will these be used by check-consistency or do I need to supply them >>>>> elsewhere ? >>>>> >>>>> Wayne. >>>>> >>>>> >>>>> On Monday, 27 February 2017 07:27:33 UTC, [email protected] >>>>> wrote: >>>>>> >>>>>> Michael, >>>>>> >>>>>> neo4j-admin check-consistency --database=test.db --verbose >>>>>> >>>>>> dbms.memory.heap.initial_size=120000m >>>>>> dbms.memory.heap.max_size=120000m >>>>>> >>>>>> Wayne. >>>>>> >>>>>> >>>>>> >>>>>> On Monday, 27 February 2017 02:47:26 UTC, Michael Hunger wrote: >>>>>>> >>>>>>> How did you call the consistency checker? >>>>>>> >>>>>>> How much heap did you provide for it? >>>>>>> >>>>>>> Cheers, Michael >>>>>>> >>>>>>> >>>>>>> On Sun, Feb 26, 2017 at 8:28 PM, unrealadmin23 via Neo4j < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> The following o/p was obtained: >>>>>>>> >>>>>>>> . >>>>>>>> . >>>>>>>> . >>>>>>>> >>>>>>>> .................... 90% >>>>>>>> 2017-02-26 00:03:16.883+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> === Stage7_RS_Backward === >>>>>>>> 2017-02-26 00:03:16.885+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> I/Os >>>>>>>> RelationshipStore >>>>>>>> Reads: 3374851294 >>>>>>>> Random Reads: 2743390177 >>>>>>>> ScatterIndex: 81 >>>>>>>> >>>>>>>> 2017-02-26 00:03:16.886+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> Counts: >>>>>>>> 10338005177 skipCheck >>>>>>>> 1697668360 missCheck >>>>>>>> 5621138678 checked >>>>>>>> 10338005177 correctSkipCheck >>>>>>>> 1688855306 skipBackup >>>>>>>> 3951022795 overwrite >>>>>>>> 2247865 noCacheSkip >>>>>>>> 239346598 activeCache >>>>>>>> 119509521 clearCache >>>>>>>> 2429587416 relSourcePrevCheck >>>>>>>> 995786837 relSourceNextCheck >>>>>>>> 2058354842 relTargetPrevCheck >>>>>>>> 137409583 relTargetNextCheck >>>>>>>> 6917470274 forwardLinks >>>>>>>> 7991190672 backLinks >>>>>>>> 1052730774 nullLinks >>>>>>>> 2017-02-26 00:03:16.887+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> Memory[used:1.09 GB, free:1.07 GB, total:2.17 GB, max:26.67 GB] >>>>>>>> 2017-02-26 00:03:16.887+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> Done in 1h 36m 37s 219ms >>>>>>>> .........2017-02-26 00:23:26.188+0000 INFO >>>>>>>> [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp === >>>>>>>> 2017-02-26 00:23:26.189+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> I/Os >>>>>>>> NodeStore >>>>>>>> Reads: 231527337 >>>>>>>> Random Reads: 228593774 >>>>>>>> ScatterIndex: 98 >>>>>>>> RelationshipStore >>>>>>>> Reads: 420334193 >>>>>>>> Random Reads: 143404207 >>>>>>>> ScatterIndex: 34 >>>>>>>> RelationshipGroupStore >>>>>>>> Reads: 409845841 >>>>>>>> Random Reads: 105935972 >>>>>>>> ScatterIndex: 25 >>>>>>>> >>>>>>>> 2017-02-26 00:23:26.189+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> Counts: >>>>>>>> 2017-02-26 00:23:26.190+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> Memory[used:751.21 MB, free:1.29 GB, total:2.02 GB, max:26.67 GB] >>>>>>>> 2017-02-26 00:23:26.191+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>>>> Done in 20m 9s 303ms >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-11" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:131) >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158) >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391) >>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>>>>> at >>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>>>> at >>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>>>>> at >>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>>>>> at >>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>>>>> at >>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-21" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnumFrame.<init>(OrdsSegmentTermsEnumFrame.java:52) >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.<init>(OrdsSegmentTermsEnum.java:84) >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsFieldReader.iterator(OrdsFieldReader.java:141) >>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:93) >>>>>>>> at >>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>>>> at >>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>>>>> at >>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>>>>> at >>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>>>>> at >>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-8" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:128) >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158) >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391) >>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>>>>> at >>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>>>> at >>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>>>>> at >>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>>>>> at >>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>>>>> at >>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-46" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.newOutput(FSTOrdsOutputs.java:225) >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.add(FSTOrdsOutputs.java:162) >>>>>>>> at >>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:450) >>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>>>>> at >>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>>>> at >>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>>>>> at >>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>>>>> at >>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>>>>> at >>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>>>>> at >>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>>>>> at >>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-22" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-10" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-40" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-58" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-61" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-18" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-25" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-45" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-28" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-50" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-39" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-51" >>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>> >>>>>>>> -- >>>>>>>> You received this message because you are subscribed to the Google >>>>>>>> Groups "Neo4j" group. >>>>>>>> To unsubscribe from this group and stop receiving emails from it, >>>>>>>> send an email to [email protected]. >>>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>>> >>>>>>> >>>>>>> -- >>>>> You received this message because you are subscribed to the Google >>>>> Groups "Neo4j" group. >>>>> To unsubscribe from this group and stop receiving emails from it, send >>>>> an email to [email protected]. >>>>> For more options, visit https://groups.google.com/d/optout. >>>>> >>>> >>>> -- >>> You received this message because you are subscribed to the Google >>> Groups "Neo4j" group. >>> To unsubscribe from this group and stop receiving emails from it, send >>> an email to [email protected]. >>> For more options, visit https://groups.google.com/d/optout. >>> >>> -- You received this message because you are subscribed to the Google Groups "Neo4j" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/d/optout.
