As an update - after running for 48 hours the GC again runs out of heap. I can increase the heap size again (above 124GB) but getting a bit tedious now. There are no indicators that the DB has a problem - I was just interested to ensure that its integrity was good.
Let me know if you wish me to try anything else.... Wayne On Sunday, 5 March 2017 08:04:36 UTC, [email protected] wrote: > > Looks like this fixed the problem - for the Heap size. > Setting Env variables doesn't appear to work - will you be coding this in > to the script ? > > In my case the 90% phase comprises a busy machine, with little IO > activity. The data appears to have been loaded into the memory and the > consistency check is running. Presumably this will run for many days ? > > .................... 90% > 2017-03-04 15:14:41.886+0000 INFO [o.n.c.ConsistencyCheckService] === > Stage7_RS_Backward === > 2017-03-04 15:14:41.887+0000 INFO [o.n.c.ConsistencyCheckService] I/Os > RelationshipStore > Reads: 3358829271 > Random Reads: 2730096948 > ScatterIndex: 81 > > 2017-03-04 15:14:41.888+0000 INFO [o.n.c.ConsistencyCheckService] Counts: > 10338220915 skipCheck > 1697668358 missCheck > 5621138677 checked > 10338220915 correctSkipCheck > 1688855306 skipBackup > 3951022794 overwrite > 2032128 noCacheSkip > 239346600 activeCache > 119509522 clearCache > 2429587415 relSourcePrevCheck > 995786837 relSourceNextCheck > 2058354842 relTargetPrevCheck > 137409583 relTargetNextCheck > 6917470274 forwardLinks > 7991190672 backLinks > 1052730774 nullLinks > 2017-03-04 15:14:41.888+0000 INFO [o.n.c.ConsistencyCheckService] > Memory[used:33.59 GB, free:90.41 GB, total:124.00 GB, max:124.00 GB] > 2017-03-04 15:14:41.888+0000 INFO [o.n.c.ConsistencyCheckService] Done in > 1h 40m 354ms > .........2017-03-04 15:37:20.050+0000 INFO > [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp === > 2017-03-04 15:37:20.051+0000 INFO [o.n.c.ConsistencyCheckService] I/Os > RelationshipGroupStore > Reads: 411311642 > Random Reads: 71933550 > ScatterIndex: 17 > NodeStore > Reads: 208717760 > Random Reads: 205603260 > ScatterIndex: 98 > RelationshipStore > Reads: 419830207 > Random Reads: 112104577 > ScatterIndex: 26 > > 2017-03-04 15:37:20.051+0000 INFO [o.n.c.ConsistencyCheckService] Counts: > 2017-03-04 15:37:20.052+0000 INFO [o.n.c.ConsistencyCheckService] > Memory[used:6.03 GB, free:117.97 GB, total:124.00 GB, max:124.00 GB] > 2017-03-04 15:37:20.052+0000 INFO [o.n.c.ConsistencyCheckService] Done in > 22m 38s 163ms > . > > > > Wayne > > > > > > On Saturday, 4 March 2017 09:16:43 UTC, [email protected] wrote: >> >> So I have added: >> $ cat neo4j-admin >> #!/usr/bin/env bash >> # Copyright (c) 2016 "Neo Technology," >> # Network Engine for Objects in Lund AB [http://neotechnology.com] >> # >> # This file is part of Neo4j. >> # >> # Neo4j is free software: you can redistribute it and/or modify >> # it under the terms of the GNU Affero General Public License as >> # published by the Free Software Foundation, either version 3 of the >> # License, or (at your option) any later version. >> # >> # This program is distributed in the hope that it will be useful, >> # but WITHOUT ANY WARRANTY; without even the implied warranty of >> # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the >> # GNU Affero General Public License for more details. >> # >> # You should have received a copy of the GNU Affero General Public License >> # along with this program. If not, see <http://www.gnu.org/licenses/>. >> set -o errexit -o nounset -o pipefail >> [[ "${TRACE:-}" ]] && set -o xtrace >> >> : "${NEO4J_BIN:=$(dirname "$0")}" >> readonly NEO4J_BIN >> . "${NEO4J_BIN}/neo4j-shared.sh" >> >> main() { >> setup_environment >> check_java >> build_classpath >> export NEO4J_HOME NEO4J_CONF >> exec "${JAVA_CMD}" -Xmx124G -Xms124G -cp "${CLASSPATH}" >> -Dfile.encoding=UTF-8 "org.neo4j.commandline.admin.AdminTool" "$@" >> } >> >> main "$@" >> >> I'll let you know in 24 hours..... >> >> Wayne >> >> On Saturday, 4 March 2017 00:03:26 UTC, Michael Hunger wrote: >>> >>> Can you try to edit the script directly and add the memory parameters >>> there? >>> >>> On Fri, Mar 3, 2017 at 8:49 PM, unrealadmin23 via Neo4j < >>> [email protected]> wrote: >>> >>>> Yes >>>> >>>> Also, in the 90% scan, what every Java memory parameter I use, htop >>>> shows the same memory foot print. Its as if the heap isn't being set as >>>> per the env parameters that you area asking me to set. >>>> >>>> Wayne >>>> >>>> >>>> On Friday, 3 March 2017 07:51:48 UTC, Mattias Persson wrote: >>>>> >>>>> Querying Lucene, at the very least the way consistency checker uses >>>>> it, has a drawback that all matching documents will be read and kept in >>>>> heap before going through them. >>>>> >>>>> So let me ask you something about your data: are there certain >>>>> property values that are very common and also indexed? >>>>> >>>>> On Thursday, March 2, 2017 at 7:07:31 PM UTC+1, >>>>> [email protected] wrote: >>>>>> >>>>>> It appears not: >>>>>> >>>>>> $env >>>>>> JAVA_MEMORY_OPTS=-Xmx32G -Xms32G >>>>>> >>>>>> . >>>>>> . >>>>>> . >>>>>> >>>>>> >>>>>> .................... 90% >>>>>> 2017-03-01 23:24:55.705+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> === Stage7_RS_Backward === >>>>>> 2017-03-01 23:24:55.706+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> I/Os >>>>>> RelationshipStore >>>>>> Reads: 3373036269 >>>>>> Random Reads: 2732592348 >>>>>> ScatterIndex: 81 >>>>>> >>>>>> 2017-03-01 23:24:55.707+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> Counts: >>>>>> 10338061780 skipCheck >>>>>> 1697668359 missCheck >>>>>> 5621138678 checked >>>>>> 10338061780 correctSkipCheck >>>>>> 1688855306 skipBackup >>>>>> 3951022794 overwrite >>>>>> 2191262 noCacheSkip >>>>>> 239346600 activeCache >>>>>> 119509522 clearCache >>>>>> 2429587416 relSourcePrevCheck >>>>>> 995786837 relSourceNextCheck >>>>>> 2058354842 relTargetPrevCheck >>>>>> 137409583 relTargetNextCheck >>>>>> 6917470274 forwardLinks >>>>>> 7991190672 backLinks >>>>>> 1052730774 nullLinks >>>>>> 2017-03-01 23:24:55.708+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> Memory[used:404.70 MB, free:1.63 GB, total:2.03 GB, max:26.67 GB] >>>>>> 2017-03-01 23:24:55.708+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> Done in 1h 37m 39s 828ms >>>>>> .........2017-03-01 23:45:36.032+0000 INFO >>>>>> [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp === >>>>>> 2017-03-01 23:45:36.032+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> I/Os >>>>>> RelationshipGroupStore >>>>>> Reads: 410800979 >>>>>> Random Reads: 102164662 >>>>>> ScatterIndex: 24 >>>>>> NodeStore >>>>>> Reads: 229862945 >>>>>> Random Reads: 226895703 >>>>>> ScatterIndex: 98 >>>>>> RelationshipStore >>>>>> Reads: 423304043 >>>>>> Random Reads: 139746630 >>>>>> ScatterIndex: 33 >>>>>> >>>>>> 2017-03-01 23:45:36.032+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> Counts: >>>>>> 2017-03-01 23:45:36.033+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> Memory[used:661.75 MB, free:1.39 GB, total:2.03 GB, max:26.67 GB] >>>>>> 2017-03-01 23:45:36.034+0000 INFO [o.n.c.ConsistencyCheckService] >>>>>> Done in 20m 40s 326ms >>>>>> .Exception in thread "ParallelRecordScanner-Stage8_PS_Props-19" >>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>> at org.apache.lucene.util.BytesRef.<init>(BytesRef.java:73) >>>>>> at >>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(FSTOrdsOutputs.java:181) >>>>>> at >>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(FSTOrdsOutputs.java:32) >>>>>> at org.apache.lucene.util.fst.Outputs.readFinalOutput(Outputs.java:77) >>>>>> at org.apache.lucene.util.fst.FST.readNextRealArc(FST.java:1094) >>>>>> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1262) >>>>>> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1186) >>>>>> at >>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:405) >>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>> at >>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>> at >>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>>> at >>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>> at >>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>> >>>>>> >>>>>> I have also tried larger memory values. >>>>>> >>>>>> Wayne. >>>>>> >>>>>> >>>>>> On Wednesday, 1 March 2017 01:52:47 UTC, Michael Hunger wrote: >>>>>>> >>>>>>> Sorry I just learned that neo4j-admin uses a different variable >>>>>>> >>>>>>> "You can pass memory options to the JVM via the `JAVA_MEMORY_OPTS` >>>>>>> variable as a workaround though." >>>>>>> >>>>>>> >>>>>>> >>>>>>> Von meinem iPhone gesendet >>>>>>> >>>>>>> Am 28.02.2017 um 18:50 schrieb unrealadmin23 via Neo4j < >>>>>>> [email protected]>: >>>>>>> >>>>>>> Michael, >>>>>>> >>>>>>> After running the check_consistency command for 1 day with the above >>>>>>> parameters, it failed in exactly the same manner. >>>>>>> >>>>>>> $env | grep -i java >>>>>>> JAVA_OPTS=-Xmx32G -Xms32G >>>>>>> >>>>>>> Any other ideas ? >>>>>>> >>>>>>> Wayne >>>>>>> >>>>>>> >>>>>>> On Monday, 27 February 2017 16:57:49 UTC, Michael Hunger wrote: >>>>>>>> >>>>>>>> Do you have really that much RAM in your machine ? 120G usually >>>>>>>> doesn't make sense. Most people run with 32G as large heap. >>>>>>>> >>>>>>>> That said. I asked and currently the numbers from the config are >>>>>>>> not used, you have to do: >>>>>>>> >>>>>>>> export JAVA_OPTS=-Xmx24G -Xms24G >>>>>>>> neo4j-admin ... >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Feb 27, 2017 at 8:32 AM, unrealadmin23 via Neo4j < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> I should have said, that the head sizes are the ones that I have >>>>>>>>> set in neo4j.conf. >>>>>>>>> >>>>>>>>> Will these be used by check-consistency or do I need to supply >>>>>>>>> them elsewhere ? >>>>>>>>> >>>>>>>>> Wayne. >>>>>>>>> >>>>>>>>> >>>>>>>>> On Monday, 27 February 2017 07:27:33 UTC, [email protected] >>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>> Michael, >>>>>>>>>> >>>>>>>>>> neo4j-admin check-consistency --database=test.db --verbose >>>>>>>>>> >>>>>>>>>> dbms.memory.heap.initial_size=120000m >>>>>>>>>> dbms.memory.heap.max_size=120000m >>>>>>>>>> >>>>>>>>>> Wayne. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Monday, 27 February 2017 02:47:26 UTC, Michael Hunger wrote: >>>>>>>>>>> >>>>>>>>>>> How did you call the consistency checker? >>>>>>>>>>> >>>>>>>>>>> How much heap did you provide for it? >>>>>>>>>>> >>>>>>>>>>> Cheers, Michael >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Sun, Feb 26, 2017 at 8:28 PM, unrealadmin23 via Neo4j < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> The following o/p was obtained: >>>>>>>>>>>> >>>>>>>>>>>> . >>>>>>>>>>>> . >>>>>>>>>>>> . >>>>>>>>>>>> >>>>>>>>>>>> .................... 90% >>>>>>>>>>>> 2017-02-26 00:03:16.883+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] === Stage7_RS_Backward === >>>>>>>>>>>> 2017-02-26 00:03:16.885+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] I/Os >>>>>>>>>>>> RelationshipStore >>>>>>>>>>>> Reads: 3374851294 >>>>>>>>>>>> Random Reads: 2743390177 >>>>>>>>>>>> ScatterIndex: 81 >>>>>>>>>>>> >>>>>>>>>>>> 2017-02-26 00:03:16.886+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] Counts: >>>>>>>>>>>> 10338005177 skipCheck >>>>>>>>>>>> 1697668360 missCheck >>>>>>>>>>>> 5621138678 checked >>>>>>>>>>>> 10338005177 correctSkipCheck >>>>>>>>>>>> 1688855306 skipBackup >>>>>>>>>>>> 3951022795 overwrite >>>>>>>>>>>> 2247865 noCacheSkip >>>>>>>>>>>> 239346598 activeCache >>>>>>>>>>>> 119509521 clearCache >>>>>>>>>>>> 2429587416 relSourcePrevCheck >>>>>>>>>>>> 995786837 relSourceNextCheck >>>>>>>>>>>> 2058354842 relTargetPrevCheck >>>>>>>>>>>> 137409583 relTargetNextCheck >>>>>>>>>>>> 6917470274 forwardLinks >>>>>>>>>>>> 7991190672 backLinks >>>>>>>>>>>> 1052730774 nullLinks >>>>>>>>>>>> 2017-02-26 00:03:16.887+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] Memory[used:1.09 GB, free:1.07 >>>>>>>>>>>> GB, >>>>>>>>>>>> total:2.17 GB, max:26.67 GB] >>>>>>>>>>>> 2017-02-26 00:03:16.887+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] Done in 1h 36m 37s 219ms >>>>>>>>>>>> .........2017-02-26 00:23:26.188+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp >>>>>>>>>>>> === >>>>>>>>>>>> 2017-02-26 00:23:26.189+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] I/Os >>>>>>>>>>>> NodeStore >>>>>>>>>>>> Reads: 231527337 >>>>>>>>>>>> Random Reads: 228593774 >>>>>>>>>>>> ScatterIndex: 98 >>>>>>>>>>>> RelationshipStore >>>>>>>>>>>> Reads: 420334193 >>>>>>>>>>>> Random Reads: 143404207 >>>>>>>>>>>> ScatterIndex: 34 >>>>>>>>>>>> RelationshipGroupStore >>>>>>>>>>>> Reads: 409845841 >>>>>>>>>>>> Random Reads: 105935972 >>>>>>>>>>>> ScatterIndex: 25 >>>>>>>>>>>> >>>>>>>>>>>> 2017-02-26 00:23:26.189+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] Counts: >>>>>>>>>>>> 2017-02-26 00:23:26.190+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] Memory[used:751.21 MB, free:1.29 >>>>>>>>>>>> GB, >>>>>>>>>>>> total:2.02 GB, max:26.67 GB] >>>>>>>>>>>> 2017-02-26 00:23:26.191+0000 INFO >>>>>>>>>>>> [o.n.c.ConsistencyCheckService] Done in 20m 9s 303ms >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-11" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:131) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-21" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnumFrame.<init>(OrdsSegmentTermsEnumFrame.java:52) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.<init>(OrdsSegmentTermsEnum.java:84) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsFieldReader.iterator(OrdsFieldReader.java:141) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.index.TermContext.build(TermContext.java:93) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-8" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:128) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-46" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.newOutput(FSTOrdsOutputs.java:225) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.add(FSTOrdsOutputs.java:162) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:450) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.index.TermContext.build(TermContext.java:94) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887) >>>>>>>>>>>> at >>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535) >>>>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co >>>>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77) >>>>>>>>>>>> at >>>>>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137) >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-22" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-10" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-40" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-58" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-61" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-18" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-25" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-45" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-28" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-50" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-39" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-51" >>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> You received this message because you are subscribed to the >>>>>>>>>>>> Google Groups "Neo4j" group. >>>>>>>>>>>> To unsubscribe from this group and stop receiving emails from >>>>>>>>>>>> it, send an email to [email protected]. >>>>>>>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>> You received this message because you are subscribed to the Google >>>>>>>>> Groups "Neo4j" group. >>>>>>>>> To unsubscribe from this group and stop receiving emails from it, >>>>>>>>> send an email to [email protected]. >>>>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>> You received this message because you are subscribed to the Google >>>>>>> Groups "Neo4j" group. >>>>>>> To unsubscribe from this group and stop receiving emails from it, >>>>>>> send an email to [email protected]. >>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>> >>>>>>> -- >>>> You received this message because you are subscribed to the Google >>>> Groups "Neo4j" group. >>>> To unsubscribe from this group and stop receiving emails from it, send >>>> an email to [email protected]. >>>> For more options, visit https://groups.google.com/d/optout. >>>> >>> >>> -- You received this message because you are subscribed to the Google Groups "Neo4j" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/d/optout.
