Looks like this fixed the problem - for the Heap size.
Setting Env variables doesn't appear to work - will you be coding this in 
to the script ?

In my case the 90% phase comprises a busy machine,  with little IO 
activity.   The data appears to have been loaded into the memory and the 
consistency check is running.  Presumably this will run for many days ?

....................  90%
2017-03-04 15:14:41.886+0000 INFO  [o.n.c.ConsistencyCheckService] === 
Stage7_RS_Backward ===
2017-03-04 15:14:41.887+0000 INFO  [o.n.c.ConsistencyCheckService] I/Os
RelationshipStore
  Reads: 3358829271
  Random Reads: 2730096948
  ScatterIndex: 81

2017-03-04 15:14:41.888+0000 INFO  [o.n.c.ConsistencyCheckService] Counts:
  10338220915 skipCheck
  1697668358 missCheck
  5621138677 checked
  10338220915 correctSkipCheck
  1688855306 skipBackup
  3951022794 overwrite
  2032128 noCacheSkip
  239346600 activeCache
  119509522 clearCache
  2429587415 relSourcePrevCheck
  995786837 relSourceNextCheck
  2058354842 relTargetPrevCheck
  137409583 relTargetNextCheck
  6917470274 forwardLinks
  7991190672 backLinks
  1052730774 nullLinks
2017-03-04 15:14:41.888+0000 INFO  [o.n.c.ConsistencyCheckService] 
Memory[used:33.59 GB, free:90.41 GB, total:124.00 GB, max:124.00 GB]
2017-03-04 15:14:41.888+0000 INFO  [o.n.c.ConsistencyCheckService] Done in 
 1h 40m 354ms
.........2017-03-04 15:37:20.050+0000 INFO  [o.n.c.ConsistencyCheckService] 
=== RelationshipGroupStore-RelGrp ===
2017-03-04 15:37:20.051+0000 INFO  [o.n.c.ConsistencyCheckService] I/Os
RelationshipGroupStore
  Reads: 411311642
  Random Reads: 71933550
  ScatterIndex: 17
NodeStore
  Reads: 208717760
  Random Reads: 205603260
  ScatterIndex: 98
RelationshipStore
  Reads: 419830207
  Random Reads: 112104577
  ScatterIndex: 26

2017-03-04 15:37:20.051+0000 INFO  [o.n.c.ConsistencyCheckService] Counts:
2017-03-04 15:37:20.052+0000 INFO  [o.n.c.ConsistencyCheckService] 
Memory[used:6.03 GB, free:117.97 GB, total:124.00 GB, max:124.00 GB]
2017-03-04 15:37:20.052+0000 INFO  [o.n.c.ConsistencyCheckService] Done in 
 22m 38s 163ms
.



Wayne





On Saturday, 4 March 2017 09:16:43 UTC, [email protected] wrote:
>
> So I have added:
> $ cat neo4j-admin
> #!/usr/bin/env bash
> # Copyright (c) 2016 "Neo Technology,"
> # Network Engine for Objects in Lund AB [http://neotechnology.com]
> #
> # This file is part of Neo4j.
> #
> # Neo4j is free software: you can redistribute it and/or modify
> # it under the terms of the GNU Affero General Public License as
> # published by the Free Software Foundation, either version 3 of the
> # License, or (at your option) any later version.
> #
> # This program is distributed in the hope that it will be useful,
> # but WITHOUT ANY WARRANTY; without even the implied warranty of
> # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
> # GNU Affero General Public License for more details.
> #
> # You should have received a copy of the GNU Affero General Public License
> # along with this program. If not, see <http://www.gnu.org/licenses/>.
> set -o errexit -o nounset -o pipefail
> [[ "${TRACE:-}" ]] && set -o xtrace
>
> : "${NEO4J_BIN:=$(dirname "$0")}"
> readonly NEO4J_BIN
> . "${NEO4J_BIN}/neo4j-shared.sh"
>
> main() {
>   setup_environment
>   check_java
>   build_classpath
>   export NEO4J_HOME NEO4J_CONF
>   exec "${JAVA_CMD}" -Xmx124G -Xms124G -cp "${CLASSPATH}" 
> -Dfile.encoding=UTF-8 "org.neo4j.commandline.admin.AdminTool" "$@"
> }
>
> main "$@"
>
> I'll let you know in 24 hours.....
>
> Wayne
>
> On Saturday, 4 March 2017 00:03:26 UTC, Michael Hunger wrote:
>>
>> Can you try to edit the script directly and add the memory parameters 
>> there?
>>
>> On Fri, Mar 3, 2017 at 8:49 PM, unrealadmin23 via Neo4j <
>> [email protected]> wrote:
>>
>>> Yes 
>>>
>>> Also, in the 90% scan, what every Java memory parameter I use,   htop 
>>> shows the same memory foot print.   Its as if the heap isn't being set as 
>>> per the env parameters that you area asking me to set.
>>>
>>> Wayne
>>>
>>>
>>> On Friday, 3 March 2017 07:51:48 UTC, Mattias Persson wrote:
>>>>
>>>> Querying Lucene, at the very least the way consistency checker uses it, 
>>>> has a drawback that all matching documents will be read and kept in heap 
>>>> before going through them.
>>>>
>>>> So let me ask you something about your data: are there certain property 
>>>> values that are very common and also indexed?
>>>>
>>>> On Thursday, March 2, 2017 at 7:07:31 PM UTC+1, 
>>>> [email protected] wrote:
>>>>>
>>>>> It appears not:
>>>>>
>>>>> $env
>>>>> JAVA_MEMORY_OPTS=-Xmx32G -Xms32G
>>>>>
>>>>> .
>>>>> .
>>>>> .
>>>>>
>>>>>
>>>>> ....................  90%
>>>>> 2017-03-01 23:24:55.705+0000 INFO  [o.n.c.ConsistencyCheckService] === 
>>>>> Stage7_RS_Backward ===
>>>>> 2017-03-01 23:24:55.706+0000 INFO  [o.n.c.ConsistencyCheckService] I/Os
>>>>> RelationshipStore
>>>>>   Reads: 3373036269
>>>>>   Random Reads: 2732592348
>>>>>   ScatterIndex: 81
>>>>>
>>>>> 2017-03-01 23:24:55.707+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>> Counts:
>>>>>   10338061780 skipCheck
>>>>>   1697668359 missCheck
>>>>>   5621138678 checked
>>>>>   10338061780 correctSkipCheck
>>>>>   1688855306 skipBackup
>>>>>   3951022794 overwrite
>>>>>   2191262 noCacheSkip
>>>>>   239346600 activeCache
>>>>>   119509522 clearCache
>>>>>   2429587416 relSourcePrevCheck
>>>>>   995786837 relSourceNextCheck
>>>>>   2058354842 relTargetPrevCheck
>>>>>   137409583 relTargetNextCheck
>>>>>   6917470274 forwardLinks
>>>>>   7991190672 backLinks
>>>>>   1052730774 nullLinks
>>>>> 2017-03-01 23:24:55.708+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>> Memory[used:404.70 MB, free:1.63 GB, total:2.03 GB, max:26.67 GB]
>>>>> 2017-03-01 23:24:55.708+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>> Done in  1h 37m 39s 828ms
>>>>> .........2017-03-01 23:45:36.032+0000 INFO 
>>>>>  [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp ===
>>>>> 2017-03-01 23:45:36.032+0000 INFO  [o.n.c.ConsistencyCheckService] I/Os
>>>>> RelationshipGroupStore
>>>>>   Reads: 410800979
>>>>>   Random Reads: 102164662
>>>>>   ScatterIndex: 24
>>>>> NodeStore
>>>>>   Reads: 229862945
>>>>>   Random Reads: 226895703
>>>>>   ScatterIndex: 98
>>>>> RelationshipStore
>>>>>   Reads: 423304043
>>>>>   Random Reads: 139746630
>>>>>   ScatterIndex: 33
>>>>>
>>>>> 2017-03-01 23:45:36.032+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>> Counts:
>>>>> 2017-03-01 23:45:36.033+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>> Memory[used:661.75 MB, free:1.39 GB, total:2.03 GB, max:26.67 GB]
>>>>> 2017-03-01 23:45:36.034+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>> Done in  20m 40s 326ms
>>>>> .Exception in thread "ParallelRecordScanner-Stage8_PS_Props-19" 
>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>> at org.apache.lucene.util.BytesRef.<init>(BytesRef.java:73)
>>>>> at 
>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(FSTOrdsOutputs.java:181)
>>>>> at 
>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(FSTOrdsOutputs.java:32)
>>>>> at org.apache.lucene.util.fst.Outputs.readFinalOutput(Outputs.java:77)
>>>>> at org.apache.lucene.util.fst.FST.readNextRealArc(FST.java:1094)
>>>>> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1262)
>>>>> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1186)
>>>>> at 
>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:405)
>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>> at 
>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>> at 
>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119)
>>>>> at 
>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>
>>>>>
>>>>> I have also tried larger memory values.
>>>>>
>>>>> Wayne.
>>>>>
>>>>>
>>>>> On Wednesday, 1 March 2017 01:52:47 UTC, Michael Hunger wrote:
>>>>>>
>>>>>> Sorry I just learned that neo4j-admin uses a different variable 
>>>>>>
>>>>>> "You can pass memory options to the JVM via the `JAVA_MEMORY_OPTS` 
>>>>>> variable as a workaround though."
>>>>>>
>>>>>>
>>>>>>
>>>>>> Von meinem iPhone gesendet
>>>>>>
>>>>>> Am 28.02.2017 um 18:50 schrieb unrealadmin23 via Neo4j <
>>>>>> [email protected]>:
>>>>>>
>>>>>> Michael,
>>>>>>
>>>>>> After running the check_consistency command for 1 day with the above 
>>>>>> parameters, it failed in exactly the same manner.
>>>>>>
>>>>>> $env | grep -i java
>>>>>> JAVA_OPTS=-Xmx32G -Xms32G
>>>>>>
>>>>>> Any other ideas ?
>>>>>>
>>>>>> Wayne
>>>>>>
>>>>>>
>>>>>> On Monday, 27 February 2017 16:57:49 UTC, Michael Hunger wrote:
>>>>>>>
>>>>>>> Do you have really that much RAM in your machine ? 120G usually 
>>>>>>> doesn't make sense. Most people run with 32G as large heap.
>>>>>>>
>>>>>>> That said. I asked and currently the numbers from the config are not 
>>>>>>> used, you have to do:
>>>>>>>
>>>>>>> export JAVA_OPTS=-Xmx24G -Xms24G
>>>>>>> neo4j-admin ...
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Feb 27, 2017 at 8:32 AM, unrealadmin23 via Neo4j <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>>
>>>>>>>> I should have said, that the head sizes are the ones that I have 
>>>>>>>> set in neo4j.conf.
>>>>>>>>
>>>>>>>> Will these be used by check-consistency or do I need to supply them 
>>>>>>>> elsewhere ?
>>>>>>>>
>>>>>>>> Wayne.
>>>>>>>>
>>>>>>>>
>>>>>>>> On Monday, 27 February 2017 07:27:33 UTC, [email protected] 
>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>> Michael,
>>>>>>>>>
>>>>>>>>> neo4j-admin check-consistency --database=test.db --verbose
>>>>>>>>>
>>>>>>>>> dbms.memory.heap.initial_size=120000m
>>>>>>>>> dbms.memory.heap.max_size=120000m
>>>>>>>>>
>>>>>>>>> Wayne.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Monday, 27 February 2017 02:47:26 UTC, Michael Hunger wrote:
>>>>>>>>>>
>>>>>>>>>> How did you call the consistency checker?
>>>>>>>>>>
>>>>>>>>>> How much heap did you provide for it?
>>>>>>>>>>
>>>>>>>>>> Cheers, Michael
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Sun, Feb 26, 2017 at 8:28 PM, unrealadmin23 via Neo4j <
>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>
>>>>>>>>>>> The following o/p was obtained:
>>>>>>>>>>>
>>>>>>>>>>> .
>>>>>>>>>>> .
>>>>>>>>>>> .
>>>>>>>>>>>
>>>>>>>>>>> ....................  90%
>>>>>>>>>>> 2017-02-26 00:03:16.883+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] === Stage7_RS_Backward ===
>>>>>>>>>>> 2017-02-26 00:03:16.885+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] I/Os
>>>>>>>>>>> RelationshipStore
>>>>>>>>>>>   Reads: 3374851294
>>>>>>>>>>>   Random Reads: 2743390177
>>>>>>>>>>>   ScatterIndex: 81
>>>>>>>>>>>
>>>>>>>>>>> 2017-02-26 00:03:16.886+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] Counts:
>>>>>>>>>>>   10338005177 skipCheck
>>>>>>>>>>>   1697668360 missCheck
>>>>>>>>>>>   5621138678 checked
>>>>>>>>>>>   10338005177 correctSkipCheck
>>>>>>>>>>>   1688855306 skipBackup
>>>>>>>>>>>   3951022795 overwrite
>>>>>>>>>>>   2247865 noCacheSkip
>>>>>>>>>>>   239346598 activeCache
>>>>>>>>>>>   119509521 clearCache
>>>>>>>>>>>   2429587416 relSourcePrevCheck
>>>>>>>>>>>   995786837 relSourceNextCheck
>>>>>>>>>>>   2058354842 relTargetPrevCheck
>>>>>>>>>>>   137409583 relTargetNextCheck
>>>>>>>>>>>   6917470274 forwardLinks
>>>>>>>>>>>   7991190672 backLinks
>>>>>>>>>>>   1052730774 nullLinks
>>>>>>>>>>> 2017-02-26 00:03:16.887+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] Memory[used:1.09 GB, free:1.07 GB, 
>>>>>>>>>>> total:2.17 GB, max:26.67 GB]
>>>>>>>>>>> 2017-02-26 00:03:16.887+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] Done in  1h 36m 37s 219ms
>>>>>>>>>>> .........2017-02-26 00:23:26.188+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp 
>>>>>>>>>>> ===
>>>>>>>>>>> 2017-02-26 00:23:26.189+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] I/Os
>>>>>>>>>>> NodeStore
>>>>>>>>>>>   Reads: 231527337
>>>>>>>>>>>   Random Reads: 228593774
>>>>>>>>>>>   ScatterIndex: 98
>>>>>>>>>>> RelationshipStore
>>>>>>>>>>>   Reads: 420334193
>>>>>>>>>>>   Random Reads: 143404207
>>>>>>>>>>>   ScatterIndex: 34
>>>>>>>>>>> RelationshipGroupStore
>>>>>>>>>>>   Reads: 409845841
>>>>>>>>>>>   Random Reads: 105935972
>>>>>>>>>>>   ScatterIndex: 25
>>>>>>>>>>>
>>>>>>>>>>> 2017-02-26 00:23:26.189+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] Counts:
>>>>>>>>>>> 2017-02-26 00:23:26.190+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] Memory[used:751.21 MB, free:1.29 
>>>>>>>>>>> GB, 
>>>>>>>>>>> total:2.02 GB, max:26.67 GB]
>>>>>>>>>>> 2017-02-26 00:23:26.191+0000 INFO 
>>>>>>>>>>>  [o.n.c.ConsistencyCheckService] Done in  20m 9s 303ms
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-11" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:131)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391)
>>>>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535)
>>>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137)
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-21" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnumFrame.<init>(OrdsSegmentTermsEnumFrame.java:52)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.<init>(OrdsSegmentTermsEnum.java:84)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsFieldReader.iterator(OrdsFieldReader.java:141)
>>>>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:93)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535)
>>>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137)
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-8" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:128)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391)
>>>>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535)
>>>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137)
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-46" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.newOutput(FSTOrdsOutputs.java:225)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.add(FSTOrdsOutputs.java:162)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:450)
>>>>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535)
>>>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77)
>>>>>>>>>>> at 
>>>>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137)
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-22" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-10" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-40" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-58" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-61" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-18" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-25" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-45" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-28" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-50" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-39" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-51" 
>>>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>>>
>>>>>>>>>>> -- 
>>>>>>>>>>> You received this message because you are subscribed to the 
>>>>>>>>>>> Google Groups "Neo4j" group.
>>>>>>>>>>> To unsubscribe from this group and stop receiving emails from 
>>>>>>>>>>> it, send an email to [email protected].
>>>>>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> -- 
>>>>>>>> You received this message because you are subscribed to the Google 
>>>>>>>> Groups "Neo4j" group.
>>>>>>>> To unsubscribe from this group and stop receiving emails from it, 
>>>>>>>> send an email to [email protected].
>>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>>
>>>>>>>
>>>>>>> -- 
>>>>>> You received this message because you are subscribed to the Google 
>>>>>> Groups "Neo4j" group.
>>>>>> To unsubscribe from this group and stop receiving emails from it, 
>>>>>> send an email to [email protected].
>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>
>>>>>> -- 
>>> You received this message because you are subscribed to the Google 
>>> Groups "Neo4j" group.
>>> To unsubscribe from this group and stop receiving emails from it, send 
>>> an email to [email protected].
>>> For more options, visit https://groups.google.com/d/optout.
>>>
>>
>>

-- 
You received this message because you are subscribed to the Google Groups 
"Neo4j" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to