Yes 

Also, in the 90% scan, what every Java memory parameter I use,   htop shows 
the same memory foot print.   Its as if the heap isn't being set as per the 
env parameters that you area asking me to set.

Wayne

On Friday, 3 March 2017 07:51:48 UTC, Mattias Persson wrote:
>
> Querying Lucene, at the very least the way consistency checker uses it, 
> has a drawback that all matching documents will be read and kept in heap 
> before going through them.
>
> So let me ask you something about your data: are there certain property 
> values that are very common and also indexed?
>
> On Thursday, March 2, 2017 at 7:07:31 PM UTC+1, [email protected] 
> wrote:
>>
>> It appears not:
>>
>> $env
>> JAVA_MEMORY_OPTS=-Xmx32G -Xms32G
>>
>> .
>> .
>> .
>>
>>
>> ....................  90%
>> 2017-03-01 23:24:55.705+0000 INFO  [o.n.c.ConsistencyCheckService] === 
>> Stage7_RS_Backward ===
>> 2017-03-01 23:24:55.706+0000 INFO  [o.n.c.ConsistencyCheckService] I/Os
>> RelationshipStore
>>   Reads: 3373036269
>>   Random Reads: 2732592348
>>   ScatterIndex: 81
>>
>> 2017-03-01 23:24:55.707+0000 INFO  [o.n.c.ConsistencyCheckService] Counts:
>>   10338061780 skipCheck
>>   1697668359 missCheck
>>   5621138678 checked
>>   10338061780 correctSkipCheck
>>   1688855306 skipBackup
>>   3951022794 overwrite
>>   2191262 noCacheSkip
>>   239346600 activeCache
>>   119509522 clearCache
>>   2429587416 relSourcePrevCheck
>>   995786837 relSourceNextCheck
>>   2058354842 relTargetPrevCheck
>>   137409583 relTargetNextCheck
>>   6917470274 forwardLinks
>>   7991190672 backLinks
>>   1052730774 nullLinks
>> 2017-03-01 23:24:55.708+0000 INFO  [o.n.c.ConsistencyCheckService] 
>> Memory[used:404.70 MB, free:1.63 GB, total:2.03 GB, max:26.67 GB]
>> 2017-03-01 23:24:55.708+0000 INFO  [o.n.c.ConsistencyCheckService] Done 
>> in  1h 37m 39s 828ms
>> .........2017-03-01 23:45:36.032+0000 INFO 
>>  [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp ===
>> 2017-03-01 23:45:36.032+0000 INFO  [o.n.c.ConsistencyCheckService] I/Os
>> RelationshipGroupStore
>>   Reads: 410800979
>>   Random Reads: 102164662
>>   ScatterIndex: 24
>> NodeStore
>>   Reads: 229862945
>>   Random Reads: 226895703
>>   ScatterIndex: 98
>> RelationshipStore
>>   Reads: 423304043
>>   Random Reads: 139746630
>>   ScatterIndex: 33
>>
>> 2017-03-01 23:45:36.032+0000 INFO  [o.n.c.ConsistencyCheckService] Counts:
>> 2017-03-01 23:45:36.033+0000 INFO  [o.n.c.ConsistencyCheckService] 
>> Memory[used:661.75 MB, free:1.39 GB, total:2.03 GB, max:26.67 GB]
>> 2017-03-01 23:45:36.034+0000 INFO  [o.n.c.ConsistencyCheckService] Done 
>> in  20m 40s 326ms
>> .Exception in thread "ParallelRecordScanner-Stage8_PS_Props-19" 
>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>> at org.apache.lucene.util.BytesRef.<init>(BytesRef.java:73)
>> at 
>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(FSTOrdsOutputs.java:181)
>> at 
>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(FSTOrdsOutputs.java:32)
>> at org.apache.lucene.util.fst.Outputs.readFinalOutput(Outputs.java:77)
>> at org.apache.lucene.util.fst.FST.readNextRealArc(FST.java:1094)
>> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1262)
>> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1186)
>> at 
>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:405)
>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>> at 
>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>> at 
>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119)
>> at 
>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>
>>
>> I have also tried larger memory values.
>>
>> Wayne.
>>
>>
>> On Wednesday, 1 March 2017 01:52:47 UTC, Michael Hunger wrote:
>>>
>>> Sorry I just learned that neo4j-admin uses a different variable 
>>>
>>> "You can pass memory options to the JVM via the `JAVA_MEMORY_OPTS` 
>>> variable as a workaround though."
>>>
>>>
>>>
>>> Von meinem iPhone gesendet
>>>
>>> Am 28.02.2017 um 18:50 schrieb unrealadmin23 via Neo4j <
>>> [email protected]>:
>>>
>>> Michael,
>>>
>>> After running the check_consistency command for 1 day with the above 
>>> parameters, it failed in exactly the same manner.
>>>
>>> $env | grep -i java
>>> JAVA_OPTS=-Xmx32G -Xms32G
>>>
>>> Any other ideas ?
>>>
>>> Wayne
>>>
>>>
>>> On Monday, 27 February 2017 16:57:49 UTC, Michael Hunger wrote:
>>>>
>>>> Do you have really that much RAM in your machine ? 120G usually doesn't 
>>>> make sense. Most people run with 32G as large heap.
>>>>
>>>> That said. I asked and currently the numbers from the config are not 
>>>> used, you have to do:
>>>>
>>>> export JAVA_OPTS=-Xmx24G -Xms24G
>>>> neo4j-admin ...
>>>>
>>>>
>>>> On Mon, Feb 27, 2017 at 8:32 AM, unrealadmin23 via Neo4j <
>>>> [email protected]> wrote:
>>>>
>>>>>
>>>>> I should have said, that the head sizes are the ones that I have set 
>>>>> in neo4j.conf.
>>>>>
>>>>> Will these be used by check-consistency or do I need to supply them 
>>>>> elsewhere ?
>>>>>
>>>>> Wayne.
>>>>>
>>>>>
>>>>> On Monday, 27 February 2017 07:27:33 UTC, [email protected] 
>>>>> wrote:
>>>>>>
>>>>>> Michael,
>>>>>>
>>>>>> neo4j-admin check-consistency --database=test.db --verbose
>>>>>>
>>>>>> dbms.memory.heap.initial_size=120000m
>>>>>> dbms.memory.heap.max_size=120000m
>>>>>>
>>>>>> Wayne.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Monday, 27 February 2017 02:47:26 UTC, Michael Hunger wrote:
>>>>>>>
>>>>>>> How did you call the consistency checker?
>>>>>>>
>>>>>>> How much heap did you provide for it?
>>>>>>>
>>>>>>> Cheers, Michael
>>>>>>>
>>>>>>>
>>>>>>> On Sun, Feb 26, 2017 at 8:28 PM, unrealadmin23 via Neo4j <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> The following o/p was obtained:
>>>>>>>>
>>>>>>>> .
>>>>>>>> .
>>>>>>>> .
>>>>>>>>
>>>>>>>> ....................  90%
>>>>>>>> 2017-02-26 00:03:16.883+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> === Stage7_RS_Backward ===
>>>>>>>> 2017-02-26 00:03:16.885+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> I/Os
>>>>>>>> RelationshipStore
>>>>>>>>   Reads: 3374851294
>>>>>>>>   Random Reads: 2743390177
>>>>>>>>   ScatterIndex: 81
>>>>>>>>
>>>>>>>> 2017-02-26 00:03:16.886+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> Counts:
>>>>>>>>   10338005177 skipCheck
>>>>>>>>   1697668360 missCheck
>>>>>>>>   5621138678 checked
>>>>>>>>   10338005177 correctSkipCheck
>>>>>>>>   1688855306 skipBackup
>>>>>>>>   3951022795 overwrite
>>>>>>>>   2247865 noCacheSkip
>>>>>>>>   239346598 activeCache
>>>>>>>>   119509521 clearCache
>>>>>>>>   2429587416 relSourcePrevCheck
>>>>>>>>   995786837 relSourceNextCheck
>>>>>>>>   2058354842 relTargetPrevCheck
>>>>>>>>   137409583 relTargetNextCheck
>>>>>>>>   6917470274 forwardLinks
>>>>>>>>   7991190672 backLinks
>>>>>>>>   1052730774 nullLinks
>>>>>>>> 2017-02-26 00:03:16.887+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> Memory[used:1.09 GB, free:1.07 GB, total:2.17 GB, max:26.67 GB]
>>>>>>>> 2017-02-26 00:03:16.887+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> Done in  1h 36m 37s 219ms
>>>>>>>> .........2017-02-26 00:23:26.188+0000 INFO 
>>>>>>>>  [o.n.c.ConsistencyCheckService] === RelationshipGroupStore-RelGrp ===
>>>>>>>> 2017-02-26 00:23:26.189+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> I/Os
>>>>>>>> NodeStore
>>>>>>>>   Reads: 231527337
>>>>>>>>   Random Reads: 228593774
>>>>>>>>   ScatterIndex: 98
>>>>>>>> RelationshipStore
>>>>>>>>   Reads: 420334193
>>>>>>>>   Random Reads: 143404207
>>>>>>>>   ScatterIndex: 34
>>>>>>>> RelationshipGroupStore
>>>>>>>>   Reads: 409845841
>>>>>>>>   Random Reads: 105935972
>>>>>>>>   ScatterIndex: 25
>>>>>>>>
>>>>>>>> 2017-02-26 00:23:26.189+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> Counts:
>>>>>>>> 2017-02-26 00:23:26.190+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> Memory[used:751.21 MB, free:1.29 GB, total:2.02 GB, max:26.67 GB]
>>>>>>>> 2017-02-26 00:23:26.191+0000 INFO  [o.n.c.ConsistencyCheckService] 
>>>>>>>> Done in  20m 9s 303ms
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-11" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:131)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391)
>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535)
>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77)
>>>>>>>> at 
>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137)
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-21" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnumFrame.<init>(OrdsSegmentTermsEnumFrame.java:52)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.<init>(OrdsSegmentTermsEnum.java:84)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsFieldReader.iterator(OrdsFieldReader.java:141)
>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:93)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535)
>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77)
>>>>>>>> at 
>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137)
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-8" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.getFrame(OrdsSegmentTermsEnum.java:128)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.pushFrame(OrdsSegmentTermsEnum.java:158)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:391)
>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535)
>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77)
>>>>>>>> at 
>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137)
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-46" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.newOutput(FSTOrdsOutputs.java:225)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.add(FSTOrdsOutputs.java:162)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.seekExact(OrdsSegmentTermsEnum.java:450)
>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.ConstantScoreQuery.createWeight(ConstantScoreQuery.java:119)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.BooleanQuery.createWeight(BooleanQuery.java:239)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createWeight(IndexSearcher.java:904)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.createNormalizedWeight(IndexSearcher.java:887)
>>>>>>>> at 
>>>>>>>> org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:535)
>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.java:171)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCheck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.dispatch(ConsistencyReporter.java:124)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.report.ConsistencyReporter.forNode(ConsistencyReporter.java:440)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.PropertyAndNode2LabelIndexProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>> at 
>>>>>>>> org.neo4j.consistency.checking.full.RecordCheckWorker.run(RecordCheckWorker.java:77)
>>>>>>>> at 
>>>>>>>> org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.Workers$Worker.run(Workers.java:137)
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-22" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-10" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-40" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-58" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-61" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-18" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-25" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-45" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-28" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-50" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-39" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-51" 
>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>
>>>>>>>> -- 
>>>>>>>> You received this message because you are subscribed to the Google 
>>>>>>>> Groups "Neo4j" group.
>>>>>>>> To unsubscribe from this group and stop receiving emails from it, 
>>>>>>>> send an email to [email protected].
>>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>>
>>>>>>>
>>>>>>> -- 
>>>>> You received this message because you are subscribed to the Google 
>>>>> Groups "Neo4j" group.
>>>>> To unsubscribe from this group and stop receiving emails from it, send 
>>>>> an email to [email protected].
>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>
>>>>
>>>> -- 
>>> You received this message because you are subscribed to the Google 
>>> Groups "Neo4j" group.
>>> To unsubscribe from this group and stop receiving emails from it, send 
>>> an email to [email protected].
>>> For more options, visit https://groups.google.com/d/optout.
>>>
>>>

-- 
You received this message because you are subscribed to the Google Groups 
"Neo4j" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to