Can you try to edit the script directly and add the memory parameters there?

On Fri, Mar 3, 2017 at 8:49 PM, unrealadmin23 via Neo4j <
neo4j@googlegroups.com> wrote:

> Yes
>
> Also, in the 90% scan, what every Java memory parameter I use,   htop
> shows the same memory foot print.   Its as if the heap isn't being set as
> per the env parameters that you area asking me to set.
>
> Wayne
>
>
> On Friday, 3 March 2017 07:51:48 UTC, Mattias Persson wrote:
>>
>> Querying Lucene, at the very least the way consistency checker uses it,
>> has a drawback that all matching documents will be read and kept in heap
>> before going through them.
>>
>> So let me ask you something about your data: are there certain property
>> values that are very common and also indexed?
>>
>> On Thursday, March 2, 2017 at 7:07:31 PM UTC+1, unreal...@googlemail.com
>> wrote:
>>>
>>> It appears not:
>>>
>>> $env
>>> JAVA_MEMORY_OPTS=-Xmx32G -Xms32G
>>>
>>> .
>>> .
>>> .
>>>
>>>
>>> ....................  90%
>>> 2017-03-01 23:24:55.705+0000 INFO  [o.n.c.ConsistencyCheckService] ===
>>> Stage7_RS_Backward ===
>>> 2017-03-01 23:24:55.706+0000 INFO  [o.n.c.ConsistencyCheckService] I/Os
>>> RelationshipStore
>>>   Reads: 3373036269 <(337)%20303-6269>
>>>   Random Reads: 2732592348
>>>   ScatterIndex: 81
>>>
>>> 2017-03-01 23:24:55.707+0000 INFO  [o.n.c.ConsistencyCheckService]
>>> Counts:
>>>   10338061780 skipCheck
>>>   1697668359 missCheck
>>>   5621138678 checked
>>>   10338061780 correctSkipCheck
>>>   1688855306 skipBackup
>>>   3951022794 overwrite
>>>   2191262 noCacheSkip
>>>   239346600 activeCache
>>>   119509522 clearCache
>>>   2429587416 relSourcePrevCheck
>>>   995786837 relSourceNextCheck
>>>   2058354842 <(205)%20835-4842> relTargetPrevCheck
>>>   137409583 relTargetNextCheck
>>>   6917470274 forwardLinks
>>>   7991190672 backLinks
>>>   1052730774 nullLinks
>>> 2017-03-01 23:24:55.708+0000 INFO  [o.n.c.ConsistencyCheckService]
>>> Memory[used:404.70 MB, free:1.63 GB, total:2.03 GB, max:26.67 GB]
>>> 2017-03-01 23:24:55.708+0000 INFO  [o.n.c.ConsistencyCheckService] Done
>>> in  1h 37m 39s 828ms
>>> .........2017-03-01 23:45:36.032+0000 INFO  [o.n.c.ConsistencyCheckService]
>>> === RelationshipGroupStore-RelGrp ===
>>> 2017-03-01 23:45:36.032+0000 INFO  [o.n.c.ConsistencyCheckService] I/Os
>>> RelationshipGroupStore
>>>   Reads: 410800979
>>>   Random Reads: 102164662
>>>   ScatterIndex: 24
>>> NodeStore
>>>   Reads: 229862945
>>>   Random Reads: 226895703
>>>   ScatterIndex: 98
>>> RelationshipStore
>>>   Reads: 423304043
>>>   Random Reads: 139746630
>>>   ScatterIndex: 33
>>>
>>> 2017-03-01 23:45:36.032+0000 INFO  [o.n.c.ConsistencyCheckService]
>>> Counts:
>>> 2017-03-01 23:45:36.033+0000 INFO  [o.n.c.ConsistencyCheckService]
>>> Memory[used:661.75 MB, free:1.39 GB, total:2.03 GB, max:26.67 GB]
>>> 2017-03-01 23:45:36.034+0000 INFO  [o.n.c.ConsistencyCheckService] Done
>>> in  20m 40s 326ms
>>> .Exception in thread "ParallelRecordScanner-Stage8_PS_Props-19"
>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>> at org.apache.lucene.util.BytesRef.<init>(BytesRef.java:73)
>>> at org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(
>>> FSTOrdsOutputs.java:181)
>>> at org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.read(
>>> FSTOrdsOutputs.java:32)
>>> at org.apache.lucene.util.fst.Outputs.readFinalOutput(Outputs.java:77)
>>> at org.apache.lucene.util.fst.FST.readNextRealArc(FST.java:1094)
>>> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1262)
>>> at org.apache.lucene.util.fst.FST.findTargetArc(FST.java:1186)
>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>> seekExact(OrdsSegmentTermsEnum.java:405)
>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.java:192)
>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>> rcher.java:904)
>>> at org.apache.lucene.search.ConstantScoreQuery.createWeight(Con
>>> stantScoreQuery.java:119)
>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>> rcher.java:904)
>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.java:57)
>>>
>>>
>>> I have also tried larger memory values.
>>>
>>> Wayne.
>>>
>>>
>>> On Wednesday, 1 March 2017 01:52:47 UTC, Michael Hunger wrote:
>>>>
>>>> Sorry I just learned that neo4j-admin uses a different variable
>>>>
>>>> "You can pass memory options to the JVM via the `JAVA_MEMORY_OPTS`
>>>> variable as a workaround though."
>>>>
>>>>
>>>>
>>>> Von meinem iPhone gesendet
>>>>
>>>> Am 28.02.2017 um 18:50 schrieb unrealadmin23 via Neo4j <
>>>> ne...@googlegroups.com>:
>>>>
>>>> Michael,
>>>>
>>>> After running the check_consistency command for 1 day with the above
>>>> parameters, it failed in exactly the same manner.
>>>>
>>>> $env | grep -i java
>>>> JAVA_OPTS=-Xmx32G -Xms32G
>>>>
>>>> Any other ideas ?
>>>>
>>>> Wayne
>>>>
>>>>
>>>> On Monday, 27 February 2017 16:57:49 UTC, Michael Hunger wrote:
>>>>>
>>>>> Do you have really that much RAM in your machine ? 120G usually
>>>>> doesn't make sense. Most people run with 32G as large heap.
>>>>>
>>>>> That said. I asked and currently the numbers from the config are not
>>>>> used, you have to do:
>>>>>
>>>>> export JAVA_OPTS=-Xmx24G -Xms24G
>>>>> neo4j-admin ...
>>>>>
>>>>>
>>>>> On Mon, Feb 27, 2017 at 8:32 AM, unrealadmin23 via Neo4j <
>>>>> ne...@googlegroups.com> wrote:
>>>>>
>>>>>>
>>>>>> I should have said, that the head sizes are the ones that I have set
>>>>>> in neo4j.conf.
>>>>>>
>>>>>> Will these be used by check-consistency or do I need to supply them
>>>>>> elsewhere ?
>>>>>>
>>>>>> Wayne.
>>>>>>
>>>>>>
>>>>>> On Monday, 27 February 2017 07:27:33 UTC, unreal...@googlemail.com
>>>>>> wrote:
>>>>>>>
>>>>>>> Michael,
>>>>>>>
>>>>>>> neo4j-admin check-consistency --database=test.db --verbose
>>>>>>>
>>>>>>> dbms.memory.heap.initial_size=120000m
>>>>>>> dbms.memory.heap.max_size=120000m
>>>>>>>
>>>>>>> Wayne.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Monday, 27 February 2017 02:47:26 UTC, Michael Hunger wrote:
>>>>>>>>
>>>>>>>> How did you call the consistency checker?
>>>>>>>>
>>>>>>>> How much heap did you provide for it?
>>>>>>>>
>>>>>>>> Cheers, Michael
>>>>>>>>
>>>>>>>>
>>>>>>>> On Sun, Feb 26, 2017 at 8:28 PM, unrealadmin23 via Neo4j <
>>>>>>>> ne...@googlegroups.com> wrote:
>>>>>>>>
>>>>>>>>> The following o/p was obtained:
>>>>>>>>>
>>>>>>>>> .
>>>>>>>>> .
>>>>>>>>> .
>>>>>>>>>
>>>>>>>>> ....................  90%
>>>>>>>>> 2017-02-26 00:03:16.883+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> === Stage7_RS_Backward ===
>>>>>>>>> 2017-02-26 00:03:16.885+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> I/Os
>>>>>>>>> RelationshipStore
>>>>>>>>>   Reads: 3374851294
>>>>>>>>>   Random Reads: 2743390177
>>>>>>>>>   ScatterIndex: 81
>>>>>>>>>
>>>>>>>>> 2017-02-26 00:03:16.886+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> Counts:
>>>>>>>>>   10338005177 skipCheck
>>>>>>>>>   1697668360 missCheck
>>>>>>>>>   5621138678 checked
>>>>>>>>>   10338005177 correctSkipCheck
>>>>>>>>>   1688855306 skipBackup
>>>>>>>>>   3951022795 overwrite
>>>>>>>>>   2247865 noCacheSkip
>>>>>>>>>   239346598 activeCache
>>>>>>>>>   119509521 clearCache
>>>>>>>>>   2429587416 relSourcePrevCheck
>>>>>>>>>   995786837 relSourceNextCheck
>>>>>>>>>   2058354842 relTargetPrevCheck
>>>>>>>>>   137409583 relTargetNextCheck
>>>>>>>>>   6917470274 forwardLinks
>>>>>>>>>   7991190672 backLinks
>>>>>>>>>   1052730774 nullLinks
>>>>>>>>> 2017-02-26 00:03:16.887+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> Memory[used:1.09 GB, free:1.07 GB, total:2.17 GB, max:26.67 GB]
>>>>>>>>> 2017-02-26 00:03:16.887+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> Done in  1h 36m 37s 219ms
>>>>>>>>> .........2017-02-26 00:23:26.188+0000 INFO
>>>>>>>>>  [o.n.c.ConsistencyCheckService] ===
>>>>>>>>> RelationshipGroupStore-RelGrp ===
>>>>>>>>> 2017-02-26 00:23:26.189+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> I/Os
>>>>>>>>> NodeStore
>>>>>>>>>   Reads: 231527337
>>>>>>>>>   Random Reads: 228593774
>>>>>>>>>   ScatterIndex: 98
>>>>>>>>> RelationshipStore
>>>>>>>>>   Reads: 420334193
>>>>>>>>>   Random Reads: 143404207
>>>>>>>>>   ScatterIndex: 34
>>>>>>>>> RelationshipGroupStore
>>>>>>>>>   Reads: 409845841
>>>>>>>>>   Random Reads: 105935972
>>>>>>>>>   ScatterIndex: 25
>>>>>>>>>
>>>>>>>>> 2017-02-26 00:23:26.189+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> Counts:
>>>>>>>>> 2017-02-26 00:23:26.190+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> Memory[used:751.21 MB, free:1.29 GB, total:2.02 GB, max:26.67 GB]
>>>>>>>>> 2017-02-26 00:23:26.191+0000 INFO  [o.n.c.ConsistencyCheckService]
>>>>>>>>> Done in  20m 9s 303ms
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-11"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>>>>>>>> getFrame(OrdsSegmentTermsEnum.java:131)
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>>>>>>>> pushFrame(OrdsSegmentTermsEnum.java:158)
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>>>>>>>> seekExact(OrdsSegmentTermsEnum.java:391)
>>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.
>>>>>>>>> java:192)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.ConstantScoreQuery.createWeight(Con
>>>>>>>>> stantScoreQuery.java:119)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.
>>>>>>>>> java:57)
>>>>>>>>> at org.apache.lucene.search.BooleanQuery.createWeight(BooleanQu
>>>>>>>>> ery.java:239)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createNormalizedWeigh
>>>>>>>>> t(IndexSearcher.java:887)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.
>>>>>>>>> java:535)
>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.j
>>>>>>>>> ava:171)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>>> at org.neo4j.consistency.report.ConsistencyReporter.dispatch(Co
>>>>>>>>> nsistencyReporter.java:124)
>>>>>>>>> at org.neo4j.consistency.report.ConsistencyReporter.forNode(Con
>>>>>>>>> sistencyReporter.java:440)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNode2LabelInd
>>>>>>>>> exProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNode2LabelInd
>>>>>>>>> exProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>>> at org.neo4j.consistency.checking.full.RecordCheckWorker.run(Re
>>>>>>>>> cordCheckWorker.java:77)
>>>>>>>>> at org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.
>>>>>>>>> Workers$Worker.run(Workers.java:137)
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-21"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnumF
>>>>>>>>> rame.<init>(OrdsSegmentTermsEnumFrame.java:52)
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>>>>>>>> <init>(OrdsSegmentTermsEnum.java:84)
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsFieldReader.itera
>>>>>>>>> tor(OrdsFieldReader.java:141)
>>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:93)
>>>>>>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.
>>>>>>>>> java:192)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.
>>>>>>>>> java:57)
>>>>>>>>> at org.apache.lucene.search.BooleanQuery.createWeight(BooleanQu
>>>>>>>>> ery.java:239)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createNormalizedWeigh
>>>>>>>>> t(IndexSearcher.java:887)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.
>>>>>>>>> java:535)
>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.j
>>>>>>>>> ava:171)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>>> at org.neo4j.consistency.report.ConsistencyReporter.dispatch(Co
>>>>>>>>> nsistencyReporter.java:124)
>>>>>>>>> at org.neo4j.consistency.report.ConsistencyReporter.forNode(Con
>>>>>>>>> sistencyReporter.java:440)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNode2LabelInd
>>>>>>>>> exProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNode2LabelInd
>>>>>>>>> exProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>>> at org.neo4j.consistency.checking.full.RecordCheckWorker.run(Re
>>>>>>>>> cordCheckWorker.java:77)
>>>>>>>>> at org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.
>>>>>>>>> Workers$Worker.run(Workers.java:137)
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-8"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>>>>>>>> getFrame(OrdsSegmentTermsEnum.java:128)
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>>>>>>>> pushFrame(OrdsSegmentTermsEnum.java:158)
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>>>>>>>> seekExact(OrdsSegmentTermsEnum.java:391)
>>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.
>>>>>>>>> java:192)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.ConstantScoreQuery.createWeight(Con
>>>>>>>>> stantScoreQuery.java:119)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.
>>>>>>>>> java:57)
>>>>>>>>> at org.apache.lucene.search.BooleanQuery.createWeight(BooleanQu
>>>>>>>>> ery.java:239)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createNormalizedWeigh
>>>>>>>>> t(IndexSearcher.java:887)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.
>>>>>>>>> java:535)
>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.j
>>>>>>>>> ava:171)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>>> at org.neo4j.consistency.report.ConsistencyReporter.dispatch(Co
>>>>>>>>> nsistencyReporter.java:124)
>>>>>>>>> at org.neo4j.consistency.report.ConsistencyReporter.forNode(Con
>>>>>>>>> sistencyReporter.java:440)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNode2LabelInd
>>>>>>>>> exProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNode2LabelInd
>>>>>>>>> exProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>>> at org.neo4j.consistency.checking.full.RecordCheckWorker.run(Re
>>>>>>>>> cordCheckWorker.java:77)
>>>>>>>>> at org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.
>>>>>>>>> Workers$Worker.run(Workers.java:137)
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-46"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.newOut
>>>>>>>>> put(FSTOrdsOutputs.java:225)
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.FSTOrdsOutputs.add(
>>>>>>>>> FSTOrdsOutputs.java:162)
>>>>>>>>> at org.apache.lucene.codecs.blocktreeords.OrdsSegmentTermsEnum.
>>>>>>>>> seekExact(OrdsSegmentTermsEnum.java:450)
>>>>>>>>> at org.apache.lucene.index.TermContext.build(TermContext.java:94)
>>>>>>>>> at org.apache.lucene.search.TermQuery.createWeight(TermQuery.
>>>>>>>>> java:192)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.ConstantScoreQuery.createWeight(Con
>>>>>>>>> stantScoreQuery.java:119)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.BooleanWeight.<init>(BooleanWeight.
>>>>>>>>> java:57)
>>>>>>>>> at org.apache.lucene.search.BooleanQuery.createWeight(BooleanQu
>>>>>>>>> ery.java:239)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createWeight(IndexSea
>>>>>>>>> rcher.java:904)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.createNormalizedWeigh
>>>>>>>>> t(IndexSearcher.java:887)
>>>>>>>>> at org.apache.lucene.search.IndexSearcher.search(IndexSearcher.
>>>>>>>>> java:535)
>>>>>>>>> at org.neo4j.kernel.api.impl.schema.reader.SimpleIndexReader.co
>>>>>>>>> untIndexedNodes(SimpleIndexReader.java:136)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.verifyNodeCorrectlyIndexed(PropertyAndNodeIndexedCheck.j
>>>>>>>>> ava:171)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.checkIndexToLabels(PropertyAndNodeIndexedCheck.java:113)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.check(PropertyAndNodeIndexedCheck.java:71)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNodeIndexedCh
>>>>>>>>> eck.check(PropertyAndNodeIndexedCheck.java:48)
>>>>>>>>> at org.neo4j.consistency.report.ConsistencyReporter.dispatch(Co
>>>>>>>>> nsistencyReporter.java:124)
>>>>>>>>> at org.neo4j.consistency.report.ConsistencyReporter.forNode(Con
>>>>>>>>> sistencyReporter.java:440)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNode2LabelInd
>>>>>>>>> exProcessor.process(PropertyAndNode2LabelIndexProcessor.java:63)
>>>>>>>>> at org.neo4j.consistency.checking.full.PropertyAndNode2LabelInd
>>>>>>>>> exProcessor.process(PropertyAndNode2LabelIndexProcessor.java:39)
>>>>>>>>> at org.neo4j.consistency.checking.full.RecordCheckWorker.run(Re
>>>>>>>>> cordCheckWorker.java:77)
>>>>>>>>> at org.neo4j.unsafe.impl.batchimport.cache.idmapping.string.
>>>>>>>>> Workers$Worker.run(Workers.java:137)
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-22"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-10"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-40"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-58"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-61"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-18"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-25"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-45"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-28"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-50"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-39"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>> Exception in thread "ParallelRecordScanner-Stage8_PS_Props-51"
>>>>>>>>> java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> You received this message because you are subscribed to the Google
>>>>>>>>> Groups "Neo4j" group.
>>>>>>>>> To unsubscribe from this group and stop receiving emails from it,
>>>>>>>>> send an email to neo4j+un...@googlegroups.com.
>>>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>> You received this message because you are subscribed to the Google
>>>>>> Groups "Neo4j" group.
>>>>>> To unsubscribe from this group and stop receiving emails from it,
>>>>>> send an email to neo4j+un...@googlegroups.com.
>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>>>
>>>>>
>>>>> --
>>>> You received this message because you are subscribed to the Google
>>>> Groups "Neo4j" group.
>>>> To unsubscribe from this group and stop receiving emails from it, send
>>>> an email to neo4j+un...@googlegroups.com.
>>>> For more options, visit https://groups.google.com/d/optout.
>>>>
>>>> --
> You received this message because you are subscribed to the Google Groups
> "Neo4j" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to neo4j+unsubscr...@googlegroups.com.
> For more options, visit https://groups.google.com/d/optout.
>

-- 
You received this message because you are subscribed to the Google Groups 
"Neo4j" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to neo4j+unsubscr...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to