I am very sorry to point out that those Lucene queries actually have a changeable behaviour, usually they are slower than the past, and in most case I get that error again (using 6GB heap of 8GB total RAM).
I am updating the graph with delete and update of nodes and relationships. I decreased the number of operations per transaction but most of times I still got this error. The insertion with Batch Inserter instead seems to be ok! What do you suggest me please? Keep 1.9.9 version or upgrade to 2.2.1 could be a solution? The new version covers packages related to these problems? Thanks in advance Rita Il giorno giovedì 16 aprile 2015 14:22:34 UTC+2, Rita ha scritto: > > Thank you for the reply Michael. I have just published the issue. > I was using as usual -Xmx4g. Now I've just tried with 6GB and open and > close a single transaction for every query like that on the different > indexes and I do not get this exception. > So now it has more need of memory for the same operation. I try on other > cases. Tell me if there are news on the issue please. > Thank you. > > Regards > Rita > > Il giorno giovedì 16 aprile 2015 13:04:01 UTC+2, Michael Hunger ha scritto: >> >> This seems to be a like a bug. >> >> How much heap do you have? >> >> Could you raise an issue on github.com/neo4j/neo4j/issues ? >> >> Thanks so much >> >> Michael >> >> Am 16.04.2015 um 11:26 schrieb Rita <rita...@gmail.com>: >> >> Hi all, >> I am passing from Neo4j 1.9.9 to Neo4j 2.2.0, with embedded mode using >> java. I have inserted the transactions also for read operations, but now >> when I query my Lucene indexes as this >> >> rhits = index.query("cs", "*"); >> out.println("#" + rhits.size()); >> rhits.close(); >> >> >> as you can see I do not have to iterate over the result, I need only the >> number of results but this new Neo4j version looks like loading all in >> memory and I get the following error in the first instruction. >> >> Exception in thread "main" java.lang.OutOfMemoryError: Java heap space >> at org.neo4j.collection.primitive.hopscotch.IntArrayBasedKeyTable >> .initia >> lizeTable(IntArrayBasedKeyTable.java:54) >> at org.neo4j.collection.primitive.hopscotch.IntArrayBasedKeyTable >> .<init> >> (IntArrayBasedKeyTable.java:48) >> at org.neo4j.collection.primitive.hopscotch.LongKeyTable.<init>( >> LongKeyT >> able.java:27) >> at org.neo4j.collection.primitive.Primitive.longSet(Primitive. >> java:66) >> at org.neo4j.kernel.impl.coreapi.LegacyIndexProxy$1.<init>( >> LegacyIndexPr >> oxy.java:296) >> at org.neo4j.kernel.impl.coreapi.LegacyIndexProxy.wrapIndexHits( >> LegacyIn >> dexProxy.java:294) >> at org.neo4j.kernel.impl.coreapi.LegacyIndexProxy.query( >> LegacyIndexProxy >> .java:352) >> >> I never get this with older versions of Neo4j! I always did this >> operation until version 1.9.9. >> Could you please help me to avoid this? Is it a bug of library >> implementation or I have to change the way of querying? >> >> Thanks in advance, >> Rita >> >> -- >> You received this message because you are subscribed to the Google Groups >> "Neo4j" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to neo4j+un...@googlegroups.com. >> For more options, visit https://groups.google.com/d/optout. >> >> >> -- You received this message because you are subscribed to the Google Groups "Neo4j" group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.