[Neo4j] HTTP Status Codes and Errors
Hi, When using the transactional endpoint, when an error occurs, such as for instance a unique constraint violation, the status code returned is still 200. Is this by design? Is the preferred method to always have to examine errors property in body to make sure it's empty to guarantee success? Thanks. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.
Re: [Neo4j] HTTP Status Codes and Errors
Yes it is, as the results are streamed, the headers are already sent out immediately before query execution happens. Also as you can send many queries, the error contains more information on which (the last) query it happened. There is an error field in the response. Not sure how much effort it would be to rewrite the impl to wait for the first query to start and on failure during parsing/semantic check send an appropriate error code (the query results are also streamed directly so if an error happens during fetching query results the same as above applies). But then the behavior would be inconsistent depending on your offending query is the first or second which would be even more confusing. Michael Am 16.06.2014 um 23:07 schrieb Hadi Hariri hadihar...@gmail.com: Hi, When using the transactional endpoint, when an error occurs, such as for instance a unique constraint violation, the status code returned is still 200. Is this by design? Is the preferred method to always have to examine errors property in body to make sure it's empty to guarantee success? Thanks. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.
Re: [Neo4j] Re: dijkstra bidirectional
Sorry for not replying but time is a scarse resource :( I don't expect my getting time for this the closest months. Perhaps there are others willing to help out! Take care Best, Mattias Den 10 jun 2014 12:06 skrev Antonio Grimaldi antonio.grimaldim...@gmail.com: Is* org.neo4j.graphalgo.impl.shortestpath.DijkstraCostType a *Bidirectional Dijkstra's implementation??? Il giorno giovedì 8 maggio 2014 17:04:32 UTC+2, Antonio Grimaldi ha scritto: Hi, Is there an implementation of Bidirectional Dijkstra? Thanks Antonio -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.
[Neo4j] Re: Neo4j database ALWAYS shuts down incorrectly if start/stop as a service from a list of windows services
Here is a stack trace that I get when I try to do upgrade from 1.9.5 to 2.0: 2014-06-17 09:48:27.319+ INFO [API] Setting startup timeout to: 12ms based on -1 Detected incorrectly shut down database, performing recovery.. 2014-06-17 09:48:28.108+ DEBUG [API] org.neo4j.server.ServerStartupException: Starting Neo4j Server failed: Error starting org.neo4j.kernel.EmbeddedGraphDatabase, D:\Neo4j\neo4j-enterprise-2.0.0\data\graph.db at org.neo4j.server.AbstractNeoServer.start(AbstractNeoServer.java:209) ~[neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.Bootstrapper.start(Bootstrapper.java:87) [neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.Bootstrapper.main(Bootstrapper.java:50) [neo4j-server-2.0.0.jar:2.0.0] Caused by: java.lang.RuntimeException: Error starting org.neo4j.kernel.EmbeddedGraphDatabase, D:\Neo4j\neo4j-enterprise-2.0.0\data\graph.db at org.neo4j.kernel.InternalAbstractGraphDatabase.run(InternalAbstractGraphDatabase.java:333) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.EmbeddedGraphDatabase.init(EmbeddedGraphDatabase.java:63) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.graphdb.factory.GraphDatabaseFactory$1.newDatabase(GraphDatabaseFactory.java:92) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.graphdb.factory.GraphDatabaseBuilder.newGraphDatabase(GraphDatabaseBuilder.java:198) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.impl.recovery.StoreRecoverer.recover(StoreRecoverer.java:115) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.server.preflight.PerformRecoveryIfNecessary.run(PerformRecoveryIfNecessary.java:59) ~[neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.preflight.PreFlightTasks.run(PreFlightTasks.java:70) ~[neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.AbstractNeoServer.runPreflightTasks(AbstractNeoServer.java:319) ~[neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.AbstractNeoServer.start(AbstractNeoServer.java:144) ~[neo4j-server-2.0.0.jar:2.0.0] ... 2 common frames omitted Caused by: org.neo4j.kernel.lifecycle.LifecycleException: Component 'org.neo4j.kernel.impl.transaction.XaDataSourceManager@2b1eb67d' was successfully initialized, but failed to start. Please see attached cause exception. at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.start(LifeSupport.java:504) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.lifecycle.LifeSupport.start(LifeSupport.java:115) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.InternalAbstractGraphDatabase.run(InternalAbstractGraphDatabase.java:310) ~[neo4j-kernel-2.0.0.jar:2.0.0] ... 10 common frames omitted Caused by: org.neo4j.kernel.lifecycle.LifecycleException: Component 'org.neo4j.kernel.impl.nioneo.xa.NeoStoreXaDataSource@1bf5df6a' was successfully initialized, but failed to start. P lease see attached cause exception. at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.start(LifeSupport.java:504) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.lifecycle.LifeSupport.start(LifeSupport.java:115) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.impl.transaction.XaDataSourceManager.start(XaDataSourceManager.java:164) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.start(LifeSupport.java:498) ~[neo4j-kernel-2.0.0.jar:2.0.0] ... 12 common frames omitted Caused by: org.neo4j.kernel.impl.storemigration.StoreUpgrader$UpgradingStoreVersionNotFoundException: 'neostore' does not contain a store version, please ensure that the original datab ase was shut down in a clean state. at org.neo4j.kernel.impl.storemigration.UpgradableDatabase.checkUpgradeable(UpgradableDatabase.java:85) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.impl.storemigration.StoreUpgrader.attemptUpgrade(StoreUpgrader.java:72) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.impl.nioneo.store.StoreFactory.tryToUpgradeStores(StoreFactory.java:143) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.impl.nioneo.store.StoreFactory.newNeoStore(StoreFactory.java:123) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.impl.nioneo.xa.NeoStoreXaDataSource.start(NeoStoreXaDataSource.java:323) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.start(LifeSupport.java:498) ~[neo4j-kernel-2.0.0.jar:2.0.0] ... 15 common frames omitted 2014-06-17 09:48:28.109+ DEBUG [API] Failed to start Neo Server on port [7474] -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.
[Neo4j] Upgrading neo4j 1.9.3 to 2.0.3 fails with Invalid log format version found, expected 3 but was 2
Hi, I have been trying to upgrade neo4j from 1.9.3 to 2.0.3. SDN from 2.3.1.RELEASE to 3.1.0.RELEASE. Followed the steps listed @ http://docs.neo4j.org/chunked/stable/deployment-upgrading.html#explicit-upgrade I try bringing up the server with the upgrade configuration.There are a few new folders created in the db store. One among which is upgrade_backup and messages log there says upgrade happened. 2014-06-17 07:16:55.286+ INFO [o.n.k.i.DiagnosticsManager]: --- INITIALIZED diagnostics END --- 2014-06-17 07:17:00.216+ INFO [o.n.k.i.n.s.StoreFactory]: Starting upgrade of database store files 2014-06-17 07:17:00.225+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 10% complete 2014-06-17 07:17:00.228+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 20% complete 2014-06-17 07:17:00.231+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 30% complete 2014-06-17 07:17:00.233+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 40% complete 2014-06-17 07:17:00.236+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 50% complete 2014-06-17 07:17:00.239+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 60% complete 2014-06-17 07:17:00.241+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 70% complete 2014-06-17 07:17:00.244+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 80% complete 2014-06-17 07:17:00.247+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 90% complete 2014-06-17 07:17:00.249+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 100% complete 2014-06-17 07:17:03.776+ INFO [o.n.k.i.n.s.StoreFactory]: Finished upgrade of database store files But I get the error with log/index. Exception when stopping org.neo4j.index.impl.lucene.LuceneDataSource@42a792f0 org.neo4j.kernel.impl.tran saction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but was 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database java.lang.RuntimeException: org.neo4j.kernel.impl.transaction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but wa s 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy$1.reached(LogPruneStrategies.java:250) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$AbstractPruneStrategy.prune(LogPruneStrategies.java:78) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy.prune(LogPruneStrategies.java:222) at org.neo4j.kernel.impl.transaction.xaframework.XaLogicalLog.close(XaLogicalLog.java:742) at org.neo4j.kernel.impl.transaction.xaframework.LogBackedXaDataSource.stop(LogBackedXaDataSource.java:69) at org.neo4j.index.impl.lucene.LuceneDataSource.stop(LuceneDataSource.java:310) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.shutdown(LifeSupport.java:547) at org.neo4j.kernel.lifecycle.LifeSupport.remove(LifeSupport.java:339) at org.neo4j.kernel.impl.transaction.XaDataSourceManager.unregisterDataSource(XaDataSourceManager.java:272) at org.neo4j.index.lucene.LuceneKernelExtension.stop(LuceneKernelExtension.java:92) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport.stop(LifeSupport.java:155) at org.neo4j.kernel.extension.KernelExtensions.stop(KernelExtensions.java:124) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport.stop(LifeSupport.java:155) at org.neo4j.kernel.lifecycle.LifeSupport.shutdown(LifeSupport.java:185) at org.neo4j.kernel.InternalAbstractGraphDatabase.shutdown(InternalAbstractGraphDatabase.java:801) at org.springframework.data.neo4j.support.DelegatingGraphDatabase.shutdown(DelegatingGraphDatabase.java:270) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.springframework.beans.factory.support.DisposableBeanAdapter.invokeCustomDestroyMethod(DisposableBeanAdapter.java:327) at org.springframework.beans.factory.support.DisposableBeanAdapter.destroy(DisposableBeanAdapter.java:253) at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.destroyBean(DefaultSingletonBeanRegistry.java:510) at
Re: [Neo4j] LOAD CSV creates nodes but does not set properties
if they don't have a value for city id, do they then have empty columns there still? like user-id,, You probably want to filter these rows? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c WHERE coalesce(c.CitiId,) ... Am 17.06.2014 um 11:23 schrieb Paul Damian pauldamia...@gmail.com: Well, the csv file contains some rows that do not have a value for CityId, and the rows are unique regarding the clientID. There are 11M clients living in 14K Cities. Is there a limit of links/node? Now I've created a piece of code that reads from file and creates each relationship, but, as you can imagine, it works really slow in this scenario. did you create an index on :Client(Id) and :City(Id) what happens if you do: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) each count should be equivalent to the # of rows in the file. Michael Am 16.06.2014 um 17:47 schrieb Paul Damian paulda...@gmail.com: Somehow I've managed to load all the nodes and now I'm trying to load the links as well. I read the nodes from csv file and create the relation between them. I run the following command: USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}), (city: City { Id: toInt(c.CityId)}) CREATE (client)-[r:LOCATED_IN]-(city) Running with a smaller commit size returns this error Neo.DatabaseError.Statement.ExecutionFailure, while increasing the commit size to 1 throws Neo.DatabaseError.General.UnknownFailure. Can you help me with this? joi, 5 iunie 2014, 12:05:18 UTC+3, Michael Hunger a scris: Perhaps something with field or line terminators? I assume it blows up the field separation. Try to run: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c RETURN { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone: c.Phone, Fax: c.Fax, BusinessName: c.BusinessName, URL: c.URL, Latitude: toFloat(c.Latitude), Longitude: toFloat(c.Longitude), AgencyId: toInt(c.AgencyId), RowStatus: toInt(c.RowStatus)} as data, c as line LIMIT 3 On Thu, Jun 5, 2014 at 10:51 AM, Paul Damian paulda...@gmail.com wrote: I've tried using the shell and I get the same results: nodes with no properties. I've created the csv file using MsSQL Server Export. Is it relevant? About you curiosity: I figured I would import first the nodes, then the relationships from the connection tables. Am I doing it wrong? Thanks joi, 5 iunie 2014, 09:54:31 UTC+3, Michael Hunger a scris: I'd probably use a commit size in your case of 50k or 100k. Try to use the neo4j-shell and not the web-interface. Connect to neo4j using bin/neo4j-shell Then run your commands ending with a semicolon. Just curious: Your data is imported as one node per row? That's not really a graph structure. On Wed, Jun 4, 2014 at 6:56 PM, Paul Damian paulda...@gmail.com wrote: Hi there, I'm experimenting with Neo4j while benchmarking a bunch of NoSQL databases for my graduation paper. I'm using the web interface to populate the database. I've been able to load the smaller tables from my SQL database and LOAD CSV works fine. By small, I mean a few columns (4-5) and some rows (1 million). However, when I try to upload a larger table (15 columns, 12 million rows), it creates the nodes but it doesn't set any properties. I've tried to reduce the number of records (to 100) and also the number of columns( just the Id property ), but no luck so far. The cypher command used is this one USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c CREATE (:Client { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone: c.Phone, Fax: c.Fax, BusinessName: c.BusinessName, URL: c.URL, Latitude: toFloat(c.Latitude), Longitude: toFloat(c.Longitude), AgencyId: toInt(c.AgencyId), RowStatus: toInt(c.RowStatus)}) Any help and indication is welcomed, Paul -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+un...@googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+un...@googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message
Re: [Neo4j] Hierarchical facets
Hi Michael, have you been able to look at the profiling info that I sent you ? Perhaps we can have a chat on it tomorrow in Amsterdam .. Best, Tom Met vriendelijke groet / With kind regards Ir. T. Zeppenfeldt van der Waalsstraat 30 6706 JR Wageningen The Netherlands Mobile: +31 6 23 28 78 06 Phone: +31 3 17 84 22 17 E-mail: t.zeppenfe...@ophileon.com t.zeppenfe...@ophileon.comWeb: www.ophileon.com Twitter: tomzeppenfeldt Skype: tomzeppenfeldt 2014-05-21 21:23 GMT+02:00 Michael Hunger michael.hun...@neotechnology.com : That's why I suggested UNION. So the two individual queries take 14s? Still way too long. On Wed, May 21, 2014 at 3:40 PM, Tom Zeppenfeldt tomzeppenfe...@gmail.com wrote: I have some problems starting the shell from my Mac Terminal (it's giving me a out of mem error) , but form the webadmin powerconsole. Can't find any documentation either on how to set shell to have me return profile. neo4j-sh (?)$ export termname=Eurovoc *Your first Query* neo4j-sh (?)$ match (j:jurt)-[:HAS_TERM]-()-[:BT*0..]-(t:term)-[:BT]-(t2:term {name:{termname}}) return t.name, count(distinct j) as count order by count desc limit 10 ; == +-+ == | t.name | count | == +-+ == | gezondheidsbeleid | 1823 | == | overtreding | 1393 | == | Europese organisatie | 1389 | == | EU-instantie | 1323 | == | mondiale organisatie | 1277 | == | gespecialiseerde instelling van de VN | 1143 | == | handeling van de EU | 1129 | == | internationaal publiekrecht | 1091 | == | sociaal beleid| 971 | == | rechtsvordering | 915 | == +-+ == 10 rows *== 8775 ms* *Your second Query* neo4j-sh (?)$ match (j:jurt)-[:HAS_TERM]-()-[:BT*0..]-(t2:term {name:{termname}}) return t.name, count(distinct j) as count order by count desc limit 10; == SyntaxException: t not defined (line 1, column 79) == match (j:jurt)-[:HAS_TERM]-()-[:BT*0..]-(t2:term {name:{termname}}) return t.name, count(distinct j) as count order by count desc limit 10 == ^ neo4j-sh (?)$ match (j:jurt)-[:HAS_TERM]-()-[:BT*0..]-(t2:term {name:{termname}}) return t2.name, count(distinct j) as count order by count desc limit 10; == +---+ == | t2.name | count | == +---+ == | Eurovoc | 9576 | == +---+ == 1 row *== 3668 ms* But what I need is to include the docs on both the term I request and the count on its children, like this. I notice that the combination takes longer than the two separate queries combined. neo4j-sh (?)$ match (j:jurt)-[:HAS_TERM]-()-[:BT*0..]-(t:term)-[:BT*0..1]-(t2:term {name:{termname}}) return t.name, count(distinct j) as count order by count desc limit 10; == +-+ == | t.name | count | == +-+ == | Eurovoc | 9576 | == | gezondheidsbeleid | 1823 | == | overtreding | 1393 | == | Europese organisatie | 1389 | == | EU-instantie | 1323 | == | mondiale organisatie | 1277 | == | gespecialiseerde instelling van de VN | 1143 | == | handeling van de EU | 1129 | == | internationaal publiekrecht | 1091 | == | sociaal beleid| 971 | == +-+ == 10 rows *== 17802 ms* -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.
Re: [Neo4j] 'Using Periodic Commit' throws a invalid syntax exception
oh, that was a really useful command when updating large datasets. +1 for this feature please. On Tuesday, 17 June 2014 14:08:08 UTC+1, Michael Hunger wrote: Sorry, that feature was removed between M06 and 2.1.0 :( So what you have to do is to run this repeatedly: MATCH (a) LIMIT 1 OPTIONAL MATCH (a)-[r]-() DELETE a,r RETURN count(*); until it returns 0. You can try higher limits though, depeding on the number of relationships per node, with 10 rels per node this will be 100k ops, with 100 - 1M ops. Michael Am 17.06.2014 um 15:05 schrieb ducky rohit.a...@gmail.com javascript:: Hi, I am using Neo4j 2.1.2 and when I try to run the example given by Michael Hunger here http://jexp.de/blog/2014/02/cleaning-out-your-graph/: Query: USING PERIODIC COMMIT MATCH (a) OPTIONAL MATCH (a)-[r]-() DELETE a,r; Error: Neo.ClientError.Statement.InvalidSyntax Invalid input 'M': expected whitespace, comment, an integer or LoadCSVQuery (line 2, column 1) MATCH (a) My question is: has the use of this command changed since this blog post? cheers -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+un...@googlegroups.com javascript:. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.
Re: [Neo4j] Upgrading neo4j 1.9.3 to 2.0.3 fails with Invalid log format version found, expected 3 but was 2
Btw. just got the info that it is fixed and will be part of 2.0.4 https://github.com/neo4j/neo4j/commit/37371aa (Thanks Jake!) Michael Am 17.06.2014 um 14:55 schrieb Michael Hunger michael.hun...@neotechnology.com: This is a know issue which is currently worked on, can you delete the logical log files of the lucene index after your upgrade? that means rm graph.db/index/lucene.log.* and you _might_ need to create a transaction against an index, like creating an index and deleting it again, e.g. from java code or the shell. db.index().forNodex(foo).delete() Thanks a lot Michael Am 17.06.2014 um 09:59 schrieb Mamta Thakur ghazal.ma...@gmail.com: Hi, I have been trying to upgrade neo4j from 1.9.3 to 2.0.3. SDN from 2.3.1.RELEASE to 3.1.0.RELEASE. Followed the steps listed @ http://docs.neo4j.org/chunked/stable/deployment-upgrading.html#explicit-upgrade I try bringing up the server with the upgrade configuration.There are a few new folders created in the db store. One among which is upgrade_backup and messages log there says upgrade happened. 2014-06-17 07:16:55.286+ INFO [o.n.k.i.DiagnosticsManager]: --- INITIALIZED diagnostics END --- 2014-06-17 07:17:00.216+ INFO [o.n.k.i.n.s.StoreFactory]: Starting upgrade of database store files 2014-06-17 07:17:00.225+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 10% complete 2014-06-17 07:17:00.228+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 20% complete 2014-06-17 07:17:00.231+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 30% complete 2014-06-17 07:17:00.233+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 40% complete 2014-06-17 07:17:00.236+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 50% complete 2014-06-17 07:17:00.239+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 60% complete 2014-06-17 07:17:00.241+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 70% complete 2014-06-17 07:17:00.244+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 80% complete 2014-06-17 07:17:00.247+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 90% complete 2014-06-17 07:17:00.249+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 100% complete 2014-06-17 07:17:03.776+ INFO [o.n.k.i.n.s.StoreFactory]: Finished upgrade of database store files But I get the error with log/index. Exception when stopping org.neo4j.index.impl.lucene.LuceneDataSource@42a792f0 org.neo4j.kernel.impl.tran saction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but was 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database java.lang.RuntimeException: org.neo4j.kernel.impl.transaction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but wa s 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy$1.reached(LogPruneStrategies.java:250) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$AbstractPruneStrategy.prune(LogPruneStrategies.java:78) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy.prune(LogPruneStrategies.java:222) at org.neo4j.kernel.impl.transaction.xaframework.XaLogicalLog.close(XaLogicalLog.java:742) at org.neo4j.kernel.impl.transaction.xaframework.LogBackedXaDataSource.stop(LogBackedXaDataSource.java:69) at org.neo4j.index.impl.lucene.LuceneDataSource.stop(LuceneDataSource.java:310) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.shutdown(LifeSupport.java:547) at org.neo4j.kernel.lifecycle.LifeSupport.remove(LifeSupport.java:339) at org.neo4j.kernel.impl.transaction.XaDataSourceManager.unregisterDataSource(XaDataSourceManager.java:272) at org.neo4j.index.lucene.LuceneKernelExtension.stop(LuceneKernelExtension.java:92) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport.stop(LifeSupport.java:155) at org.neo4j.kernel.extension.KernelExtensions.stop(KernelExtensions.java:124) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport.stop(LifeSupport.java:155) at org.neo4j.kernel.lifecycle.LifeSupport.shutdown(LifeSupport.java:185) at org.neo4j.kernel.InternalAbstractGraphDatabase.shutdown(InternalAbstractGraphDatabase.java:801) at org.springframework.data.neo4j.support.DelegatingGraphDatabase.shutdown(DelegatingGraphDatabase.java:270)
Re: [Neo4j] LOAD CSV creates nodes but does not set properties
in the file I only have 2 columns, one for client id, which is always not null and CityId, which may be sometimes null. Should I export the records from SQL database leaving out the columns that contain null values? marți, 17 iunie 2014, 15:39:14 UTC+3, Michael Hunger a scris: if they don't have a value for city id, do they then have empty columns there still? like user-id,, You probably want to filter these rows? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c WHERE coalesce(c.CitiId,) ... Am 17.06.2014 um 11:23 schrieb Paul Damian paulda...@gmail.com javascript:: Well, the csv file contains some rows that do not have a value for CityId, and the rows are unique regarding the clientID. There are 11M clients living in 14K Cities. Is there a limit of links/node? Now I've created a piece of code that reads from file and creates each relationship, but, as you can imagine, it works really slow in this scenario. did you create an index on :Client(Id) and :City(Id) what happens if you do: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) each count should be equivalent to the # of rows in the file. Michael Am 16.06.2014 um 17:47 schrieb Paul Damian paulda...@gmail.com: Somehow I've managed to load all the nodes and now I'm trying to load the links as well. I read the nodes from csv file and create the relation between them. I run the following command: USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}), (city: City { Id: toInt(c.CityId)}) CREATE (client)-[r:LOCATED_IN]-(city) Running with a smaller commit size returns this error Neo.DatabaseError.Statement.ExecutionFailure, while increasing the commit size to 1 throws Neo.DatabaseError.General.UnknownFailure. Can you help me with this? joi, 5 iunie 2014, 12:05:18 UTC+3, Michael Hunger a scris: Perhaps something with field or line terminators? I assume it blows up the field separation. Try to run: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c RETURN { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone: c.Phone, Fax: c.Fax, BusinessName: c.BusinessName, URL: c.URL, Latitude: toFloat(c.Latitude), Longitude: toFloat(c.Longitude), AgencyId: toInt(c.AgencyId), RowStatus: toInt(c.RowStatus)} as data, c as line LIMIT 3 On Thu, Jun 5, 2014 at 10:51 AM, Paul Damian paulda...@gmail.com wrote: I've tried using the shell and I get the same results: nodes with no properties. I've created the csv file using MsSQL Server Export. Is it relevant? About you curiosity: I figured I would import first the nodes, then the relationships from the connection tables. Am I doing it wrong? Thanks joi, 5 iunie 2014, 09:54:31 UTC+3, Michael Hunger a scris: I'd probably use a commit size in your case of 50k or 100k. Try to use the neo4j-shell and not the web-interface. Connect to neo4j using bin/neo4j-shell Then run your commands ending with a semicolon. Just curious: Your data is imported as one node per row? That's not really a graph structure. On Wed, Jun 4, 2014 at 6:56 PM, Paul Damian paulda...@gmail.com wrote: Hi there, I'm experimenting with Neo4j while benchmarking a bunch of NoSQL databases for my graduation paper. I'm using the web interface to populate the database. I've been able to load the smaller tables from my SQL database and LOAD CSV works fine. By small, I mean a few columns (4-5) and some rows (1 million). However, when I try to upload a larger table (15 columns, 12 million rows), it creates the nodes but it doesn't set any properties. I've tried to reduce the number of records (to 100) and also the number of columns( just the Id property ), but no luck so far. The cypher command used is this one USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c CREATE (:Client { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone: c.Phone, Fax: c.Fax, BusinessName: c.BusinessName, URL: c.URL, Latitude: toFloat(c.Latitude), Longitude: toFloat(c.Longitude), AgencyId: toInt(c.AgencyId), RowStatus: toInt(c.RowStatus)}) Any help and indication is welcomed, Paul -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+un...@googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You
Re: [Neo4j] LOAD CSV creates nodes but does not set properties
No you can just filter out the lines with no cityid Did you run my suggested commands? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c return c limit 10 Am 17.06.2014 um 16:37 schrieb Paul Damian pauldamia...@gmail.com: in the file I only have 2 columns, one for client id, which is always not null and CityId, which may be sometimes null. Should I export the records from SQL database leaving out the columns that contain null values? marți, 17 iunie 2014, 15:39:14 UTC+3, Michael Hunger a scris: if they don't have a value for city id, do they then have empty columns there still? like user-id,, You probably want to filter these rows? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c WHERE coalesce(c.CitiId,) ... Am 17.06.2014 um 11:23 schrieb Paul Damian paulda...@gmail.com: Well, the csv file contains some rows that do not have a value for CityId, and the rows are unique regarding the clientID. There are 11M clients living in 14K Cities. Is there a limit of links/node? Now I've created a piece of code that reads from file and creates each relationship, but, as you can imagine, it works really slow in this scenario. did you create an index on :Client(Id) and :City(Id) what happens if you do: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) each count should be equivalent to the # of rows in the file. Michael Am 16.06.2014 um 17:47 schrieb Paul Damian paulda...@gmail.com: Somehow I've managed to load all the nodes and now I'm trying to load the links as well. I read the nodes from csv file and create the relation between them. I run the following command: USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}), (city: City { Id: toInt(c.CityId)}) CREATE (client)-[r:LOCATED_IN]-(city) Running with a smaller commit size returns this error Neo.DatabaseError.Statement.ExecutionFailure, while increasing the commit size to 1 throws Neo.DatabaseError.General.UnknownFailure. Can you help me with this? joi, 5 iunie 2014, 12:05:18 UTC+3, Michael Hunger a scris: Perhaps something with field or line terminators? I assume it blows up the field separation. Try to run: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c RETURN { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone: c.Phone, Fax: c.Fax, BusinessName: c.BusinessName, URL: c.URL, Latitude: toFloat(c.Latitude), Longitude: toFloat(c.Longitude), AgencyId: toInt(c.AgencyId), RowStatus: toInt(c.RowStatus)} as data, c as line LIMIT 3 On Thu, Jun 5, 2014 at 10:51 AM, Paul Damian paulda...@gmail.com wrote: I've tried using the shell and I get the same results: nodes with no properties. I've created the csv file using MsSQL Server Export. Is it relevant? About you curiosity: I figured I would import first the nodes, then the relationships from the connection tables. Am I doing it wrong? Thanks joi, 5 iunie 2014, 09:54:31 UTC+3, Michael Hunger a scris: I'd probably use a commit size in your case of 50k or 100k. Try to use the neo4j-shell and not the web-interface. Connect to neo4j using bin/neo4j-shell Then run your commands ending with a semicolon. Just curious: Your data is imported as one node per row? That's not really a graph structure. On Wed, Jun 4, 2014 at 6:56 PM, Paul Damian paulda...@gmail.com wrote: Hi there, I'm experimenting with Neo4j while benchmarking a bunch of NoSQL databases for my graduation paper. I'm using the web interface to populate the database. I've been able to load the smaller tables from my SQL database and LOAD CSV works fine. By small, I mean a few columns (4-5) and some rows (1 million). However, when I try to upload a larger table (15 columns, 12 million rows), it creates the nodes but it doesn't set any properties. I've tried to reduce the number of records (to 100) and also the number of columns( just the Id property ), but no luck so far. The cypher command used is this one USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c CREATE (:Client { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone:
Re: [Neo4j] Upgrading neo4j 1.9.3 to 2.0.3 fails with Invalid log format version found, expected 3 but was 2
Hi Michael, Does that mean I will loose the existing index as well? How would any of my queries work? I fixed some of the @RelationshipEntity that did not have an @GraphId required with 2.0.3. Now I don't get this error on start up rather I get this error when shutting down the server. ~Mamta. On Tuesday, June 17, 2014 7:36:10 PM UTC+5:30, Michael Hunger wrote: Btw. just got the info that it is fixed and will be part of 2.0.4 https://github.com/neo4j/neo4j/commit/37371aa (Thanks Jake!) Michael Am 17.06.2014 um 14:55 schrieb Michael Hunger michael...@neotechnology.com javascript:: This is a know issue which is currently worked on, can you delete the logical log files of the lucene index after your upgrade? that means rm graph.db/index/lucene.log.* and you _might_ need to create a transaction against an index, like creating an index and deleting it again, e.g. from java code or the shell. db.index().forNodex(foo).delete() Thanks a lot Michael Am 17.06.2014 um 09:59 schrieb Mamta Thakur ghazal...@gmail.com javascript:: Hi, I have been trying to upgrade neo4j from 1.9.3 to 2.0.3. SDN from 2.3.1.RELEASE to 3.1.0.RELEASE. Followed the steps listed @ http://docs.neo4j.org/chunked/stable/deployment-upgrading.html#explicit-upgrade I try bringing up the server with the upgrade configuration.There are a few new folders created in the db store. One among which is upgrade_backup and messages log there says upgrade happened. 2014-06-17 07:16:55.286+ INFO [o.n.k.i.DiagnosticsManager]: --- INITIALIZED diagnostics END --- 2014-06-17 07:17:00.216+ INFO [o.n.k.i.n.s.StoreFactory]: Starting upgrade of database store files 2014-06-17 07:17:00.225+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 10% complete 2014-06-17 07:17:00.228+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 20% complete 2014-06-17 07:17:00.231+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 30% complete 2014-06-17 07:17:00.233+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 40% complete 2014-06-17 07:17:00.236+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 50% complete 2014-06-17 07:17:00.239+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 60% complete 2014-06-17 07:17:00.241+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 70% complete 2014-06-17 07:17:00.244+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 80% complete 2014-06-17 07:17:00.247+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 90% complete 2014-06-17 07:17:00.249+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 100% complete 2014-06-17 07:17:03.776+ INFO [o.n.k.i.n.s.StoreFactory]: Finished upgrade of database store files But I get the error with log/index. Exception when stopping org.neo4j.index.impl.lucene.LuceneDataSource@42a792f0 org.neo4j.kernel.impl.tran saction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but was 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database java.lang.RuntimeException: org.neo4j.kernel.impl.transaction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but wa s 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy$1.reached(LogPruneStrategies.java:250) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$AbstractPruneStrategy.prune(LogPruneStrategies.java:78) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy.prune(LogPruneStrategies.java:222) at org.neo4j.kernel.impl.transaction.xaframework.XaLogicalLog.close(XaLogicalLog.java:742) at org.neo4j.kernel.impl.transaction.xaframework.LogBackedXaDataSource.stop(LogBackedXaDataSource.java:69) at org.neo4j.index.impl.lucene.LuceneDataSource.stop(LuceneDataSource.java:310) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.shutdown(LifeSupport.java:547) at org.neo4j.kernel.lifecycle.LifeSupport.remove(LifeSupport.java:339) at org.neo4j.kernel.impl.transaction.XaDataSourceManager.unregisterDataSource(XaDataSourceManager.java:272) at org.neo4j.index.lucene.LuceneKernelExtension.stop(LuceneKernelExtension.java:92) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport.stop(LifeSupport.java:155) at org.neo4j.kernel.extension.KernelExtensions.stop(KernelExtensions.java:124) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527)
Re: [Neo4j] Upgrading neo4j 1.9.3 to 2.0.3 fails with Invalid log format version found, expected 3 but was 2
No, as you have a clean shutdown all the data is in the store. Sent from mobile device Am 17.06.2014 um 16:55 schrieb Mamta Thakur ghazal.ma...@gmail.com: Hi Michael, Does that mean I will loose the existing index as well? How would any of my queries work? I fixed some of the @RelationshipEntity that did not have an @GraphId required with 2.0.3. Now I don't get this error on start up rather I get this error when shutting down the server. ~Mamta. On Tuesday, June 17, 2014 7:36:10 PM UTC+5:30, Michael Hunger wrote: Btw. just got the info that it is fixed and will be part of 2.0.4 https://github.com/neo4j/neo4j/commit/37371aa (Thanks Jake!) Michael Am 17.06.2014 um 14:55 schrieb Michael Hunger michael...@neotechnology.com: This is a know issue which is currently worked on, can you delete the logical log files of the lucene index after your upgrade? that means rm graph.db/index/lucene.log.* and you _might_ need to create a transaction against an index, like creating an index and deleting it again, e.g. from java code or the shell. db.index().forNodex(foo).delete() Thanks a lot Michael Am 17.06.2014 um 09:59 schrieb Mamta Thakur ghazal...@gmail.com: Hi, I have been trying to upgrade neo4j from 1.9.3 to 2.0.3. SDN from 2.3.1.RELEASE to 3.1.0.RELEASE. Followed the steps listed @ http://docs.neo4j.org/chunked/stable/deployment-upgrading.html#explicit-upgrade I try bringing up the server with the upgrade configuration.There are a few new folders created in the db store. One among which is upgrade_backup and messages log there says upgrade happened. 2014-06-17 07:16:55.286+ INFO [o.n.k.i.DiagnosticsManager]: --- INITIALIZED diagnostics END --- 2014-06-17 07:17:00.216+ INFO [o.n.k.i.n.s.StoreFactory]: Starting upgrade of database store files 2014-06-17 07:17:00.225+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 10% complete 2014-06-17 07:17:00.228+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 20% complete 2014-06-17 07:17:00.231+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 30% complete 2014-06-17 07:17:00.233+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 40% complete 2014-06-17 07:17:00.236+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 50% complete 2014-06-17 07:17:00.239+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 60% complete 2014-06-17 07:17:00.241+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 70% complete 2014-06-17 07:17:00.244+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 80% complete 2014-06-17 07:17:00.247+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 90% complete 2014-06-17 07:17:00.249+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 100% complete 2014-06-17 07:17:03.776+ INFO [o.n.k.i.n.s.StoreFactory]: Finished upgrade of database store files But I get the error with log/index. Exception when stopping org.neo4j.index.impl.lucene.LuceneDataSource@42a792f0 org.neo4j.kernel.impl.tran saction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but was 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database java.lang.RuntimeException: org.neo4j.kernel.impl.transaction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but wa s 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy$1.reached(LogPruneStrategies.java:250) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$AbstractPruneStrategy.prune(LogPruneStrategies.java:78) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy.prune(LogPruneStrategies.java:222) at org.neo4j.kernel.impl.transaction.xaframework.XaLogicalLog.close(XaLogicalLog.java:742) at org.neo4j.kernel.impl.transaction.xaframework.LogBackedXaDataSource.stop(LogBackedXaDataSource.java:69) at org.neo4j.index.impl.lucene.LuceneDataSource.stop(LuceneDataSource.java:310) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.shutdown(LifeSupport.java:547) at org.neo4j.kernel.lifecycle.LifeSupport.remove(LifeSupport.java:339) at org.neo4j.kernel.impl.transaction.XaDataSourceManager.unregisterDataSource(XaDataSourceManager.java:272) at org.neo4j.index.lucene.LuceneKernelExtension.stop(LuceneKernelExtension.java:92) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport.stop(LifeSupport.java:155) at
Re: [Neo4j] LOAD CSV creates nodes but does not set properties
The first query returns 96 which is the number of rows in the file and the second one returns Neo.DatabaseError.Statement.ExecutionFailure probably because of the null values. But then I run the following command: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city:City { Id: toInt(c.CityId)}) WHERE coalesce(c.CityId,) RETURN count(*) and I get 992980 marți, 17 iunie 2014, 17:55:56 UTC+3, Michael Hunger a scris: No you can just filter out the lines with no cityid Did you run my suggested commands? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c return c limit 10 Am 17.06.2014 um 16:37 schrieb Paul Damian paulda...@gmail.com javascript:: in the file I only have 2 columns, one for client id, which is always not null and CityId, which may be sometimes null. Should I export the records from SQL database leaving out the columns that contain null values? marți, 17 iunie 2014, 15:39:14 UTC+3, Michael Hunger a scris: if they don't have a value for city id, do they then have empty columns there still? like user-id,, You probably want to filter these rows? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c WHERE coalesce(c.CitiId,) ... Am 17.06.2014 um 11:23 schrieb Paul Damian paulda...@gmail.com: Well, the csv file contains some rows that do not have a value for CityId, and the rows are unique regarding the clientID. There are 11M clients living in 14K Cities. Is there a limit of links/node? Now I've created a piece of code that reads from file and creates each relationship, but, as you can imagine, it works really slow in this scenario. did you create an index on :Client(Id) and :City(Id) what happens if you do: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) each count should be equivalent to the # of rows in the file. Michael Am 16.06.2014 um 17:47 schrieb Paul Damian paulda...@gmail.com: Somehow I've managed to load all the nodes and now I'm trying to load the links as well. I read the nodes from csv file and create the relation between them. I run the following command: USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}), (city: City { Id: toInt(c.CityId)}) CREATE (client)-[r:LOCATED_IN]-(city) Running with a smaller commit size returns this error Neo.DatabaseError.Statement.ExecutionFailure, while increasing the commit size to 1 throws Neo.DatabaseError.General.UnknownFailure. Can you help me with this? joi, 5 iunie 2014, 12:05:18 UTC+3, Michael Hunger a scris: Perhaps something with field or line terminators? I assume it blows up the field separation. Try to run: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c RETURN { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone: c.Phone, Fax: c.Fax, BusinessName: c.BusinessName, URL: c.URL, Latitude: toFloat(c.Latitude), Longitude: toFloat(c.Longitude), AgencyId: toInt(c.AgencyId), RowStatus: toInt(c.RowStatus)} as data, c as line LIMIT 3 On Thu, Jun 5, 2014 at 10:51 AM, Paul Damian paulda...@gmail.com wrote: I've tried using the shell and I get the same results: nodes with no properties. I've created the csv file using MsSQL Server Export. Is it relevant? About you curiosity: I figured I would import first the nodes, then the relationships from the connection tables. Am I doing it wrong? Thanks joi, 5 iunie 2014, 09:54:31 UTC+3, Michael Hunger a scris: I'd probably use a commit size in your case of 50k or 100k. Try to use the neo4j-shell and not the web-interface. Connect to neo4j using bin/neo4j-shell Then run your commands ending with a semicolon. Just curious: Your data is imported as one node per row? That's not really a graph structure. On Wed, Jun 4, 2014 at 6:56 PM, Paul Damian paulda...@gmail.com wrote: Hi there, I'm experimenting with Neo4j while benchmarking a bunch of NoSQL databases for my graduation paper. I'm using the web interface to populate the database. I've been able to load the smaller tables from my SQL database and LOAD CSV works fine. By small, I mean a few columns (4-5) and some rows (1 million). However, when I try to upload a larger table (15 columns, 12 million rows), it
[Neo4j] Large scale network analysis - best strategy?
I am preparing a Neo4j database on which I would like to do some network analysis. It is a representation of a weakly connected and static physical system, and will have in the region of 50 million nodes where, lets say, about 50 nodes will connect to a parent node, which in turn is linked (think streets and intersections) to a network of other parent nodes. For most of the analysis, I will be using a weighted distance decay, so analysis of things like betweenness or centrality will be computed for the parent node network, but only to a limited extent. So, for example, if (a)--(b)--(c)--(d)--(e), then the computation will only be based up to, say, two steps away. So (a) will consider (b) and (c), whereas (c) will consider two steps in either direction. My question is a conceptual and strategic one: What is the best approach for doing this kind of analysis with neo4j? I currently work with Python, but it appears that for speed, flexibility, and use of complex graph algorithms, I am better off working with the embedded Java API for direct and powerful access to the graph? Or is an approach using something like bulb flow with gremlin also feasible? How does the power and flexibility of the different embedded tools compare - e.g. Python embedded vs. Java vs. Node.js? Thanks. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.
Re: [Neo4j] LOAD CSV creates nodes but does not set properties
Ok, cool and you have the indexes for both :City(Id) and :Client(Id) ? Michael Am 17.06.2014 um 18:15 schrieb Paul Damian pauldamia...@gmail.com: The first query returns 96 which is the number of rows in the file and the second one returns Neo.DatabaseError.Statement.ExecutionFailure probably because of the null values. But then I run the following command: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city:City { Id: toInt(c.CityId)}) WHERE coalesce(c.CityId,) RETURN count(*) and I get 992980 marți, 17 iunie 2014, 17:55:56 UTC+3, Michael Hunger a scris: No you can just filter out the lines with no cityid Did you run my suggested commands? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c return c limit 10 Am 17.06.2014 um 16:37 schrieb Paul Damian paulda...@gmail.com: in the file I only have 2 columns, one for client id, which is always not null and CityId, which may be sometimes null. Should I export the records from SQL database leaving out the columns that contain null values? marți, 17 iunie 2014, 15:39:14 UTC+3, Michael Hunger a scris: if they don't have a value for city id, do they then have empty columns there still? like user-id,, You probably want to filter these rows? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c WHERE coalesce(c.CitiId,) ... Am 17.06.2014 um 11:23 schrieb Paul Damian paulda...@gmail.com: Well, the csv file contains some rows that do not have a value for CityId, and the rows are unique regarding the clientID. There are 11M clients living in 14K Cities. Is there a limit of links/node? Now I've created a piece of code that reads from file and creates each relationship, but, as you can imagine, it works really slow in this scenario. did you create an index on :Client(Id) and :City(Id) what happens if you do: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) each count should be equivalent to the # of rows in the file. Michael Am 16.06.2014 um 17:47 schrieb Paul Damian paulda...@gmail.com: Somehow I've managed to load all the nodes and now I'm trying to load the links as well. I read the nodes from csv file and create the relation between them. I run the following command: USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}), (city: City { Id: toInt(c.CityId)}) CREATE (client)-[r:LOCATED_IN]-(city) Running with a smaller commit size returns this error Neo.DatabaseError.Statement.ExecutionFailure, while increasing the commit size to 1 throws Neo.DatabaseError.General.UnknownFailure. Can you help me with this? joi, 5 iunie 2014, 12:05:18 UTC+3, Michael Hunger a scris: Perhaps something with field or line terminators? I assume it blows up the field separation. Try to run: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c RETURN { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone: c.Phone, Fax: c.Fax, BusinessName: c.BusinessName, URL: c.URL, Latitude: toFloat(c.Latitude), Longitude: toFloat(c.Longitude), AgencyId: toInt(c.AgencyId), RowStatus: toInt(c.RowStatus)} as data, c as line LIMIT 3 On Thu, Jun 5, 2014 at 10:51 AM, Paul Damian paulda...@gmail.com wrote: I've tried using the shell and I get the same results: nodes with no properties. I've created the csv file using MsSQL Server Export. Is it relevant? About you curiosity: I figured I would import first the nodes, then the relationships from the connection tables. Am I doing it wrong? Thanks joi, 5 iunie 2014, 09:54:31 UTC+3, Michael Hunger a scris: I'd probably use a commit size in your case of 50k or 100k. Try to use the neo4j-shell and not the web-interface. Connect to neo4j using bin/neo4j-shell Then run your commands ending with a semicolon. Just curious: Your data is imported as one node per row? That's not really a graph structure. On Wed, Jun 4, 2014 at 6:56 PM, Paul Damian paulda...@gmail.com wrote: Hi there, I'm experimenting with Neo4j while benchmarking a bunch of NoSQL databases for my graduation paper. I'm using the web interface to populate the database. I've been able to load the smaller tables from my SQL database
Re: [Neo4j] LOAD CSV creates nodes but does not set properties
Yes, I do. I keep getting Java heap space error now. I'm using 100 commit size. marți, 17 iunie 2014, 19:28:05 UTC+3, Michael Hunger a scris: Ok, cool and you have the indexes for both :City(Id) and :Client(Id) ? Michael Am 17.06.2014 um 18:15 schrieb Paul Damian paulda...@gmail.com javascript:: The first query returns 96 which is the number of rows in the file and the second one returns Neo.DatabaseError.Statement.ExecutionFailure probably because of the null values. But then I run the following command: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city:City { Id: toInt(c.CityId)}) WHERE coalesce(c.CityId,) RETURN count(*) and I get 992980 marți, 17 iunie 2014, 17:55:56 UTC+3, Michael Hunger a scris: No you can just filter out the lines with no cityid Did you run my suggested commands? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c return c limit 10 Am 17.06.2014 um 16:37 schrieb Paul Damian paulda...@gmail.com: in the file I only have 2 columns, one for client id, which is always not null and CityId, which may be sometimes null. Should I export the records from SQL database leaving out the columns that contain null values? marți, 17 iunie 2014, 15:39:14 UTC+3, Michael Hunger a scris: if they don't have a value for city id, do they then have empty columns there still? like user-id,, You probably want to filter these rows? LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c WHERE coalesce(c.CitiId,) ... Am 17.06.2014 um 11:23 schrieb Paul Damian paulda...@gmail.com: Well, the csv file contains some rows that do not have a value for CityId, and the rows are unique regarding the clientID. There are 11M clients living in 14K Cities. Is there a limit of links/node? Now I've created a piece of code that reads from file and creates each relationship, but, as you can imagine, it works really slow in this scenario. did you create an index on :Client(Id) and :City(Id) what happens if you do: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}) RETURN count(*) LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (city: City { Id: toInt(c.CityId)}) RETURN count(*) each count should be equivalent to the # of rows in the file. Michael Am 16.06.2014 um 17:47 schrieb Paul Damian paulda...@gmail.com: Somehow I've managed to load all the nodes and now I'm trying to load the links as well. I read the nodes from csv file and create the relation between them. I run the following command: USING PERIODIC COMMIT 100 LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/LOCATED_IN.csv AS c MATCH (client: Client { Id: toInt(c.Id)}), (city: City { Id: toInt(c.CityId)}) CREATE (client)-[r:LOCATED_IN]-(city) Running with a smaller commit size returns this error Neo.DatabaseError.Statement.ExecutionFailure, while increasing the commit size to 1 throws Neo.DatabaseError.General.UnknownFailure. Can you help me with this? joi, 5 iunie 2014, 12:05:18 UTC+3, Michael Hunger a scris: Perhaps something with field or line terminators? I assume it blows up the field separation. Try to run: LOAD CSV WITH HEADERS FROM file:/Users/pauld/Documents/Client.csv AS c RETURN { Id: toInt(c.Id), FirstName: c.FirstName, LastName: c.Lastname, Address: c.Address, ZipCode: toInt(c.ZipCode), Email: c.Email, Phone: c.Phone, Fax: c.Fax, BusinessName: c.BusinessName, URL: c.URL, Latitude: toFloat(c.Latitude), Longitude: toFloat(c.Longitude), AgencyId: toInt(c.AgencyId), RowStatus: toInt(c.RowStatus)} as data, c as line LIMIT 3 On Thu, Jun 5, 2014 at 10:51 AM, Paul Damian paulda...@gmail.com wrote: I've tried using the shell and I get the same results: nodes with no properties. I've created the csv file using MsSQL Server Export. Is it relevant? About you curiosity: I figured I would import first the nodes, then the relationships from the connection tables. Am I doing it wrong? Thanks joi, 5 iunie 2014, 09:54:31 UTC+3, Michael Hunger a scris: I'd probably use a commit size in your case of 50k or 100k. Try to use the neo4j-shell and not the web-interface. Connect to neo4j using bin/neo4j-shell Then run your commands ending with a semicolon. Just curious: Your data is imported as one node per row? That's not really a graph structure. On Wed, Jun 4, 2014 at 6:56 PM, Paul Damian paulda...@gmail.com wrote: Hi there, I'm experimenting with Neo4j while benchmarking a bunch of NoSQL databases for my
[Neo4j] Re: Neo4j database ALWAYS shuts down incorrectly if start/stop as a service from a list of windows services
Hi Denys, I think you're experience is a variation of mine as related here https://groups.google.com/d/msg/neo4j/Qp2azbKy2_8/u4jDbiBLlesJ. Seems to be a long-standing issue that hasn't bitten too many but is problematic for the 2.1 database migration. There are some tips/insight on the link. The Neo folks are aware of the issue so I expect we'll see a fix at some point, maybe soon. Have you tried the Ole Out and In -- dump it from your 'corrupt' DB and load it in an empty 2.1? I had a number of tiny to small research and self-learning DBs that it was so much easier to go out and back in on a fresh 2.1 store. That tip about timeouts might help you, too. And the Neo4j Mojo that I am not qualified to comment on and I suspect is a big factor has to do with leveraging 2.0+ version indexing and constraints, etc. There might be some tweaks to the schema that you can make before migrating that, along with the increased timeout value, might give the migration process the room to work its one-time-only procedure. It's a painful problem but, as you can imagine, hits a relative small segment of the broader Neo4j community. BTW, if you run Neo4j as a Windows Service, have you tried my mini Control Panel? :-) http://jim-salmons.github.io/neo4jcp/ I've subscribed to this thread and will let you know if I learn anything more, etc. --Jim-- www.FactMiners.org and www.SoftalkApple.com On Tuesday, June 17, 2014 4:50:09 AM UTC-5, Denys Hryvastov wrote: Here is a stack trace that I get when I try to do upgrade from 1.9.5 to 2.0: 2014-06-17 09:48:27.319+ INFO [API] Setting startup timeout to: 12ms based on -1 Detected incorrectly shut down database, performing recovery.. 2014-06-17 09:48:28.108+ DEBUG [API] org.neo4j.server.ServerStartupException: Starting Neo4j Server failed: Error starting org.neo4j.kernel.EmbeddedGraphDatabase, D:\Neo4j\neo4j-enterprise-2.0.0\data\graph.db at org.neo4j.server.AbstractNeoServer.start(AbstractNeoServer.java:209) ~[neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.Bootstrapper.start(Bootstrapper.java:87) [neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.Bootstrapper.main(Bootstrapper.java:50) [neo4j-server-2.0.0.jar:2.0.0] Caused by: java.lang.RuntimeException: Error starting org.neo4j.kernel.EmbeddedGraphDatabase, D:\Neo4j\neo4j-enterprise-2.0.0\data\graph.db at org.neo4j.kernel.InternalAbstractGraphDatabase.run(InternalAbstractGraphDatabase.java:333) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.EmbeddedGraphDatabase.init(EmbeddedGraphDatabase.java:63) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.graphdb.factory.GraphDatabaseFactory$1.newDatabase(GraphDatabaseFactory.java:92) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.graphdb.factory.GraphDatabaseBuilder.newGraphDatabase(GraphDatabaseBuilder.java:198) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.impl.recovery.StoreRecoverer.recover(StoreRecoverer.java:115) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.server.preflight.PerformRecoveryIfNecessary.run(PerformRecoveryIfNecessary.java:59) ~[neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.preflight.PreFlightTasks.run(PreFlightTasks.java:70) ~[neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.AbstractNeoServer.runPreflightTasks(AbstractNeoServer.java:319) ~[neo4j-server-2.0.0.jar:2.0.0] at org.neo4j.server.AbstractNeoServer.start(AbstractNeoServer.java:144) ~[neo4j-server-2.0.0.jar:2.0.0] ... 2 common frames omitted Caused by: org.neo4j.kernel.lifecycle.LifecycleException: Component 'org.neo4j.kernel.impl.transaction.XaDataSourceManager@2b1eb67d' was successfully initialized, but failed to start. Please see attached cause exception. at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.start(LifeSupport.java:504) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.lifecycle.LifeSupport.start(LifeSupport.java:115) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.InternalAbstractGraphDatabase.run(InternalAbstractGraphDatabase.java:310) ~[neo4j-kernel-2.0.0.jar:2.0.0] ... 10 common frames omitted Caused by: org.neo4j.kernel.lifecycle.LifecycleException: Component 'org.neo4j.kernel.impl.nioneo.xa.NeoStoreXaDataSource@1bf5df6a' was successfully initialized, but failed to start. P lease see attached cause exception. at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.start(LifeSupport.java:504) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.lifecycle.LifeSupport.start(LifeSupport.java:115) ~[neo4j-kernel-2.0.0.jar:2.0.0] at org.neo4j.kernel.impl.transaction.XaDataSourceManager.start(XaDataSourceManager.java:164) ~[neo4j-kernel-2.0.0.jar:2.0.0] at
Re: [Neo4j] Large scale network analysis - best strategy?
Hi Gareth As you identify, there are certainly some differences in terms of performance and feature set that you get when working with Neo4j under different programming languages. Depending on your background, constraints and integration needs, you could consider a hybrid approach whereby you continue working with Python for your main application and build anything that requires serious performance as a server extension in Java. Neo4j plugin support is pretty comprehensive: for example, my server extension load2neo http://nigelsmall.com/load2neo provides a facility to bulk load data but also has direct support from my Python driver, py2neo http://py2neo.org/. This approach is somewhat analogous to compiling a C extension in Python and could be done as an optimisation step once you have built your end-to-end application logic. Bear in mind also that Cypher is very powerful these days. It would certainly be worth exploring some of its more recent capabilities before choosing an architectural path as you may find there is little that cannot already be achieved purely with Cypher. If this is the case, your choice of application language could then become far less critical. I'd suggest beginning with a prototype in a language you are comfortable with. Then, build a suite of queries you need to run and ascertain the bottlenecks or missing features. Once you have a list of these, you can then make an informed decision on which pieces to optimise. Kind regards Nigel On 17 June 2014 15:42, Shongololo garethsim...@gmail.com wrote: I am preparing a Neo4j database on which I would like to do some network analysis. It is a representation of a weakly connected and static physical system, and will have in the region of 50 million nodes where, lets say, about 50 nodes will connect to a parent node, which in turn is linked (think streets and intersections) to a network of other parent nodes. For most of the analysis, I will be using a weighted distance decay, so analysis of things like betweenness or centrality will be computed for the parent node network, but only to a limited extent. So, for example, if (a)--(b)--(c)--(d)--(e), then the computation will only be based up to, say, two steps away. So (a) will consider (b) and (c), whereas (c) will consider two steps in either direction. My question is a conceptual and strategic one: What is the best approach for doing this kind of analysis with neo4j? I currently work with Python, but it appears that for speed, flexibility, and use of complex graph algorithms, I am better off working with the embedded Java API for direct and powerful access to the graph? Or is an approach using something like bulb flow with gremlin also feasible? How does the power and flexibility of the different embedded tools compare - e.g. Python embedded vs. Java vs. Node.js? Thanks. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups Neo4j group. To unsubscribe from this group and stop receiving emails from it, send an email to neo4j+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.
[Neo4j] Re: Upgrading neo4j 1.9.3 to 2.0.3 fails with Invalid log format version found, expected 3 but was 2
I also encountered this problem.My solution is that: 1.put the data import single neo4j.(one instance) 2.update some data(any one can) 3.start the ha cluster. Note: First startup the neo4j instance that import the data 在 2014年6月17日星期二UTC+8下午3时59分32秒,Mamta Thakur写道: Hi, I have been trying to upgrade neo4j from 1.9.3 to 2.0.3. SDN from 2.3.1.RELEASE to 3.1.0.RELEASE. Followed the steps listed @ http://docs.neo4j.org/chunked/stable/deployment-upgrading.html#explicit-upgrade I try bringing up the server with the upgrade configuration.There are a few new folders created in the db store. One among which is upgrade_backup and messages log there says upgrade happened. 2014-06-17 07:16:55.286+ INFO [o.n.k.i.DiagnosticsManager]: --- INITIALIZED diagnostics END --- 2014-06-17 07:17:00.216+ INFO [o.n.k.i.n.s.StoreFactory]: Starting upgrade of database store files 2014-06-17 07:17:00.225+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 10% complete 2014-06-17 07:17:00.228+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 20% complete 2014-06-17 07:17:00.231+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 30% complete 2014-06-17 07:17:00.233+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 40% complete 2014-06-17 07:17:00.236+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 50% complete 2014-06-17 07:17:00.239+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 60% complete 2014-06-17 07:17:00.241+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 70% complete 2014-06-17 07:17:00.244+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 80% complete 2014-06-17 07:17:00.247+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 90% complete 2014-06-17 07:17:00.249+ INFO [o.n.k.i.n.s.StoreFactory]: Store upgrade 100% complete 2014-06-17 07:17:03.776+ INFO [o.n.k.i.n.s.StoreFactory]: Finished upgrade of database store files But I get the error with log/index. Exception when stopping org.neo4j.index.impl.lucene.LuceneDataSource@42a792f0 org.neo4j.kernel.impl.tran saction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but was 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database java.lang.RuntimeException: org.neo4j.kernel.impl.transaction.xaframework.IllegalLogFormatException: Invalid log format version found, expected 3 but wa s 2. To be able to upgrade from an older log format version there must have been a clean shutdown of the database at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy$1.reached(LogPruneStrategies.java:250) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$AbstractPruneStrategy.prune(LogPruneStrategies.java:78) at org.neo4j.kernel.impl.transaction.xaframework.LogPruneStrategies$TransactionTimeSpanPruneStrategy.prune(LogPruneStrategies.java:222) at org.neo4j.kernel.impl.transaction.xaframework.XaLogicalLog.close(XaLogicalLog.java:742) at org.neo4j.kernel.impl.transaction.xaframework.LogBackedXaDataSource.stop(LogBackedXaDataSource.java:69) at org.neo4j.index.impl.lucene.LuceneDataSource.stop(LuceneDataSource.java:310) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.shutdown(LifeSupport.java:547) at org.neo4j.kernel.lifecycle.LifeSupport.remove(LifeSupport.java:339) at org.neo4j.kernel.impl.transaction.XaDataSourceManager.unregisterDataSource(XaDataSourceManager.java:272) at org.neo4j.index.lucene.LuceneKernelExtension.stop(LuceneKernelExtension.java:92) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport.stop(LifeSupport.java:155) at org.neo4j.kernel.extension.KernelExtensions.stop(KernelExtensions.java:124) at org.neo4j.kernel.lifecycle.LifeSupport$LifecycleInstance.stop(LifeSupport.java:527) at org.neo4j.kernel.lifecycle.LifeSupport.stop(LifeSupport.java:155) at org.neo4j.kernel.lifecycle.LifeSupport.shutdown(LifeSupport.java:185) at org.neo4j.kernel.InternalAbstractGraphDatabase.shutdown(InternalAbstractGraphDatabase.java:801) at org.springframework.data.neo4j.support.DelegatingGraphDatabase.shutdown(DelegatingGraphDatabase.java:270) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at