Re: IgniteInterruptedException: Node is stopping

2017-11-27 Thread Denis Mekhanikov
Hi Hyma! Looks like you encountered a classic deadlock. It happens because you put values into cache in arbitrary order. This line causes this problem: *companyDao.nameCache.putAll(kvs)* So, when multiple threads try to acquire the same locks in different order, then these operations will be wait

Re: IgniteInterruptedException: Node is stopping

2017-11-23 Thread Hyma
Below is the corresponding code where ignite step was in hung state. logInfo("Populating the canonical name Cache on Ignite Nodes") val time = System.currentTimeMillis() companyVORDD.mapPartitions(x => { val kvs = x.map(comp => (comp.wcaId,comp)).toMap[String,CompanyVO].asJava

Re: IgniteInterruptedException: Node is stopping

2017-11-22 Thread Michael Cherkasov
Hi Hyma, Could you please show a code snippet where it is hanged? Thanks, Mike. 2017-11-22 12:48 GMT+03:00 Hyma : > Thanks Mikhail. > > I suspected to increase the spark heartbeat/network timeout. But my > question > here is if an executor is lost, corresponding ignite node also gets out of > c

Re: IgniteInterruptedException: Node is stopping

2017-11-22 Thread Hyma
Thanks Mikhail. I suspected to increase the spark heartbeat/network timeout. But my question here is if an executor is lost, corresponding ignite node also gets out of cluster. In that case, ignite takes care of re balancing between the other active nodes right. My spark job was not killed and it

Re: IgniteInterruptedException: Node is stopping

2017-11-17 Thread Mikhail
Hi Hyma, looks like your job takes too much time, you hit some timeout and spark killed your jobs. I don't see any other errors or warnings from your logs, it's very likely that you need to increase some time out in spark. thanks, Mike. -- Sent from: http://apache-ignite-users.70518.x6.nabble.

IgniteInterruptedException: Node is stopping

2017-11-17 Thread Hyma
Hi, When loading ignite cache, we saw the spark job went into hung state at this step. We see one of the executor task has been running for hours and below are the logs from this executor that had the failure. Stdout log Launch class org.apache.spark.executor.CoarseGrainedExecutorBackend by calli