Re: Java sample for using cassandra-driver-spark

2014-07-07 Thread Piotr Kołaczkowski
Hi, we're planning to add a basic Java-API very soon, possibly this week. There's a ticket for it here: https://github.com/datastax/cassandra-driver-spark/issues/11 We're open to any ideas. Just let us know what you need the API to have in the comments. Regards, Piotr Kołaczkowski

Re: How to terminate job from the task code?

2014-07-02 Thread Piotr Kołaczkowski
SparkContext is not serializable and can't be just "sent across" ;) 2014-06-21 14:14 GMT+02:00 Mayur Rustagi : > You can terminate job group from spark context, Youll have to send across > the spark context to your task. > On 21 Jun 2014 01:09, "Piotr Kołaczkow

How to terminate job from the task code?

2014-06-20 Thread Piotr Kołaczkowski
If the task detects unrecoverable error, i.e. an error that we can't expect to fix by retrying nor moving the task to another node, how to stop the job / prevent Spark from retrying it? def process(taskContext: TaskContext, data: Iterator[T]) { ... if (unrecoverableError) { ??? // ter

Re: Master not seeing recovered nodes("Got heartbeat from unregistered worker ....")

2014-06-16 Thread Piotr Kołaczkowski
We are having the same problem. We're running Spark 0.9.1 in standalone mode and on some heavy jobs workers become unresponsive and marked by master as dead, even though the worker process is still running. Then they never join the cluster again and cluster becomes essentially unusable until we res

Re: SparkContext#stop

2014-05-22 Thread Piotr Kołaczkowski
No exceptions in any logs. No errors in stdout or stderr. 2014-05-22 11:21 GMT+02:00 Andrew Or : > You should always call sc.stop(), so it cleans up state and does not fill > up your disk over time. The strange behavior you observe is mostly benign, > as it only occurs after you have supposedly

Workers disconnected from master sometimes and never reconnect back

2014-05-22 Thread Piotr Kołaczkowski
Hi, Another problem we observed that on a very heavily loaded cluster, if the worker fails to respond to the heartbeat within 60 seconds, it gets disconnected permanently from the master and never connects back again. It is very easy to reproduce - just setup a spark standalone cluster on a single

SparkContext#stop

2014-05-22 Thread Piotr Kołaczkowski
Hi, We observed strange behabiour of Spark 0.9.0 when using sc.stop(). We have a bunch of applications that perform some jobs and then issue sc.stop() at the end of main. Most of the time, everything works as desired, but sometimes the applications get marked as "FAILED" by the master and all rem