Re: Java sample for using cassandra-driver-spark

2014-07-07 Thread Piotr Kołaczkowski
Hi, we're planning to add a basic Java-API very soon, possibly this week. There's a ticket for it here: https://github.com/datastax/cassandra-driver-spark/issues/11 We're open to any ideas. Just let us know what you need the API to have in the comments. Regards, Piotr Kołaczkowski 2014-07-05 0

Re: How to terminate job from the task code?

2014-07-02 Thread Piotr Kołaczkowski
SparkContext is not serializable and can't be just sent across ;) 2014-06-21 14:14 GMT+02:00 Mayur Rustagi mayur.rust...@gmail.com: You can terminate job group from spark context, Youll have to send across the spark context to your task. On 21 Jun 2014 01:09, Piotr Kołaczkowski pkola

Re: Master not seeing recovered nodes(Got heartbeat from unregistered worker ....)

2014-06-16 Thread Piotr Kołaczkowski
We are having the same problem. We're running Spark 0.9.1 in standalone mode and on some heavy jobs workers become unresponsive and marked by master as dead, even though the worker process is still running. Then they never join the cluster again and cluster becomes essentially unusable until we

SparkContext#stop

2014-05-22 Thread Piotr Kołaczkowski
Hi, We observed strange behabiour of Spark 0.9.0 when using sc.stop(). We have a bunch of applications that perform some jobs and then issue sc.stop() at the end of main. Most of the time, everything works as desired, but sometimes the applications get marked as FAILED by the master and all

Workers disconnected from master sometimes and never reconnect back

2014-05-22 Thread Piotr Kołaczkowski
Hi, Another problem we observed that on a very heavily loaded cluster, if the worker fails to respond to the heartbeat within 60 seconds, it gets disconnected permanently from the master and never connects back again. It is very easy to reproduce - just setup a spark standalone cluster on a

Re: SparkContext#stop

2014-05-22 Thread Piotr Kołaczkowski
No exceptions in any logs. No errors in stdout or stderr. 2014-05-22 11:21 GMT+02:00 Andrew Or and...@databricks.com: You should always call sc.stop(), so it cleans up state and does not fill up your disk over time. The strange behavior you observe is mostly benign, as it only occurs after