Re: Contributing to Spark

2017-03-20 Thread cht liu
Hi Sam A great way to contribute to Spark is to help answer user questions on the user@spark.apache.org mailing list or on StackOverflow. 2017-03-20 11:50 GMT+08:00 Nick Pentreath : > If you have experience and interest in Python then PySpark is a good area > to look

Re: Huge partitioning job takes longer to close after all tasks finished

2017-03-07 Thread cht liu
Do you enable the spark fault tolerance mechanism, RDD run at the end of the job, will start a separate job, to the checkpoint data written to the file system before the persistence of high availability 2017-03-08 2:45 GMT+08:00 Swapnil Shinde : > Hello all >I have