Re: Please add me

2016-11-29 Thread Bu Jianjian
Hi Srinivas, You can subscribe the mail list in the community page by yourself http://spark.apache.org/community.html On Tue, Nov 29, 2016 at 9:59 AM, Srinivas Potluri wrote: > Hi, > > I am interested to contribute code on Spark. Could you please add me into > the

Proposal for SPARK-18278

2016-11-29 Thread Matt Cheah
Hi everyone, Kubernetes is a technology that is a key player in the cluster computing world. Currently, running Spark applications on Kubernetes requires deploying a standalone Spark cluster on the Kubernetes cluster, and then running the jobs against the standalone Spark cluster. However,

Re: How is the order ensured in the jdbc relation provider when inserting data from multiple executors

2016-11-29 Thread Sachith Withana
Hi all, To explain the scenario a bit more. We need to retain the order when writing to the RDBMS tables. The way we found was to execute the DB Write *job* for each partition which is really costly. One reason being that the partition count is really high( 200) and it seems we cannot control

Spark-9487, Need some insight

2016-11-29 Thread Saikat Kanjilal
Hello Spark dev community, I took this the following jira item (https://github.com/apache/spark/pull/15848) and am looking for some general pointers, it seems that I am running into issues where things work successfully doing local development on my macbook pro but fail on jenkins for a

Question about spark.mllib.GradientDescent

2016-11-29 Thread WangJianfei
Hi devs: I think it's unnecessary to use c1._1 += c2.1 in combOp operation, I think it's the same if we use c1._1+c2._1, see the code below : in GradientDescent.scala val (gradientSum, lossSum, miniBatchSize) = data.sample(false, miniBatchFraction, 42 + i)

Re: Can't read tables written in Spark 2.1 in Spark 2.0 (and earlier)

2016-11-29 Thread Michael Allman
This is not an issue with all tables created in Spark 2.1, though I'm not sure why some work and some do not. I have found that a table created as such sql("create table test stored as parquet as select 1") in Spark 2.1 cannot be read in previous versions of Spark. Michael > On Nov 29, 2016,

Can't read tables written in Spark 2.1 in Spark 2.0 (and earlier)

2016-11-29 Thread Michael Allman
Hello, When I try to read from a Hive table created by Spark 2.1 in Spark 2.0 or earlier, I get an error: java.lang.ClassNotFoundException: Failed to load class for data source: hive. Is there a way to get previous versions of Spark to read tables written with Spark 2.1? Cheers, Michael

Please add me

2016-11-29 Thread Srinivas Potluri
Hi, I am interested to contribute code on Spark. Could you please add me into the mailing list / DL. Thanks, *Srinivas Potluri*

Re: [VOTE] Apache Spark 2.1.0 (RC1)

2016-11-29 Thread Marcelo Vanzin
I'll send a -1 because of SPARK-18546. Haven't looked at anything else yet. On Mon, Nov 28, 2016 at 5:25 PM, Reynold Xin wrote: > Please vote on releasing the following candidate as Apache Spark version > 2.1.0. The vote is open until Thursday, December 1, 2016 at 18:00 UTC

Re: [VOTE] Apache Spark 2.1.0 (RC1)

2016-11-29 Thread Sean Owen
We still have several blockers for 2.1, so I imagine at least one will mean this won't be the final RC: SPARK-18318 ML, Graph 2.1 QA: API: New Scala APIs, docs SPARK-18319 ML, Graph 2.1 QA: API: Experimental, DeveloperApi, final, sealed audit SPARK-18326 SparkR 2.1 QA: New R APIs and API docs