spark git commit: Added inline comment for the canEqual PR by @cloud-fan.

2015-07-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 358e7bf65 -> 111c05538 Added inline comment for the canEqual PR by @cloud-fan. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/111c0553 Tree: http://git-wip-us.apache.o

spark git commit: [SPARK-9126] [MLLIB] do not assert on time taken by Thread.sleep()

2015-07-16 Thread meng
Repository: spark Updated Branches: refs/heads/master 322d286bb -> 358e7bf65 [SPARK-9126] [MLLIB] do not assert on time taken by Thread.sleep() Measure lower and upper bounds for task time and use them for validation. This PR also implements `Stopwatch.toString`. This suite should finish in l

[2/2] spark git commit: [SPARK-7131] [ML] Copy Decision Tree, Random Forest impl to spark.ml

2015-07-16 Thread meng
[SPARK-7131] [ML] Copy Decision Tree, Random Forest impl to spark.ml This PR copies the RandomForest implementation from spark.mllib to spark.ml. Note that this includes the DecisionTree implementation, but not the GradientBoostedTrees one (which will come later). I essentially copied a minima

[1/2] spark git commit: [SPARK-7131] [ML] Copy Decision Tree, Random Forest impl to spark.ml

2015-07-16 Thread meng
Repository: spark Updated Branches: refs/heads/master f893955b9 -> 322d286bb http://git-wip-us.apache.org/repos/asf/spark/blob/322d286b/mllib/src/main/scala/org/apache/spark/mllib/tree/impl/DTStatsAggregator.scala -- diff --git

spark git commit: [SPARK-8899] [SQL] remove duplicated equals method for Row

2015-07-16 Thread davies
Repository: spark Updated Branches: refs/heads/master 812b63bbe -> f893955b9 [SPARK-8899] [SQL] remove duplicated equals method for Row Author: Wenchen Fan Closes #7291 from cloud-fan/row and squashes the following commits: a11addf [Wenchen Fan] move hashCode back to internal row 2de6180 [W

spark git commit: [SPARK-8857][SPARK-8859][Core]Add an internal flag to Accumulable and send internal accumulator updates to the driver via heartbeats

2015-07-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 96aa3340f -> 812b63bbe [SPARK-8857][SPARK-8859][Core]Add an internal flag to Accumulable and send internal accumulator updates to the driver via heartbeats This PR includes the following changes: 1. Remove the thread local `Accumulators.l

spark git commit: [SPARK-8119] HeartbeatReceiver should replace executors, not kill

2015-07-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/master d86bbb4e2 -> 96aa3340f [SPARK-8119] HeartbeatReceiver should replace executors, not kill **Symptom.** If an executor in an application times out, `HeartbeatReceiver` attempts to kill it. After this happens, however, the application never g

spark git commit: [SPARK-6284] [MESOS] Add mesos role, principal and secret

2015-07-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/master 49351c7f5 -> d86bbb4e2 [SPARK-6284] [MESOS] Add mesos role, principal and secret Mesos supports framework authentication and role to be set per framework, which the role is used to identify the framework's role which impacts the sharing w

spark git commit: [SPARK-8646] PySpark does not run on YARN if master not provided in command line

2015-07-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/master 57e9b13bf -> 49351c7f5 [SPARK-8646] PySpark does not run on YARN if master not provided in command line andrewor14 davies vanzin can you take a look at this? thanks Author: Lianhui Wang Closes #7438 from lianhuiwang/SPARK-8646 and squash

spark git commit: [SPARK-8644] Include call site in SparkException stack traces thrown by job failures

2015-07-16 Thread adav
Repository: spark Updated Branches: refs/heads/master 031d7d414 -> 57e9b13bf [SPARK-8644] Include call site in SparkException stack traces thrown by job failures Example exception (new part at bottom, clearly demarcated): ``` org.apache.spark.SparkException: Job aborted due to stage failure:

spark git commit: [SPARK-6304] [STREAMING] Fix checkpointing doesn't retain driver port issue.

2015-07-16 Thread tdas
Repository: spark Updated Branches: refs/heads/master fec10f0c6 -> 031d7d414 [SPARK-6304] [STREAMING] Fix checkpointing doesn't retain driver port issue. Author: jerryshao Author: Saisai Shao Closes #5060 from jerryshao/SPARK-6304 and squashes the following commits: 89b01f5 [jerryshao] Upd

spark git commit: [SPARK-9085][SQL] Remove LeafNode, UnaryNode, BinaryNode from TreeNode.

2015-07-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 43dac2c88 -> fec10f0c6 [SPARK-9085][SQL] Remove LeafNode, UnaryNode, BinaryNode from TreeNode. This builds on #7433 but also removes LeafNode/UnaryNode. These are slightly more complicated to remove. I had to change some abstract classes t

spark git commit: [SPARK-6941] [SQL] Provide a better error message to when inserting into RDD based table

2015-07-16 Thread yhuai
Repository: spark Updated Branches: refs/heads/master b536d5dc6 -> 43dac2c88 [SPARK-6941] [SQL] Provide a better error message to when inserting into RDD based table JIRA: https://issues.apache.org/jira/browse/SPARK-6941 Author: Yijie Shen Closes #7342 from yijieshen/SPARK-6941 and squashe

spark git commit: [SPARK-9015] [BUILD] Clean project import in scala ide

2015-07-16 Thread srowen
Repository: spark Updated Branches: refs/heads/master 4ea6480a3 -> b536d5dc6 [SPARK-9015] [BUILD] Clean project import in scala ide Cleanup maven for a clean import in scala-ide / eclipse. * remove groovy plugin which is really not needed at all * add-source from build-helper-maven-plugin is

spark git commit: [SPARK-8995] [SQL] cast date strings like '2015-01-01 12:15:31' to date

2015-07-16 Thread davies
Repository: spark Updated Branches: refs/heads/master 011551620 -> 4ea6480a3 [SPARK-8995] [SQL] cast date strings like '2015-01-01 12:15:31' to date Jira https://issues.apache.org/jira/browse/SPARK-8995 In PR #6981we noticed that we cannot cast date strings that contains a time, like '2015-0

spark git commit: [SPARK-8893] Add runtime checks against non-positive number of partitions

2015-07-16 Thread srowen
Repository: spark Updated Branches: refs/heads/master 0a795336d -> 011551620 [SPARK-8893] Add runtime checks against non-positive number of partitions https://issues.apache.org/jira/browse/SPARK-8893 > What does `sc.parallelize(1 to 3).repartition(p).collect` return? I would > expect `Array(