git commit: [HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD zipPartitions"

2014-09-03 Thread ankurdave
Repository: spark
Updated Branches:
  refs/heads/master 1bed0a386 -> 00362dac9


[HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD zipPartitions"

9b225ac3072de522b40b46aba6df1f1c231f13ef has been causing GraphX tests
to fail nondeterministically, which is blocking development for others.

Author: Ankur Dave 

Closes #2271 from ankurdave/SPARK-3400 and squashes the following commits:

10c2a97 [Ankur Dave] [HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD 
zipPartitions"


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/00362dac
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/00362dac
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/00362dac

Branch: refs/heads/master
Commit: 00362dac976cd05b06638deb11d990d612429e0b
Parents: 1bed0a3
Author: Ankur Dave 
Authored: Wed Sep 3 23:49:47 2014 -0700
Committer: Ankur Dave 
Committed: Wed Sep 3 23:49:47 2014 -0700

--
 .../scala/org/apache/spark/graphx/EdgeRDD.scala |  4 ++--
 .../scala/org/apache/spark/graphx/GraphSuite.scala  | 16 
 2 files changed, 2 insertions(+), 18 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/00362dac/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
--
diff --git a/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala 
b/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
index 35fbd47..5bcb96b 100644
--- a/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
+++ b/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
@@ -19,7 +19,7 @@ package org.apache.spark.graphx
 
 import scala.reflect.{classTag, ClassTag}
 
-import org.apache.spark._
+import org.apache.spark.{OneToOneDependency, Partition, Partitioner, 
TaskContext}
 import org.apache.spark.rdd.RDD
 import org.apache.spark.storage.StorageLevel
 
@@ -55,7 +55,7 @@ class EdgeRDD[@specialized ED: ClassTag, VD: ClassTag](
* partitioner that allows co-partitioning with `partitionsRDD`.
*/
   override val partitioner =
-partitionsRDD.partitioner.orElse(Some(new 
HashPartitioner(partitionsRDD.partitions.size)))
+
partitionsRDD.partitioner.orElse(Some(Partitioner.defaultPartitioner(partitionsRDD)))
 
   override def compute(part: Partition, context: TaskContext): 
Iterator[Edge[ED]] = {
 val p = firstParent[(PartitionID, EdgePartition[ED, VD])].iterator(part, 
context)

http://git-wip-us.apache.org/repos/asf/spark/blob/00362dac/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
--
diff --git a/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala 
b/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
index eaaa449..6506bac 100644
--- a/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
+++ b/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
@@ -19,7 +19,6 @@ package org.apache.spark.graphx
 
 import org.scalatest.FunSuite
 
-import org.apache.spark.SparkConf
 import org.apache.spark.SparkContext
 import org.apache.spark.graphx.Graph._
 import org.apache.spark.graphx.PartitionStrategy._
@@ -351,19 +350,4 @@ class GraphSuite extends FunSuite with LocalSparkContext {
 }
   }
 
-  test("non-default number of edge partitions") {
-val n = 10
-val defaultParallelism = 3
-val numEdgePartitions = 4
-assert(defaultParallelism != numEdgePartitions)
-val conf = new SparkConf()
-  .set("spark.default.parallelism", defaultParallelism.toString)
-val sc = new SparkContext("local", "test", conf)
-val edges = sc.parallelize((1 to n).map(x => (x: VertexId, 0: VertexId)),
-  numEdgePartitions)
-val graph = Graph.fromEdgeTuples(edges, 1)
-val neighborAttrSums = graph.mapReduceTriplets[Int](
-  et => Iterator((et.dstId, et.srcAttr)), _ + _)
-assert(neighborAttrSums.collect.toSet === Set((0: VertexId, n)))
-  }
 }


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD zipPartitions"

2014-09-03 Thread ankurdave
Repository: spark
Updated Branches:
  refs/heads/branch-1.0 8dd7690e2 -> 4d3ab2925


[HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD zipPartitions"

9b225ac3072de522b40b46aba6df1f1c231f13ef has been causing GraphX tests
to fail nondeterministically, which is blocking development for others.

Author: Ankur Dave 

Closes #2271 from ankurdave/SPARK-3400 and squashes the following commits:

10c2a97 [Ankur Dave] [HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD 
zipPartitions"

(cherry picked from commit 00362dac976cd05b06638deb11d990d612429e0b)
Signed-off-by: Ankur Dave 


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/4d3ab292
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/4d3ab292
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/4d3ab292

Branch: refs/heads/branch-1.0
Commit: 4d3ab292576396e11d9e5c2dabb676c07b34d286
Parents: 8dd7690
Author: Ankur Dave 
Authored: Wed Sep 3 23:49:47 2014 -0700
Committer: Ankur Dave 
Committed: Wed Sep 3 23:50:22 2014 -0700

--
 .../scala/org/apache/spark/graphx/EdgeRDD.scala |  4 ++--
 .../scala/org/apache/spark/graphx/GraphSuite.scala  | 16 
 2 files changed, 2 insertions(+), 18 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/4d3ab292/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
--
diff --git a/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala 
b/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
index 4dd15bf..a8fc095 100644
--- a/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
+++ b/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
@@ -19,7 +19,7 @@ package org.apache.spark.graphx
 
 import scala.reflect.{classTag, ClassTag}
 
-import org.apache.spark._
+import org.apache.spark.{OneToOneDependency, Partition, Partitioner, 
TaskContext}
 import org.apache.spark.rdd.RDD
 import org.apache.spark.storage.StorageLevel
 
@@ -45,7 +45,7 @@ class EdgeRDD[@specialized ED: ClassTag, VD: ClassTag](
* partitioner that allows co-partitioning with `partitionsRDD`.
*/
   override val partitioner =
-partitionsRDD.partitioner.orElse(Some(new 
HashPartitioner(partitionsRDD.partitions.size)))
+
partitionsRDD.partitioner.orElse(Some(Partitioner.defaultPartitioner(partitionsRDD)))
 
   override def compute(part: Partition, context: TaskContext): 
Iterator[Edge[ED]] = {
 val p = firstParent[(PartitionID, EdgePartition[ED, VD])].iterator(part, 
context)

http://git-wip-us.apache.org/repos/asf/spark/blob/4d3ab292/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
--
diff --git a/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala 
b/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
index e1b83c2..abc25d0 100644
--- a/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
+++ b/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
@@ -19,7 +19,6 @@ package org.apache.spark.graphx
 
 import org.scalatest.FunSuite
 
-import org.apache.spark.SparkConf
 import org.apache.spark.SparkContext
 import org.apache.spark.graphx.Graph._
 import org.apache.spark.graphx.PartitionStrategy._
@@ -326,19 +325,4 @@ class GraphSuite extends FunSuite with LocalSparkContext {
 }
   }
 
-  test("non-default number of edge partitions") {
-val n = 10
-val defaultParallelism = 3
-val numEdgePartitions = 4
-assert(defaultParallelism != numEdgePartitions)
-val conf = new SparkConf()
-  .set("spark.default.parallelism", defaultParallelism.toString)
-val sc = new SparkContext("local", "test", conf)
-val edges = sc.parallelize((1 to n).map(x => (x: VertexId, 0: VertexId)),
-  numEdgePartitions)
-val graph = Graph.fromEdgeTuples(edges, 1)
-val neighborAttrSums = graph.mapReduceTriplets[Int](
-  et => Iterator((et.dstId, et.srcAttr)), _ + _)
-assert(neighborAttrSums.collect.toSet === Set((0: VertexId, n)))
-  }
 }


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD zipPartitions"

2014-09-03 Thread ankurdave
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 f41c45a75 -> 8c40ab5c0


[HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD zipPartitions"

9b225ac3072de522b40b46aba6df1f1c231f13ef has been causing GraphX tests
to fail nondeterministically, which is blocking development for others.

Author: Ankur Dave 

Closes #2271 from ankurdave/SPARK-3400 and squashes the following commits:

10c2a97 [Ankur Dave] [HOTFIX] [SPARK-3400] Revert 9b225ac "fix GraphX EdgeRDD 
zipPartitions"

(cherry picked from commit 00362dac976cd05b06638deb11d990d612429e0b)
Signed-off-by: Ankur Dave 


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8c40ab5c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8c40ab5c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8c40ab5c

Branch: refs/heads/branch-1.1
Commit: 8c40ab5c06ab72e85a8a9d4272fed0e81eca1d3a
Parents: f41c45a
Author: Ankur Dave 
Authored: Wed Sep 3 23:49:47 2014 -0700
Committer: Ankur Dave 
Committed: Wed Sep 3 23:50:11 2014 -0700

--
 .../scala/org/apache/spark/graphx/EdgeRDD.scala |  4 ++--
 .../scala/org/apache/spark/graphx/GraphSuite.scala  | 16 
 2 files changed, 2 insertions(+), 18 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/8c40ab5c/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
--
diff --git a/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala 
b/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
index 0f1a101..899a3cb 100644
--- a/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
+++ b/graphx/src/main/scala/org/apache/spark/graphx/EdgeRDD.scala
@@ -19,7 +19,7 @@ package org.apache.spark.graphx
 
 import scala.reflect.{classTag, ClassTag}
 
-import org.apache.spark._
+import org.apache.spark.{OneToOneDependency, Partition, Partitioner, 
TaskContext}
 import org.apache.spark.rdd.RDD
 import org.apache.spark.storage.StorageLevel
 
@@ -47,7 +47,7 @@ class EdgeRDD[@specialized ED: ClassTag, VD: ClassTag](
* partitioner that allows co-partitioning with `partitionsRDD`.
*/
   override val partitioner =
-partitionsRDD.partitioner.orElse(Some(new 
HashPartitioner(partitionsRDD.partitions.size)))
+
partitionsRDD.partitioner.orElse(Some(Partitioner.defaultPartitioner(partitionsRDD)))
 
   override def compute(part: Partition, context: TaskContext): 
Iterator[Edge[ED]] = {
 val p = firstParent[(PartitionID, EdgePartition[ED, VD])].iterator(part, 
context)

http://git-wip-us.apache.org/repos/asf/spark/blob/8c40ab5c/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
--
diff --git a/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala 
b/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
index eaaa449..6506bac 100644
--- a/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
+++ b/graphx/src/test/scala/org/apache/spark/graphx/GraphSuite.scala
@@ -19,7 +19,6 @@ package org.apache.spark.graphx
 
 import org.scalatest.FunSuite
 
-import org.apache.spark.SparkConf
 import org.apache.spark.SparkContext
 import org.apache.spark.graphx.Graph._
 import org.apache.spark.graphx.PartitionStrategy._
@@ -351,19 +350,4 @@ class GraphSuite extends FunSuite with LocalSparkContext {
 }
   }
 
-  test("non-default number of edge partitions") {
-val n = 10
-val defaultParallelism = 3
-val numEdgePartitions = 4
-assert(defaultParallelism != numEdgePartitions)
-val conf = new SparkConf()
-  .set("spark.default.parallelism", defaultParallelism.toString)
-val sc = new SparkContext("local", "test", conf)
-val edges = sc.parallelize((1 to n).map(x => (x: VertexId, 0: VertexId)),
-  numEdgePartitions)
-val graph = Graph.fromEdgeTuples(edges, 1)
-val neighborAttrSums = graph.mapReduceTriplets[Int](
-  et => Iterator((et.dstId, et.srcAttr)), _ + _)
-assert(neighborAttrSums.collect.toSet === Set((0: VertexId, n)))
-  }
 }


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org