Repository: hive
Updated Branches:
  refs/heads/master 705979e68 -> 3ad42415a


HIVE-17153: Flaky test: 
TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] (Sahil Takiar, 
reviewed by Rui Li)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/3ad42415
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/3ad42415
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/3ad42415

Branch: refs/heads/master
Commit: 3ad42415adad82d776c3eaa3adb6c12684c70b35
Parents: 705979e
Author: Sahil Takiar <takiar.sa...@gmail.com>
Authored: Fri Jul 28 09:40:52 2017 -0700
Committer: Sahil Takiar <stak...@cloudera.com>
Committed: Fri Jul 28 09:40:52 2017 -0700

----------------------------------------------------------------------
 .../apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java | 4 ++--
 .../spark/spark_dynamic_partition_pruning.q.out              | 8 ++++----
 2 files changed, 6 insertions(+), 6 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hive/blob/3ad42415/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java 
b/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java
index 1348d8b..933d6af 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java
@@ -19,7 +19,7 @@
 package org.apache.hadoop.hive.ql.parse.spark;
 
 import java.util.ArrayList;
-import java.util.HashSet;
+import java.util.LinkedHashSet;
 import java.util.LinkedList;
 import java.util.List;
 import java.util.Set;
@@ -140,7 +140,7 @@ public class SplitOpTreeForDPP implements NodeProcessor {
     }
 
     //Find all pruningSinkSet in new roots
-    Set<Operator<?>> sinkSet = new HashSet<>();
+    Set<Operator<?>> sinkSet = new LinkedHashSet<>();
     for (Operator<?> root : newRoots) {
       SparkUtilities.collectOp(sinkSet, root, 
SparkPartitionPruningSinkOperator.class);
     }

http://git-wip-us.apache.org/repos/asf/hive/blob/3ad42415/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out
 
b/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out
index e743af1..f1092c5 100644
--- 
a/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out
+++ 
b/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out
@@ -3700,7 +3700,7 @@ STAGE PLANS:
                           partition key expr: ds
                           Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
                           target column name: ds
-                          target work: Map 4
+                          target work: Map 1
                     Select Operator
                       expressions: _col0 (type: string)
                       outputColumnNames: _col0
@@ -3714,7 +3714,7 @@ STAGE PLANS:
                           partition key expr: ds
                           Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
                           target column name: ds
-                          target work: Map 1
+                          target work: Map 4
         Reducer 13 
             Reduce Operator Tree:
               Group By Operator
@@ -3743,7 +3743,7 @@ STAGE PLANS:
                           partition key expr: ds
                           Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
                           target column name: ds
-                          target work: Map 4
+                          target work: Map 1
                     Select Operator
                       expressions: _col0 (type: string)
                       outputColumnNames: _col0
@@ -3757,7 +3757,7 @@ STAGE PLANS:
                           partition key expr: ds
                           Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
                           target column name: ds
-                          target work: Map 1
+                          target work: Map 4
 
   Stage: Stage-1
     Spark

Reply via email to