Repository: hive Updated Branches: refs/heads/master 705979e68 -> 3ad42415a
HIVE-17153: Flaky test: TestMiniSparkOnYarnCliDriver[spark_dynamic_partition_pruning] (Sahil Takiar, reviewed by Rui Li) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/3ad42415 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/3ad42415 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/3ad42415 Branch: refs/heads/master Commit: 3ad42415adad82d776c3eaa3adb6c12684c70b35 Parents: 705979e Author: Sahil Takiar <takiar.sa...@gmail.com> Authored: Fri Jul 28 09:40:52 2017 -0700 Committer: Sahil Takiar <stak...@cloudera.com> Committed: Fri Jul 28 09:40:52 2017 -0700 ---------------------------------------------------------------------- .../apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java | 4 ++-- .../spark/spark_dynamic_partition_pruning.q.out | 8 ++++---- 2 files changed, 6 insertions(+), 6 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hive/blob/3ad42415/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java index 1348d8b..933d6af 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.parse.spark; import java.util.ArrayList; -import java.util.HashSet; +import java.util.LinkedHashSet; import java.util.LinkedList; import java.util.List; import java.util.Set; @@ -140,7 +140,7 @@ public class SplitOpTreeForDPP implements NodeProcessor { } //Find all pruningSinkSet in new roots - Set<Operator<?>> sinkSet = new HashSet<>(); + Set<Operator<?>> sinkSet = new LinkedHashSet<>(); for (Operator<?> root : newRoots) { SparkUtilities.collectOp(sinkSet, root, SparkPartitionPruningSinkOperator.class); } http://git-wip-us.apache.org/repos/asf/hive/blob/3ad42415/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out b/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out index e743af1..f1092c5 100644 --- a/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out +++ b/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out @@ -3700,7 +3700,7 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target column name: ds - target work: Map 4 + target work: Map 1 Select Operator expressions: _col0 (type: string) outputColumnNames: _col0 @@ -3714,7 +3714,7 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target column name: ds - target work: Map 1 + target work: Map 4 Reducer 13 Reduce Operator Tree: Group By Operator @@ -3743,7 +3743,7 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target column name: ds - target work: Map 4 + target work: Map 1 Select Operator expressions: _col0 (type: string) outputColumnNames: _col0 @@ -3757,7 +3757,7 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target column name: ds - target work: Map 1 + target work: Map 4 Stage: Stage-1 Spark