Repository: hive Updated Branches: refs/heads/master 90e682887 -> bff8d95e9
HIVE-17414: HoS DPP + Vectorization generates invalid explain plan due to CombineEquivalentWorkResolver (Liyun Zhang, reviewed by Sahil Takiar and Li Rui) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/bff8d95e Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/bff8d95e Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/bff8d95e Branch: refs/heads/master Commit: bff8d95e91f94770fb0ae4a96009e22d16d7ddfa Parents: 90e6828 Author: Ferdinand Xu <cheng.a...@intel.com> Authored: Tue Sep 5 13:36:08 2017 +0800 Committer: Ferdinand Xu <cheng.a...@intel.com> Committed: Tue Sep 5 13:36:08 2017 +0800 ---------------------------------------------------------------------- .../hive/ql/exec/spark/SparkUtilities.java | 4 +-- ...k_vectorized_dynamic_partition_pruning.q.out | 28 -------------------- 2 files changed, 2 insertions(+), 30 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hive/blob/bff8d95e/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java index fac3cea..ca19fd0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java @@ -191,7 +191,7 @@ public class SparkUtilities { } /** - * Recursively find all operators under root, that are of class clazz, and + * Recursively find all operators under root, that are of class clazz or are the sub-class of clazz, and * put them in result. * @param result all operators under root that are of class clazz * @param root the root operator under which all operators will be examined @@ -202,7 +202,7 @@ public class SparkUtilities { if (root == null) { return; } - if (clazz.equals(root.getClass())) { + if (clazz.isAssignableFrom(root.getClass())) { result.add(root); } for (Operator<?> child : root.getChildOperators()) { http://git-wip-us.apache.org/repos/asf/hive/blob/bff8d95e/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out b/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out index 7df1b9e..7cafb52 100644 --- a/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out +++ b/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out @@ -4353,20 +4353,6 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target work: Map 1 - Select Operator - expressions: _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: string) - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Spark Partition Pruning Sink Operator - Target column: ds (string) - partition key expr: ds - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE - target work: Map 4 Reducer 13 Execution mode: vectorized Reduce Operator Tree: @@ -4397,20 +4383,6 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target work: Map 1 - Select Operator - expressions: _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: string) - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Spark Partition Pruning Sink Operator - Target column: ds (string) - partition key expr: ds - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE - target work: Map 4 Stage: Stage-1 Spark