Repository: hive
Updated Branches:
  refs/heads/master 90e682887 -> bff8d95e9


HIVE-17414: HoS DPP + Vectorization generates invalid explain plan due to 
CombineEquivalentWorkResolver (Liyun Zhang, reviewed by Sahil Takiar and Li Rui)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/bff8d95e
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/bff8d95e
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/bff8d95e

Branch: refs/heads/master
Commit: bff8d95e91f94770fb0ae4a96009e22d16d7ddfa
Parents: 90e6828
Author: Ferdinand Xu <cheng.a...@intel.com>
Authored: Tue Sep 5 13:36:08 2017 +0800
Committer: Ferdinand Xu <cheng.a...@intel.com>
Committed: Tue Sep 5 13:36:08 2017 +0800

----------------------------------------------------------------------
 .../hive/ql/exec/spark/SparkUtilities.java      |  4 +--
 ...k_vectorized_dynamic_partition_pruning.q.out | 28 --------------------
 2 files changed, 2 insertions(+), 30 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hive/blob/bff8d95e/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java
----------------------------------------------------------------------
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java
index fac3cea..ca19fd0 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/SparkUtilities.java
@@ -191,7 +191,7 @@ public class SparkUtilities {
   }
 
   /**
-   * Recursively find all operators under root, that are of class clazz, and
+   * Recursively find all operators under root, that are of class clazz or are 
the sub-class of clazz, and
    * put them in result.
    * @param result all operators under root that are of class clazz
    * @param root the root operator under which all operators will be examined
@@ -202,7 +202,7 @@ public class SparkUtilities {
     if (root == null) {
       return;
     }
-    if (clazz.equals(root.getClass())) {
+    if (clazz.isAssignableFrom(root.getClass())) {
       result.add(root);
     }
     for (Operator<?> child : root.getChildOperators()) {

http://git-wip-us.apache.org/repos/asf/hive/blob/bff8d95e/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out
 
b/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out
index 7df1b9e..7cafb52 100644
--- 
a/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out
+++ 
b/ql/src/test/results/clientpositive/spark/spark_vectorized_dynamic_partition_pruning.q.out
@@ -4353,20 +4353,6 @@ STAGE PLANS:
                           partition key expr: ds
                           Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
                           target work: Map 1
-                    Select Operator
-                      expressions: _col0 (type: string)
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
-                      Group By Operator
-                        keys: _col0 (type: string)
-                        mode: hash
-                        outputColumnNames: _col0
-                        Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
-                        Spark Partition Pruning Sink Operator
-                          Target column: ds (string)
-                          partition key expr: ds
-                          Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
-                          target work: Map 4
         Reducer 13 
             Execution mode: vectorized
             Reduce Operator Tree:
@@ -4397,20 +4383,6 @@ STAGE PLANS:
                           partition key expr: ds
                           Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
                           target work: Map 1
-                    Select Operator
-                      expressions: _col0 (type: string)
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
-                      Group By Operator
-                        keys: _col0 (type: string)
-                        mode: hash
-                        outputColumnNames: _col0
-                        Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
-                        Spark Partition Pruning Sink Operator
-                          Target column: ds (string)
-                          partition key expr: ds
-                          Statistics: Num rows: 2 Data size: 368 Basic stats: 
COMPLETE Column stats: NONE
-                          target work: Map 4
 
   Stage: Stage-1
     Spark

Reply via email to