http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/remove_exprs_stats.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/remove_exprs_stats.q.out 
b/ql/src/test/results/clientpositive/remove_exprs_stats.q.out
index b1ba942..567e6b2 100644
--- a/ql/src/test/results/clientpositive/remove_exprs_stats.q.out
+++ b/ql/src/test/results/clientpositive/remove_exprs_stats.q.out
@@ -74,11 +74,11 @@ STAGE PLANS:
       Processor Tree:
         TableScan
           alias: loc_orc
-          Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE Column 
stats: COMPLETE
+          Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column 
stats: COMPLETE
           Select Operator
             expressions: state (type: string), locid (type: int), zip (type: 
bigint), year (type: int)
             outputColumnNames: _col0, _col1, _col2, _col3
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             ListSink
 
 PREHOOK: query: explain select * from loc_orc where locid > 30
@@ -95,7 +95,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: false (type: boolean)
               Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -131,11 +131,11 @@ STAGE PLANS:
       Processor Tree:
         TableScan
           alias: loc_orc
-          Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE Column 
stats: COMPLETE
+          Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column 
stats: COMPLETE
           Select Operator
             expressions: state (type: string), locid (type: int), zip (type: 
bigint), year (type: int)
             outputColumnNames: _col0, _col1, _col2, _col3
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             ListSink
 
 PREHOOK: query: explain select * from loc_orc where locid >= 30
@@ -152,7 +152,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: false (type: boolean)
               Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -188,7 +188,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: (locid < 6) (type: boolean)
               Statistics: Num rows: 2 Data size: 204 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -224,7 +224,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: false (type: boolean)
               Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -260,11 +260,11 @@ STAGE PLANS:
       Processor Tree:
         TableScan
           alias: loc_orc
-          Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE Column 
stats: COMPLETE
+          Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column 
stats: COMPLETE
           Select Operator
             expressions: state (type: string), locid (type: int), zip (type: 
bigint), year (type: int)
             outputColumnNames: _col0, _col1, _col2, _col3
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             ListSink
 
 PREHOOK: query: explain select * from loc_orc where locid >= 6
@@ -281,7 +281,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: (locid >= 6) (type: boolean)
               Statistics: Num rows: 2 Data size: 204 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -317,7 +317,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: false (type: boolean)
               Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -353,7 +353,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: (locid > 1) (type: boolean)
               Statistics: Num rows: 2 Data size: 204 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -389,7 +389,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: (locid <= 1) (type: boolean)
               Statistics: Num rows: 2 Data size: 204 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -425,11 +425,11 @@ STAGE PLANS:
       Processor Tree:
         TableScan
           alias: loc_orc
-          Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE Column 
stats: COMPLETE
+          Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE Column 
stats: COMPLETE
           Select Operator
             expressions: state (type: string), locid (type: int), zip (type: 
bigint), year (type: int)
             outputColumnNames: _col0, _col1, _col2, _col3
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             ListSink
 
 PREHOOK: query: explain select * from loc_orc where locid IN (-4,5,30,40)
@@ -446,7 +446,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: (locid) IN (5) (type: boolean)
               Statistics: Num rows: 2 Data size: 204 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -482,17 +482,17 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: (locid) IN (5, 2, 3) (type: boolean)
-              Statistics: Num rows: 5 Data size: 498 Basic stats: COMPLETE 
Column stats: COMPLETE
+              Statistics: Num rows: 5 Data size: 510 Basic stats: COMPLETE 
Column stats: COMPLETE
               Select Operator
                 expressions: state (type: string), locid (type: int), zip 
(type: bigint), year (type: int)
                 outputColumnNames: _col0, _col1, _col2, _col3
-                Statistics: Num rows: 5 Data size: 498 Basic stats: COMPLETE 
Column stats: COMPLETE
+                Statistics: Num rows: 5 Data size: 510 Basic stats: COMPLETE 
Column stats: COMPLETE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 5 Data size: 498 Basic stats: COMPLETE 
Column stats: COMPLETE
+                  Statistics: Num rows: 5 Data size: 510 Basic stats: COMPLETE 
Column stats: COMPLETE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -518,7 +518,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: (locid) IN (1, 6) (type: boolean)
               Statistics: Num rows: 3 Data size: 306 Basic stats: COMPLETE 
Column stats: COMPLETE
@@ -554,7 +554,7 @@ STAGE PLANS:
       Map Operator Tree:
           TableScan
             alias: loc_orc
-            Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+            Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
             Filter Operator
               predicate: false (type: boolean)
               Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE 
Column stats: COMPLETE

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out 
b/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out
index 5f06358..31ccc5c 100644
--- a/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out
+++ b/ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out
@@ -565,19 +565,19 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: l
-                  Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+                  Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
                   Filter Operator
                     predicate: locid is not null (type: boolean)
-                    Statistics: Num rows: 8 Data size: 804 Basic stats: 
COMPLETE Column stats: COMPLETE
+                    Statistics: Num rows: 8 Data size: 816 Basic stats: 
COMPLETE Column stats: COMPLETE
                     Select Operator
                       expressions: state (type: string), locid (type: int), 
zip (type: bigint), year (type: int)
                       outputColumnNames: _col0, _col1, _col2, _col3
-                      Statistics: Num rows: 8 Data size: 804 Basic stats: 
COMPLETE Column stats: COMPLETE
+                      Statistics: Num rows: 8 Data size: 816 Basic stats: 
COMPLETE Column stats: COMPLETE
                       Reduce Output Operator
                         key expressions: _col1 (type: int)
                         sort order: +
                         Map-reduce partition columns: _col1 (type: int)
-                        Statistics: Num rows: 8 Data size: 804 Basic stats: 
COMPLETE Column stats: COMPLETE
+                        Statistics: Num rows: 8 Data size: 816 Basic stats: 
COMPLETE Column stats: COMPLETE
                         value expressions: _col0 (type: string), _col2 (type: 
bigint), _col3 (type: int)
         Reducer 2 
             Reduce Operator Tree:
@@ -659,19 +659,19 @@ STAGE PLANS:
             Map Operator Tree:
                 TableScan
                   alias: l
-                  Statistics: Num rows: 8 Data size: 804 Basic stats: COMPLETE 
Column stats: COMPLETE
+                  Statistics: Num rows: 8 Data size: 816 Basic stats: COMPLETE 
Column stats: COMPLETE
                   Filter Operator
                     predicate: (state is not null and locid is not null) 
(type: boolean)
-                    Statistics: Num rows: 8 Data size: 804 Basic stats: 
COMPLETE Column stats: COMPLETE
+                    Statistics: Num rows: 8 Data size: 816 Basic stats: 
COMPLETE Column stats: COMPLETE
                     Select Operator
                       expressions: state (type: string), locid (type: int), 
zip (type: bigint), year (type: int)
                       outputColumnNames: _col0, _col1, _col2, _col3
-                      Statistics: Num rows: 8 Data size: 804 Basic stats: 
COMPLETE Column stats: COMPLETE
+                      Statistics: Num rows: 8 Data size: 816 Basic stats: 
COMPLETE Column stats: COMPLETE
                       Reduce Output Operator
                         key expressions: _col0 (type: string), _col1 (type: 
int)
                         sort order: ++
                         Map-reduce partition columns: _col0 (type: string), 
_col1 (type: int)
-                        Statistics: Num rows: 8 Data size: 804 Basic stats: 
COMPLETE Column stats: COMPLETE
+                        Statistics: Num rows: 8 Data size: 816 Basic stats: 
COMPLETE Column stats: COMPLETE
                         value expressions: _col2 (type: bigint), _col3 (type: 
int)
         Reducer 2 
             Reduce Operator Tree:

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/groupby3.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/groupby3.q.out 
b/ql/src/test/results/clientpositive/spark/groupby3.q.out
index e48018c..23871ba 100644
--- a/ql/src/test/results/clientpositive/spark/groupby3.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby3.q.out
@@ -65,10 +65,10 @@ STAGE PLANS:
                 aggregations: sum(KEY._col0:0._col0), avg(KEY._col0:0._col0), 
avg(DISTINCT KEY._col0:0._col0), max(KEY._col0:0._col0), 
min(KEY._col0:0._col0), std(KEY._col0:0._col0), stddev_samp(KEY._col0:0._col0), 
variance(KEY._col0:0._col0), var_samp(KEY._col0:0._col0)
                 mode: partial1
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                Statistics: Num rows: 1 Data size: 176 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 1208 Basic stats: COMPLETE 
Column stats: NONE
                 Reduce Output Operator
                   sort order: 
-                  Statistics: Num rows: 1 Data size: 176 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 1208 Basic stats: 
COMPLETE Column stats: NONE
                   value expressions: _col0 (type: double), _col1 (type: 
struct<count:bigint,sum:double,input:string>), _col2 (type: 
struct<count:bigint,sum:double,input:string>), _col3 (type: string), _col4 
(type: string), _col5 (type: struct<count:bigint,sum:double,variance:double>), 
_col6 (type: struct<count:bigint,sum:double,variance:double>), _col7 (type: 
struct<count:bigint,sum:double,variance:double>), _col8 (type: 
struct<count:bigint,sum:double,variance:double>)
         Reducer 3 
             Reduce Operator Tree:
@@ -76,14 +76,14 @@ STAGE PLANS:
                 aggregations: sum(VALUE._col0), avg(VALUE._col1), 
avg(VALUE._col2), max(VALUE._col3), min(VALUE._col4), std(VALUE._col5), 
stddev_samp(VALUE._col6), variance(VALUE._col7), var_samp(VALUE._col8)
                 mode: final
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                Statistics: Num rows: 1 Data size: 224 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 1208 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), _col1 (type: double), 
_col2 (type: double), UDFToDouble(_col3) (type: double), UDFToDouble(_col4) 
(type: double), _col5 (type: double), _col6 (type: double), _col7 (type: 
double), _col8 (type: double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                  Statistics: Num rows: 1 Data size: 224 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 1208 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 224 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 1208 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/groupby3_map.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/groupby3_map.q.out 
b/ql/src/test/results/clientpositive/spark/groupby3_map.q.out
index f806303..71f8dc0 100644
--- a/ql/src/test/results/clientpositive/spark/groupby3_map.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby3_map.q.out
@@ -70,14 +70,14 @@ STAGE PLANS:
                 aggregations: sum(VALUE._col0), avg(VALUE._col1), avg(DISTINCT 
KEY._col0:0._col0), max(VALUE._col3), min(VALUE._col4), std(VALUE._col5), 
stddev_samp(VALUE._col6), variance(VALUE._col7), var_samp(VALUE._col8)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                Statistics: Num rows: 1 Data size: 224 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 1216 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), _col1 (type: double), 
_col2 (type: double), UDFToDouble(_col3) (type: double), UDFToDouble(_col4) 
(type: double), _col5 (type: double), _col6 (type: double), _col7 (type: 
double), _col8 (type: double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                  Statistics: Num rows: 1 Data size: 224 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 1216 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 224 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 1216 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/groupby3_map_multi_distinct.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/groupby3_map_multi_distinct.q.out 
b/ql/src/test/results/clientpositive/spark/groupby3_map_multi_distinct.q.out
index 3b31dfe..47ef5cb 100644
--- a/ql/src/test/results/clientpositive/spark/groupby3_map_multi_distinct.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby3_map_multi_distinct.q.out
@@ -74,14 +74,14 @@ STAGE PLANS:
                 aggregations: sum(VALUE._col0), avg(VALUE._col1), avg(DISTINCT 
KEY._col0:0._col0), max(VALUE._col3), min(VALUE._col4), std(VALUE._col5), 
stddev_samp(VALUE._col6), variance(VALUE._col7), var_samp(VALUE._col8), 
sum(DISTINCT KEY._col0:1._col0), count(DISTINCT KEY._col0:2._col0)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10
-                Statistics: Num rows: 1 Data size: 248 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), _col1 (type: double), 
_col2 (type: double), UDFToDouble(_col3) (type: double), UDFToDouble(_col4) 
(type: double), _col5 (type: double), _col6 (type: double), _col7 (type: 
double), _col8 (type: double), _col9 (type: double), UDFToDouble(_col10) (type: 
double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10
-                  Statistics: Num rows: 1 Data size: 248 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 1232 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 248 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 1232 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out 
b/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out
index bbad6e7..7cfca81 100644
--- a/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby3_map_skew.q.out
@@ -72,10 +72,10 @@ STAGE PLANS:
                 aggregations: sum(VALUE._col0), avg(VALUE._col1), avg(DISTINCT 
KEY._col0:0._col0), max(VALUE._col3), min(VALUE._col4), std(VALUE._col5), 
stddev_samp(VALUE._col6), variance(VALUE._col7), var_samp(VALUE._col8)
                 mode: partials
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 1464 Basic stats: COMPLETE 
Column stats: NONE
                 Reduce Output Operator
                   sort order: 
-                  Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 1464 Basic stats: 
COMPLETE Column stats: NONE
                   value expressions: _col0 (type: double), _col1 (type: 
struct<count:bigint,sum:double,input:string>), _col2 (type: 
struct<count:bigint,sum:double,input:string>), _col3 (type: string), _col4 
(type: string), _col5 (type: struct<count:bigint,sum:double,variance:double>), 
_col6 (type: struct<count:bigint,sum:double,variance:double>), _col7 (type: 
struct<count:bigint,sum:double,variance:double>), _col8 (type: 
struct<count:bigint,sum:double,variance:double>)
         Reducer 3 
             Reduce Operator Tree:
@@ -83,14 +83,14 @@ STAGE PLANS:
                 aggregations: sum(VALUE._col0), avg(VALUE._col1), 
avg(VALUE._col2), max(VALUE._col3), min(VALUE._col4), std(VALUE._col5), 
stddev_samp(VALUE._col6), variance(VALUE._col7), var_samp(VALUE._col8)
                 mode: final
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                Statistics: Num rows: 1 Data size: 224 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 1464 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), _col1 (type: double), 
_col2 (type: double), UDFToDouble(_col3) (type: double), UDFToDouble(_col4) 
(type: double), _col5 (type: double), _col6 (type: double), _col7 (type: 
double), _col8 (type: double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                  Statistics: Num rows: 1 Data size: 224 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 1464 Basic stats: 
COMPLETE Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 224 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 1464 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/groupby3_noskew.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/groupby3_noskew.q.out 
b/ql/src/test/results/clientpositive/spark/groupby3_noskew.q.out
index 6868eff..b2993a6 100644
--- a/ql/src/test/results/clientpositive/spark/groupby3_noskew.q.out
+++ b/ql/src/test/results/clientpositive/spark/groupby3_noskew.q.out
@@ -63,14 +63,14 @@ STAGE PLANS:
                 aggregations: sum(KEY._col0:0._col0), avg(KEY._col0:0._col0), 
avg(DISTINCT KEY._col0:0._col0), max(KEY._col0:0._col0), 
min(KEY._col0:0._col0), std(KEY._col0:0._col0), stddev_samp(KEY._col0:0._col0), 
variance(KEY._col0:0._col0), var_samp(KEY._col0:0._col0)
                 mode: complete
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                Statistics: Num rows: 1 Data size: 224 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), _col1 (type: double), 
_col2 (type: double), UDFToDouble(_col3) (type: double), UDFToDouble(_col4) 
(type: double), _col5 (type: double), _col6 (type: double), _col7 (type: 
double), _col8 (type: double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                  Statistics: Num rows: 1 Data size: 224 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 224 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 424 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/groupby3_noskew_multi_distinct.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/groupby3_noskew_multi_distinct.q.out 
b/ql/src/test/results/clientpositive/spark/groupby3_noskew_multi_distinct.q.out
index 399bfd8..d152a07 100644
--- 
a/ql/src/test/results/clientpositive/spark/groupby3_noskew_multi_distinct.q.out
+++ 
b/ql/src/test/results/clientpositive/spark/groupby3_noskew_multi_distinct.q.out
@@ -67,14 +67,14 @@ STAGE PLANS:
                 aggregations: sum(KEY._col0:0._col0), avg(KEY._col0:0._col0), 
avg(DISTINCT KEY._col0:0._col0), max(KEY._col0:0._col0), 
min(KEY._col0:0._col0), std(KEY._col0:0._col0), stddev_samp(KEY._col0:0._col0), 
variance(KEY._col0:0._col0), var_samp(KEY._col0:0._col0), sum(DISTINCT 
KEY._col0:1._col0), count(DISTINCT KEY._col0:2._col0)
                 mode: complete
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10
-                Statistics: Num rows: 1 Data size: 240 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), _col1 (type: double), 
_col2 (type: double), UDFToDouble(_col3) (type: double), UDFToDouble(_col4) 
(type: double), _col5 (type: double), _col6 (type: double), _col7 (type: 
double), _col8 (type: double), _col9 (type: double), UDFToDouble(_col10) (type: 
double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10
-                  Statistics: Num rows: 1 Data size: 240 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 240 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 440 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: org.apache.hadoop.mapred.TextInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/subquery_in.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/subquery_in.q.out 
b/ql/src/test/results/clientpositive/spark/subquery_in.q.out
index 8923a4d..e40d6e4 100644
--- a/ql/src/test/results/clientpositive/spark/subquery_in.q.out
+++ b/ql/src/test/results/clientpositive/spark/subquery_in.q.out
@@ -354,10 +354,10 @@ STAGE PLANS:
                         aggregations: avg(_col0)
                         mode: hash
                         outputColumnNames: _col0
-                        Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 76 Basic stats: 
COMPLETE Column stats: NONE
                         Reduce Output Operator
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 76 Basic stats: 
COMPLETE Column stats: NONE
                           value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:int>)
         Reducer 5 
             Reduce Operator Tree:
@@ -366,19 +366,24 @@ STAGE PLANS:
                 mode: mergepartial
                 outputColumnNames: _col0
 <<<<<<< HEAD
+<<<<<<< HEAD
                 Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+=======
+                Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
+>>>>>>> 8efe6f7... HIVE-14367 : Estimated size for constant nulls is 0 
(Ashutosh Chauhan via Jesus Camacho Rodriguez)
                 Filter Operator
                   predicate: _col0 is not null (type: boolean)
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
                   Group By Operator
                     keys: _col0 (type: double)
                     mode: hash
                     outputColumnNames: _col0
-                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 76 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       key expressions: _col0 (type: double)
                       sort order: +
                       Map-reduce partition columns: _col0 (type: double)
+<<<<<<< HEAD
                       Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
 =======
                 Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
@@ -411,6 +416,9 @@ STAGE PLANS:
 >>>>>>> 382dc20... HIVE-15192 : Use Calcite to de-correlate and plan 
 >>>>>>> subqueries (Vineet Garg via Ashutosh Chauhan)
 =======
 >>>>>>> f63dc2d... HIVE-15458 : Fix semi-join conversion rule for subquery 
 >>>>>>> (Vineet Garg via Ashutosh Chauhan)
+=======
+                      Statistics: Num rows: 1 Data size: 76 Basic stats: 
COMPLETE Column stats: NONE
+>>>>>>> 8efe6f7... HIVE-14367 : Estimated size for constant nulls is 0 
(Ashutosh Chauhan via Jesus Camacho Rodriguez)
 
   Stage: Stage-0
     Fetch Operator

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out 
b/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out
index ef9dfc4..8ef0d60 100644
--- a/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out
+++ b/ql/src/test/results/clientpositive/spark/union_remove_6_subq.q.out
@@ -279,10 +279,10 @@ STAGE PLANS:
                     aggregations: avg(_col0)
                     mode: hash
                     outputColumnNames: _col0
-                    Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL 
Column stats: COMPLETE
+                    Statistics: Num rows: 1 Data size: 80 Basic stats: 
COMPLETE Column stats: COMPLETE
                     Reduce Output Operator
                       sort order: 
-                      Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: COMPLETE
+                      Statistics: Num rows: 1 Data size: 80 Basic stats: 
COMPLETE Column stats: COMPLETE
                       value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:bigint>)
         Reducer 3 
             Reduce Operator Tree:
@@ -309,10 +309,10 @@ STAGE PLANS:
                   aggregations: avg(_col0)
                   mode: hash
                   outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL 
Column stats: COMPLETE
+                  Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE 
Column stats: COMPLETE
                   Reduce Output Operator
                     sort order: 
-                    Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL 
Column stats: COMPLETE
+                    Statistics: Num rows: 1 Data size: 80 Basic stats: 
COMPLETE Column stats: COMPLETE
                     value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:bigint>)
 
   Stage: Stage-0

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/vectorization_0.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/vectorization_0.q.out 
b/ql/src/test/results/clientpositive/spark/vectorization_0.q.out
index 2c6c52d..efca361 100644
--- a/ql/src/test/results/clientpositive/spark/vectorization_0.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorization_0.q.out
@@ -236,10 +236,10 @@ STAGE PLANS:
                       aggregations: avg(ctinyint), variance(ctinyint), 
var_pop(ctinyint), var_samp(ctinyint), std(ctinyint), stddev(ctinyint), 
stddev_pop(ctinyint), stddev_samp(ctinyint)
                       mode: hash
                       outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7
-                      Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                      Statistics: Num rows: 1 Data size: 636 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         sort order: 
-                        Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 636 Basic stats: 
COMPLETE Column stats: NONE
                         value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:tinyint>), _col1 (type: 
struct<count:bigint,sum:double,variance:double>), _col2 (type: 
struct<count:bigint,sum:double,variance:double>), _col3 (type: 
struct<count:bigint,sum:double,variance:double>), _col4 (type: 
struct<count:bigint,sum:double,variance:double>), _col5 (type: 
struct<count:bigint,sum:double,variance:double>), _col6 (type: 
struct<count:bigint,sum:double,variance:double>), _col7 (type: 
struct<count:bigint,sum:double,variance:double>)
             Execution mode: vectorized
         Reducer 2 
@@ -248,11 +248,11 @@ STAGE PLANS:
                 aggregations: avg(VALUE._col0), variance(VALUE._col1), 
var_pop(VALUE._col2), var_samp(VALUE._col3), std(VALUE._col4), 
stddev(VALUE._col5), stddev_pop(VALUE._col6), stddev_samp(VALUE._col7)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 636 Basic stats: COMPLETE 
Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col0 (type: double)
                   sort order: +
-                  Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 636 Basic stats: COMPLETE 
Column stats: NONE
                   value expressions: _col1 (type: double), _col2 (type: 
double), _col3 (type: double), _col4 (type: double), _col5 (type: double), 
_col6 (type: double), _col7 (type: double)
         Reducer 3 
             Execution mode: vectorized
@@ -260,10 +260,10 @@ STAGE PLANS:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: double), VALUE._col0 
(type: double), VALUE._col1 (type: double), VALUE._col2 (type: double), 
VALUE._col3 (type: double), VALUE._col4 (type: double), VALUE._col5 (type: 
double), VALUE._col6 (type: double)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 636 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 636 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -542,10 +542,10 @@ STAGE PLANS:
                       aggregations: avg(cbigint), variance(cbigint), 
var_pop(cbigint), var_samp(cbigint), std(cbigint), stddev(cbigint), 
stddev_pop(cbigint), stddev_samp(cbigint)
                       mode: hash
                       outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7
-                      Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                      Statistics: Num rows: 1 Data size: 640 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         sort order: 
-                        Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 640 Basic stats: 
COMPLETE Column stats: NONE
                         value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:bigint>), _col1 (type: 
struct<count:bigint,sum:double,variance:double>), _col2 (type: 
struct<count:bigint,sum:double,variance:double>), _col3 (type: 
struct<count:bigint,sum:double,variance:double>), _col4 (type: 
struct<count:bigint,sum:double,variance:double>), _col5 (type: 
struct<count:bigint,sum:double,variance:double>), _col6 (type: 
struct<count:bigint,sum:double,variance:double>), _col7 (type: 
struct<count:bigint,sum:double,variance:double>)
             Execution mode: vectorized
         Reducer 2 
@@ -554,11 +554,11 @@ STAGE PLANS:
                 aggregations: avg(VALUE._col0), variance(VALUE._col1), 
var_pop(VALUE._col2), var_samp(VALUE._col3), std(VALUE._col4), 
stddev(VALUE._col5), stddev_pop(VALUE._col6), stddev_samp(VALUE._col7)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 640 Basic stats: COMPLETE 
Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col0 (type: double)
                   sort order: +
-                  Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 640 Basic stats: COMPLETE 
Column stats: NONE
                   value expressions: _col1 (type: double), _col2 (type: 
double), _col3 (type: double), _col4 (type: double), _col5 (type: double), 
_col6 (type: double), _col7 (type: double)
         Reducer 3 
             Execution mode: vectorized
@@ -566,10 +566,10 @@ STAGE PLANS:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: double), VALUE._col0 
(type: double), VALUE._col1 (type: double), VALUE._col2 (type: double), 
VALUE._col3 (type: double), VALUE._col4 (type: double), VALUE._col5 (type: 
double), VALUE._col6 (type: double)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 640 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 640 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -848,10 +848,10 @@ STAGE PLANS:
                       aggregations: avg(cfloat), variance(cfloat), 
var_pop(cfloat), var_samp(cfloat), std(cfloat), stddev(cfloat), 
stddev_pop(cfloat), stddev_samp(cfloat)
                       mode: hash
                       outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7
-                      Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                      Statistics: Num rows: 1 Data size: 636 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         sort order: 
-                        Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 636 Basic stats: 
COMPLETE Column stats: NONE
                         value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:float>), _col1 (type: 
struct<count:bigint,sum:double,variance:double>), _col2 (type: 
struct<count:bigint,sum:double,variance:double>), _col3 (type: 
struct<count:bigint,sum:double,variance:double>), _col4 (type: 
struct<count:bigint,sum:double,variance:double>), _col5 (type: 
struct<count:bigint,sum:double,variance:double>), _col6 (type: 
struct<count:bigint,sum:double,variance:double>), _col7 (type: 
struct<count:bigint,sum:double,variance:double>)
             Execution mode: vectorized
         Reducer 2 
@@ -860,11 +860,11 @@ STAGE PLANS:
                 aggregations: avg(VALUE._col0), variance(VALUE._col1), 
var_pop(VALUE._col2), var_samp(VALUE._col3), std(VALUE._col4), 
stddev(VALUE._col5), stddev_pop(VALUE._col6), stddev_samp(VALUE._col7)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 636 Basic stats: COMPLETE 
Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col0 (type: double)
                   sort order: +
-                  Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 636 Basic stats: COMPLETE 
Column stats: NONE
                   value expressions: _col1 (type: double), _col2 (type: 
double), _col3 (type: double), _col4 (type: double), _col5 (type: double), 
_col6 (type: double), _col7 (type: double)
         Reducer 3 
             Execution mode: vectorized
@@ -872,10 +872,10 @@ STAGE PLANS:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: double), VALUE._col0 
(type: double), VALUE._col1 (type: double), VALUE._col2 (type: double), 
VALUE._col3 (type: double), VALUE._col4 (type: double), VALUE._col5 (type: 
double), VALUE._col6 (type: double)
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 636 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 636 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -998,10 +998,10 @@ STAGE PLANS:
                         aggregations: avg(cbigint), stddev_pop(cbigint), 
var_samp(cbigint), count(), sum(cfloat), min(ctinyint)
                         mode: hash
                         outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5
-                        Statistics: Num rows: 1 Data size: 20 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 260 Basic stats: 
COMPLETE Column stats: NONE
                         Reduce Output Operator
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 20 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 260 Basic stats: 
COMPLETE Column stats: NONE
                           value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:bigint>), _col1 (type: 
struct<count:bigint,sum:double,variance:double>), _col2 (type: 
struct<count:bigint,sum:double,variance:double>), _col3 (type: bigint), _col4 
(type: double), _col5 (type: tinyint)
             Execution mode: vectorized
         Reducer 2 
@@ -1010,14 +1010,14 @@ STAGE PLANS:
                 aggregations: avg(VALUE._col0), stddev_pop(VALUE._col1), 
var_samp(VALUE._col2), count(VALUE._col3), sum(VALUE._col4), min(VALUE._col5)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                Statistics: Num rows: 1 Data size: 44 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), (- _col0) (type: double), 
(-6432.0 + _col0) (type: double), _col1 (type: double), (- (-6432.0 + _col0)) 
(type: double), ((- (-6432.0 + _col0)) + (-6432.0 + _col0)) (type: double), 
_col2 (type: double), (- (-6432.0 + _col0)) (type: double), (-6432.0 + (- 
(-6432.0 + _col0))) (type: double), (- (-6432.0 + _col0)) (type: double), ((- 
(-6432.0 + _col0)) / (- (-6432.0 + _col0))) (type: double), _col3 (type: 
bigint), _col4 (type: double), (_col2 % _col1) (type: double), (- _col2) (type: 
double), ((- (-6432.0 + _col0)) * (- _col0)) (type: double), _col5 (type: 
tinyint), (- _col5) (type: tinyint)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, 
_col16, _col17
-                  Statistics: Num rows: 1 Data size: 44 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 260 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 44 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 260 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/vectorization_pushdown.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/vectorization_pushdown.q.out 
b/ql/src/test/results/clientpositive/spark/vectorization_pushdown.q.out
index ef7df76..59a3be0 100644
--- a/ql/src/test/results/clientpositive/spark/vectorization_pushdown.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorization_pushdown.q.out
@@ -31,10 +31,10 @@ STAGE PLANS:
                         aggregations: avg(cbigint)
                         mode: hash
                         outputColumnNames: _col0
-                        Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 80 Basic stats: 
COMPLETE Column stats: NONE
                         Reduce Output Operator
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 80 Basic stats: 
COMPLETE Column stats: NONE
                           value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:bigint>)
             Execution mode: vectorized
         Reducer 2 
@@ -43,10 +43,10 @@ STAGE PLANS:
                 aggregations: avg(VALUE._col0)
                 mode: mergepartial
                 outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out 
b/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out
index a969029..20524f1 100644
--- a/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorization_short_regress.q.out
@@ -134,10 +134,10 @@ STAGE PLANS:
                         aggregations: avg(cint), sum(cdouble), 
stddev_pop(cint), stddev_samp(csmallint), var_samp(cint), avg(cfloat), 
stddev_samp(cint), min(ctinyint), count(csmallint)
                         mode: hash
                         outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8
-                        Statistics: Num rows: 1 Data size: 20 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 492 Basic stats: 
COMPLETE Column stats: NONE
                         Reduce Output Operator
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 20 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 492 Basic stats: 
COMPLETE Column stats: NONE
                           value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:int>), _col1 (type: double), _col2 (type: 
struct<count:bigint,sum:double,variance:double>), _col3 (type: 
struct<count:bigint,sum:double,variance:double>), _col4 (type: 
struct<count:bigint,sum:double,variance:double>), _col5 (type: 
struct<count:bigint,sum:double,input:float>), _col6 (type: 
struct<count:bigint,sum:double,variance:double>), _col7 (type: tinyint), _col8 
(type: bigint)
             Execution mode: vectorized
         Reducer 2 
@@ -146,14 +146,14 @@ STAGE PLANS:
                 aggregations: avg(VALUE._col0), sum(VALUE._col1), 
stddev_pop(VALUE._col2), stddev_samp(VALUE._col3), var_samp(VALUE._col4), 
avg(VALUE._col5), stddev_samp(VALUE._col6), min(VALUE._col7), count(VALUE._col8)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                Statistics: Num rows: 1 Data size: 68 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 492 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), (_col0 + -3728.0) (type: 
double), (- (_col0 + -3728.0)) (type: double), (- (- (_col0 + -3728.0))) (type: 
double), ((- (- (_col0 + -3728.0))) * (_col0 + -3728.0)) (type: double), _col1 
(type: double), (- _col0) (type: double), _col2 (type: double), (((- (- (_col0 
+ -3728.0))) * (_col0 + -3728.0)) * (- (- (_col0 + -3728.0)))) (type: double), 
_col3 (type: double), (- _col2) (type: double), (_col2 - (- (- (_col0 + 
-3728.0)))) (type: double), ((_col2 - (- (- (_col0 + -3728.0)))) * _col2) 
(type: double), _col4 (type: double), _col5 (type: double), (10.175 - _col4) 
(type: double), (- (10.175 - _col4)) (type: double), ((- _col2) / -563.0) 
(type: double), _col6 (type: double), (- ((- _col2) / -563.0)) (type: double), 
(_col0 / _col1) (type: double), _col7 (type: tinyint), _col8 (type: bigint), 
(UDFToDouble(_col7) / ((- _col2) / -563.0)) (type: double), (- (_col0 / _col1)) 
(type: double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, 
_col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24
-                  Statistics: Num rows: 1 Data size: 68 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 492 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 68 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 492 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -336,10 +336,10 @@ STAGE PLANS:
                         aggregations: max(cint), var_pop(cbigint), 
stddev_pop(csmallint), max(cdouble), avg(ctinyint), min(cint), min(cdouble), 
stddev_samp(csmallint), var_samp(cint)
                         mode: hash
                         outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7, _col8
-                        Statistics: Num rows: 1 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 420 Basic stats: 
COMPLETE Column stats: NONE
                         Reduce Output Operator
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 420 Basic stats: 
COMPLETE Column stats: NONE
                           value expressions: _col0 (type: int), _col1 (type: 
struct<count:bigint,sum:double,variance:double>), _col2 (type: 
struct<count:bigint,sum:double,variance:double>), _col3 (type: double), _col4 
(type: struct<count:bigint,sum:double,input:tinyint>), _col5 (type: int), _col6 
(type: double), _col7 (type: struct<count:bigint,sum:double,variance:double>), 
_col8 (type: struct<count:bigint,sum:double,variance:double>)
             Execution mode: vectorized
         Reducer 2 
@@ -348,14 +348,14 @@ STAGE PLANS:
                 aggregations: max(VALUE._col0), var_pop(VALUE._col1), 
stddev_pop(VALUE._col2), max(VALUE._col3), avg(VALUE._col4), min(VALUE._col5), 
min(VALUE._col6), stddev_samp(VALUE._col7), var_samp(VALUE._col8)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8
-                Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 420 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: int), (UDFToDouble(_col0) / 
-3728.0) (type: double), (_col0 * -3728) (type: int), _col1 (type: double), (- 
(_col0 * -3728)) (type: int), _col2 (type: double), (-563 % (_col0 * -3728)) 
(type: int), (_col1 / _col2) (type: double), (- _col2) (type: double), _col3 
(type: double), _col4 (type: double), (_col2 - 10.175) (type: double), _col5 
(type: int), (UDFToDouble((_col0 * -3728)) % (_col2 - 10.175)) (type: double), 
(- _col3) (type: double), _col6 (type: double), (_col3 % -26.28) (type: 
double), _col7 (type: double), (- (UDFToDouble(_col0) / -3728.0)) (type: 
double), ((- (_col0 * -3728)) % (-563 % (_col0 * -3728))) (type: int), 
((UDFToDouble(_col0) / -3728.0) - _col4) (type: double), (- (_col0 * -3728)) 
(type: int), _col8 (type: double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, 
_col16, _col17, _col18, _col19, _col20, _col21, _col22
-                  Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 420 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 64 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 420 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -529,10 +529,10 @@ STAGE PLANS:
                         aggregations: var_pop(cbigint), count(), 
max(ctinyint), stddev_pop(csmallint), max(cint), stddev_samp(cdouble), 
count(ctinyint), avg(ctinyint)
                         mode: hash
                         outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7
-                        Statistics: Num rows: 1 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 340 Basic stats: 
COMPLETE Column stats: NONE
                         Reduce Output Operator
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 24 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 340 Basic stats: 
COMPLETE Column stats: NONE
                           value expressions: _col0 (type: 
struct<count:bigint,sum:double,variance:double>), _col1 (type: bigint), _col2 
(type: tinyint), _col3 (type: struct<count:bigint,sum:double,variance:double>), 
_col4 (type: int), _col5 (type: 
struct<count:bigint,sum:double,variance:double>), _col6 (type: bigint), _col7 
(type: struct<count:bigint,sum:double,input:tinyint>)
             Execution mode: vectorized
         Reducer 2 
@@ -541,14 +541,14 @@ STAGE PLANS:
                 aggregations: var_pop(VALUE._col0), count(VALUE._col1), 
max(VALUE._col2), stddev_pop(VALUE._col3), max(VALUE._col4), 
stddev_samp(VALUE._col5), count(VALUE._col6), avg(VALUE._col7)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                Statistics: Num rows: 1 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 340 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), (- _col0) (type: double), 
(_col0 - (- _col0)) (type: double), _col1 (type: bigint), (UDFToDouble(_col1) % 
79.553) (type: double), _col2 (type: tinyint), (UDFToDouble(_col1) - (- _col0)) 
(type: double), (- (- _col0)) (type: double), (-1.0 % (- _col0)) (type: 
double), _col1 (type: bigint), (- _col1) (type: bigint), _col3 (type: double), 
(- (- (- _col0))) (type: double), (762 * (- _col1)) (type: bigint), _col4 
(type: int), (UDFToLong(_col2) + (762 * (- _col1))) (type: bigint), ((- _col0) 
+ UDFToDouble(_col4)) (type: double), _col5 (type: double), ((- _col1) % _col1) 
(type: bigint), _col6 (type: bigint), _col7 (type: double), (-3728 % 
(UDFToLong(_col2) + (762 * (- _col1)))) (type: bigint)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, 
_col16, _col17, _col18, _col19, _col20, _col21
-                  Statistics: Num rows: 1 Data size: 56 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 340 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 56 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 340 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -701,10 +701,10 @@ STAGE PLANS:
                         aggregations: avg(ctinyint), max(cbigint), 
stddev_samp(cint), var_pop(cint), var_pop(cbigint), max(cfloat)
                         mode: hash
                         outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5
-                        Statistics: Num rows: 1 Data size: 12 Basic stats: 
COMPLETE Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 328 Basic stats: 
COMPLETE Column stats: NONE
                         Reduce Output Operator
                           sort order: 
-                          Statistics: Num rows: 1 Data size: 12 Basic stats: 
COMPLETE Column stats: NONE
+                          Statistics: Num rows: 1 Data size: 328 Basic stats: 
COMPLETE Column stats: NONE
                           value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:tinyint>), _col1 (type: bigint), _col2 
(type: struct<count:bigint,sum:double,variance:double>), _col3 (type: 
struct<count:bigint,sum:double,variance:double>), _col4 (type: 
struct<count:bigint,sum:double,variance:double>), _col5 (type: float)
             Execution mode: vectorized
         Reducer 2 
@@ -713,14 +713,14 @@ STAGE PLANS:
                 aggregations: avg(VALUE._col0), max(VALUE._col1), 
stddev_samp(VALUE._col2), var_pop(VALUE._col3), var_pop(VALUE._col4), 
max(VALUE._col5)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5
-                Statistics: Num rows: 1 Data size: 44 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 328 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: _col0 (type: double), (_col0 + 6981.0) (type: 
double), ((_col0 + 6981.0) + _col0) (type: double), _col1 (type: bigint), 
(((_col0 + 6981.0) + _col0) / _col0) (type: double), (- (_col0 + 6981.0)) 
(type: double), _col2 (type: double), (_col0 % (- (_col0 + 6981.0))) (type: 
double), _col3 (type: double), _col4 (type: double), (- _col1) (type: bigint), 
(UDFToDouble((- _col1)) / _col2) (type: double), _col5 (type: float), (_col4 * 
-26.28) (type: double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13
-                  Statistics: Num rows: 1 Data size: 44 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 328 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 44 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 328 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/vectorized_mapjoin.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/vectorized_mapjoin.q.out 
b/ql/src/test/results/clientpositive/spark/vectorized_mapjoin.q.out
index cf5f55a..d605185 100644
--- a/ql/src/test/results/clientpositive/spark/vectorized_mapjoin.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorized_mapjoin.q.out
@@ -72,10 +72,10 @@ STAGE PLANS:
                             aggregations: count(_col0), max(_col1), 
min(_col0), avg(_col2)
                             mode: hash
                             outputColumnNames: _col0, _col1, _col2, _col3
-                            Statistics: Num rows: 1 Data size: 16 Basic stats: 
COMPLETE Column stats: NONE
+                            Statistics: Num rows: 1 Data size: 92 Basic stats: 
COMPLETE Column stats: NONE
                             Reduce Output Operator
                               sort order: 
-                              Statistics: Num rows: 1 Data size: 16 Basic 
stats: COMPLETE Column stats: NONE
+                              Statistics: Num rows: 1 Data size: 92 Basic 
stats: COMPLETE Column stats: NONE
                               value expressions: _col0 (type: bigint), _col1 
(type: int), _col2 (type: int), _col3 (type: 
struct<count:bigint,sum:double,input:int>)
             Execution mode: vectorized
             Local Work:
@@ -86,10 +86,10 @@ STAGE PLANS:
                 aggregations: count(VALUE._col0), max(VALUE._col1), 
min(VALUE._col2), avg(VALUE._col3)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3
-                Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/vectorized_shufflejoin.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/vectorized_shufflejoin.q.out 
b/ql/src/test/results/clientpositive/spark/vectorized_shufflejoin.q.out
index 21e47b2..97f12d4 100644
--- a/ql/src/test/results/clientpositive/spark/vectorized_shufflejoin.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorized_shufflejoin.q.out
@@ -73,10 +73,10 @@ STAGE PLANS:
                     aggregations: count(_col0), max(_col1), min(_col0), 
avg(_col2)
                     mode: hash
                     outputColumnNames: _col0, _col1, _col2, _col3
-                    Statistics: Num rows: 1 Data size: 16 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 92 Basic stats: 
COMPLETE Column stats: NONE
                     Reduce Output Operator
                       sort order: 
-                      Statistics: Num rows: 1 Data size: 16 Basic stats: 
COMPLETE Column stats: NONE
+                      Statistics: Num rows: 1 Data size: 92 Basic stats: 
COMPLETE Column stats: NONE
                       value expressions: _col0 (type: bigint), _col1 (type: 
int), _col2 (type: int), _col3 (type: struct<count:bigint,sum:double,input:int>)
         Reducer 3 
             Reduce Operator Tree:
@@ -84,11 +84,11 @@ STAGE PLANS:
                 aggregations: count(VALUE._col0), max(VALUE._col1), 
min(VALUE._col2), avg(VALUE._col3)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3
-                Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE 
Column stats: NONE
                 Reduce Output Operator
                   key expressions: _col0 (type: bigint)
                   sort order: +
-                  Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE 
Column stats: NONE
                   value expressions: _col1 (type: int), _col2 (type: int), 
_col3 (type: double)
         Reducer 4 
             Execution mode: vectorized
@@ -96,10 +96,10 @@ STAGE PLANS:
               Select Operator
                 expressions: KEY.reducesinkkey0 (type: bigint), VALUE._col0 
(type: int), VALUE._col1 (type: int), VALUE._col2 (type: double)
                 outputColumnNames: _col0, _col1, _col2, _col3
-                Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE 
Column stats: NONE
                 File Output Operator
                   compressed: false
-                  Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE 
Column stats: NONE
                   table:
                       input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                       output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/spark/vectorized_timestamp_funcs.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/spark/vectorized_timestamp_funcs.q.out 
b/ql/src/test/results/clientpositive/spark/vectorized_timestamp_funcs.q.out
index c43e769..9948112 100644
--- a/ql/src/test/results/clientpositive/spark/vectorized_timestamp_funcs.q.out
+++ b/ql/src/test/results/clientpositive/spark/vectorized_timestamp_funcs.q.out
@@ -805,10 +805,10 @@ STAGE PLANS:
                       aggregations: avg(ctimestamp1), variance(ctimestamp1), 
var_pop(ctimestamp1), var_samp(ctimestamp1), std(ctimestamp1), 
stddev(ctimestamp1), stddev_pop(ctimestamp1), stddev_samp(ctimestamp1)
                       mode: hash
                       outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7
-                      Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                      Statistics: Num rows: 1 Data size: 672 Basic stats: 
COMPLETE Column stats: NONE
                       Reduce Output Operator
                         sort order: 
-                        Statistics: Num rows: 1 Data size: 0 Basic stats: 
PARTIAL Column stats: NONE
+                        Statistics: Num rows: 1 Data size: 672 Basic stats: 
COMPLETE Column stats: NONE
                         value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:timestamp>), _col1 (type: 
struct<count:bigint,sum:double,variance:double>), _col2 (type: 
struct<count:bigint,sum:double,variance:double>), _col3 (type: 
struct<count:bigint,sum:double,variance:double>), _col4 (type: 
struct<count:bigint,sum:double,variance:double>), _col5 (type: 
struct<count:bigint,sum:double,variance:double>), _col6 (type: 
struct<count:bigint,sum:double,variance:double>), _col7 (type: 
struct<count:bigint,sum:double,variance:double>)
             Execution mode: vectorized
         Reducer 2 
@@ -817,14 +817,14 @@ STAGE PLANS:
                 aggregations: avg(VALUE._col0), variance(VALUE._col1), 
var_pop(VALUE._col2), var_samp(VALUE._col3), std(VALUE._col4), 
stddev(VALUE._col5), stddev_pop(VALUE._col6), stddev_samp(VALUE._col7)
                 mode: mergepartial
                 outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 1 Data size: 672 Basic stats: COMPLETE 
Column stats: NONE
                 Select Operator
                   expressions: round(_col0, 0) (type: double), _col1 BETWEEN 
8.97077295279421E19 AND 8.97077295279422E19 (type: boolean), _col2 BETWEEN 
8.97077295279421E19 AND 8.97077295279422E19 (type: boolean), _col3 BETWEEN 
9.20684592523616E19 AND 9.20684592523617E19 (type: boolean), round(_col4, 3) 
(type: double), round(_col5, 3) (type: double), round(_col6, 3) (type: double), 
round(_col7, 3) (type: double)
                   outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, 
_col6, _col7
-                  Statistics: Num rows: 1 Data size: 64 Basic stats: COMPLETE 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 672 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
-                    Statistics: Num rows: 1 Data size: 64 Basic stats: 
COMPLETE Column stats: NONE
+                    Statistics: Num rows: 1 Data size: 672 Basic stats: 
COMPLETE Column stats: NONE
                     table:
                         input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                         output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/subquery_in.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/subquery_in.q.out 
b/ql/src/test/results/clientpositive/subquery_in.q.out
index caf2470..e2bfae0 100644
--- a/ql/src/test/results/clientpositive/subquery_in.q.out
+++ b/ql/src/test/results/clientpositive/subquery_in.q.out
@@ -284,7 +284,7 @@ STAGE PLANS:
                   aggregations: avg(_col0)
                   mode: hash
                   outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
                     table:
@@ -298,22 +298,22 @@ STAGE PLANS:
           TableScan
             Reduce Output Operator
               sort order: 
-              Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column 
stats: NONE
+              Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
               value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:int>)
       Reduce Operator Tree:
         Group By Operator
           aggregations: avg(VALUE._col0)
           mode: mergepartial
           outputColumnNames: _col0
-          Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
+          Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column 
stats: NONE
           Filter Operator
             predicate: _col0 is not null (type: boolean)
-            Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
+            Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column 
stats: NONE
             Group By Operator
               keys: _col0 (type: double)
               mode: hash
               outputColumnNames: _col0
-              Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
               File Output Operator
                 compressed: false
                 table:
@@ -345,7 +345,7 @@ STAGE PLANS:
               key expressions: _col0 (type: double)
               sort order: +
               Map-reduce partition columns: _col0 (type: double)
-              Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
       Reduce Operator Tree:
         Join Operator
           condition map:

http://git-wip-us.apache.org/repos/asf/hive/blob/47dbc005/ql/src/test/results/clientpositive/subquery_notin.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/subquery_notin.q.out 
b/ql/src/test/results/clientpositive/subquery_notin.q.out
index 4398c69..30ead87 100644
--- a/ql/src/test/results/clientpositive/subquery_notin.q.out
+++ b/ql/src/test/results/clientpositive/subquery_notin.q.out
@@ -713,7 +713,7 @@ STAGE PLANS:
                   aggregations: avg(_col0)
                   mode: hash
                   outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
                     table:
@@ -727,19 +727,19 @@ STAGE PLANS:
           TableScan
             Reduce Output Operator
               sort order: 
-              Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column 
stats: NONE
+              Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
               value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:int>)
       Reduce Operator Tree:
         Group By Operator
           aggregations: avg(VALUE._col0)
           mode: mergepartial
           outputColumnNames: _col0
-          Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
+          Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column 
stats: NONE
           Group By Operator
             aggregations: count(), count(_col0)
             mode: complete
             outputColumnNames: _col0, _col1
-            Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column 
stats: NONE
+            Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column 
stats: NONE
             File Output Operator
               compressed: false
               table:
@@ -764,7 +764,7 @@ STAGE PLANS:
           TableScan
             Reduce Output Operator
               sort order: 
-              Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE 
Column stats: NONE
               value expressions: _col0 (type: bigint), _col1 (type: bigint)
       Reduce Operator Tree:
         Join Operator
@@ -774,7 +774,7 @@ STAGE PLANS:
             0 
             1 
           outputColumnNames: _col0, _col1, _col2, _col3
-          Statistics: Num rows: 26 Data size: 3589 Basic stats: COMPLETE 
Column stats: NONE
+          Statistics: Num rows: 26 Data size: 5357 Basic stats: COMPLETE 
Column stats: NONE
           File Output Operator
             compressed: false
             table:
@@ -790,14 +790,14 @@ STAGE PLANS:
               key expressions: UDFToDouble(_col1) (type: double)
               sort order: +
               Map-reduce partition columns: UDFToDouble(_col1) (type: double)
-              Statistics: Num rows: 26 Data size: 3589 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 26 Data size: 5357 Basic stats: COMPLETE 
Column stats: NONE
               value expressions: _col0 (type: string), _col1 (type: int), 
_col2 (type: bigint), _col3 (type: bigint)
           TableScan
             Reduce Output Operator
               key expressions: _col0 (type: double)
               sort order: +
               Map-reduce partition columns: _col0 (type: double)
-              Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
               value expressions: _col1 (type: boolean)
       Reduce Operator Tree:
         Join Operator
@@ -807,17 +807,17 @@ STAGE PLANS:
             0 UDFToDouble(_col1) (type: double)
             1 _col0 (type: double)
           outputColumnNames: _col0, _col1, _col2, _col3, _col5
-          Statistics: Num rows: 28 Data size: 3947 Basic stats: COMPLETE 
Column stats: NONE
+          Statistics: Num rows: 28 Data size: 5892 Basic stats: COMPLETE 
Column stats: NONE
           Filter Operator
             predicate: ((_col2 = 0) or (_col5 is null and _col1 is not null 
and (_col3 >= _col2))) (type: boolean)
-            Statistics: Num rows: 18 Data size: 2537 Basic stats: COMPLETE 
Column stats: NONE
+            Statistics: Num rows: 18 Data size: 3787 Basic stats: COMPLETE 
Column stats: NONE
             Select Operator
               expressions: _col0 (type: string), _col1 (type: int)
               outputColumnNames: _col0, _col1
-              Statistics: Num rows: 18 Data size: 2537 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 18 Data size: 3787 Basic stats: COMPLETE 
Column stats: NONE
               File Output Operator
                 compressed: false
-                Statistics: Num rows: 18 Data size: 2537 Basic stats: COMPLETE 
Column stats: NONE
+                Statistics: Num rows: 18 Data size: 3787 Basic stats: COMPLETE 
Column stats: NONE
                 table:
                     input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
                     output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
@@ -872,7 +872,7 @@ STAGE PLANS:
                   aggregations: avg(_col0)
                   mode: hash
                   outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL 
Column stats: NONE
+                  Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
                   File Output Operator
                     compressed: false
                     table:
@@ -886,19 +886,19 @@ STAGE PLANS:
           TableScan
             Reduce Output Operator
               sort order: 
-              Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column 
stats: NONE
+              Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
               value expressions: _col0 (type: 
struct<count:bigint,sum:double,input:int>)
       Reduce Operator Tree:
         Group By Operator
           aggregations: avg(VALUE._col0)
           mode: mergepartial
           outputColumnNames: _col0
-          Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
+          Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column 
stats: NONE
           Group By Operator
             keys: _col0 (type: double)
             mode: hash
             outputColumnNames: _col0
-            Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
+            Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column 
stats: NONE
             File Output Operator
               compressed: false
               table:
@@ -914,17 +914,17 @@ STAGE PLANS:
               key expressions: _col0 (type: double)
               sort order: +
               Map-reduce partition columns: _col0 (type: double)
-              Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: NONE
+              Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE 
Column stats: NONE
       Reduce Operator Tree:
         Group By Operator
           keys: KEY._col0 (type: double)
           mode: mergepartial
           outputColumnNames: _col0
-          Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
+          Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column 
stats: NONE
           Select Operator
             expressions: _col0 (type: double), true (type: boolean)
             outputColumnNames: _col0, _col1
-            Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column 
stats: NONE
+            Statistics: Num rows: 1 Data size: 76 Basic stats: COMPLETE Column 
stats: NONE
             File Output Operator
               compressed: false
               table:

Reply via email to