http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/tez/explainanalyze_4.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/explainanalyze_4.q.out b/ql/src/test/results/clientpositive/tez/explainanalyze_4.q.out index 3426d19..2145a8a 100644 --- a/ql/src/test/results/clientpositive/tez/explainanalyze_4.q.out +++ b/ql/src/test/results/clientpositive/tez/explainanalyze_4.q.out @@ -16,8 +16,7 @@ order by a.cint POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypesorc #### A masked pattern was here #### -PREHOOK: query: -- First try with regular mergejoin -explain analyze +PREHOOK: query: explain analyze select * from alltypesorc a join alltypesorc b on a.cint = b.cint @@ -25,8 +24,7 @@ where a.cint between 1000000 and 3000000 and b.cbigint is not null order by a.cint PREHOOK: type: QUERY -POSTHOOK: query: -- First try with regular mergejoin -explain analyze +POSTHOOK: query: explain analyze select * from alltypesorc a join alltypesorc b on a.cint = b.cint @@ -46,30 +44,30 @@ Stage-0 Stage-1 Reducer 3 File Output Operator [FS_12] - Select Operator [SEL_11] (rows=7286/10 width=620) + Select Operator [SEL_11] (rows=1501/10 width=215) Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8","_col9","_col10","_col11","_col12","_col13","_col14","_col15","_col16","_col17","_col18","_col19","_col20","_col21","_col22","_col23"] <-Reducer 2 [SIMPLE_EDGE] SHUFFLE [RS_10] - Merge Join Operator [MERGEJOIN_17] (rows=7286/10 width=620) + Merge Join Operator [MERGEJOIN_17] (rows=1501/10 width=215) Conds:RS_6._col2=RS_7._col2(Inner),Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8","_col9","_col10","_col11","_col12","_col13","_col14","_col15","_col16","_col17","_col18","_col19","_col20","_col21","_col22","_col23"] <-Map 1 [SIMPLE_EDGE] SHUFFLE [RS_6] PartitionCols:_col2 - Select Operator [SEL_2] (rows=6144/10 width=251) + Select Operator [SEL_2] (rows=1365/10 width=215) Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8","_col9","_col10","_col11"] - Filter Operator [FIL_15] (rows=6144/10 width=251) + Filter Operator [FIL_15] (rows=1365/10 width=215) predicate:cint BETWEEN 1000000 AND 3000000 - TableScan [TS_0] (rows=12288/12288 width=251) - default@alltypesorc,a,Tbl:COMPLETE,Col:COMPLETE,Output:["ctinyint","csmallint","cint","cbigint","cfloat","cdouble","cstring1","cstring2","ctimestamp1","ctimestamp2","cboolean1","cboolean2"] + TableScan [TS_0] (rows=12288/12288 width=215) + default@alltypesorc,a,Tbl:COMPLETE,Col:NONE,Output:["ctinyint","csmallint","cint","cbigint","cfloat","cdouble","cstring1","cstring2","ctimestamp1","ctimestamp2","cboolean1","cboolean2"] <-Map 4 [SIMPLE_EDGE] SHUFFLE [RS_7] PartitionCols:_col2 - Select Operator [SEL_5] (rows=3424/10 width=251) + Select Operator [SEL_5] (rows=1365/10 width=215) Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8","_col9","_col10","_col11"] - Filter Operator [FIL_16] (rows=3424/10 width=251) - predicate:(cint is not null and cbigint is not null and cint BETWEEN 1000000 AND 3000000) - TableScan [TS_3] (rows=12288/12288 width=251) - default@alltypesorc,b,Tbl:COMPLETE,Col:COMPLETE,Output:["ctinyint","csmallint","cint","cbigint","cfloat","cdouble","cstring1","cstring2","ctimestamp1","ctimestamp2","cboolean1","cboolean2"] + Filter Operator [FIL_16] (rows=1365/10 width=215) + predicate:(cint BETWEEN 1000000 AND 3000000 and cbigint is not null) + TableScan [TS_3] (rows=12288/12288 width=215) + default@alltypesorc,b,Tbl:COMPLETE,Col:NONE,Output:["ctinyint","csmallint","cint","cbigint","cfloat","cdouble","cstring1","cstring2","ctimestamp1","ctimestamp2","cboolean1","cboolean2"] PREHOOK: query: select * @@ -140,33 +138,31 @@ Stage-0 limit:-1 Stage-1 Reducer 3 - File Output Operator [FS_14] - Group By Operator [GBY_12] (rows=1/1 width=8) - Output:["_col0"],aggregations:["count(VALUE._col0)"] + File Output Operator [FS_13] + Group By Operator [GBY_11] (rows=1/1 width=8) + Output:["_col0"],aggregations:["count()"] <-Reducer 2 [SIMPLE_EDGE] - SHUFFLE [RS_11] - Group By Operator [GBY_10] (rows=1/1 width=8) - Output:["_col0"],aggregations:["count()"] - Merge Join Operator [MERGEJOIN_19] (rows=7286/10 width=8) - Conds:RS_6._col0=RS_7._col0(Inner) - <-Map 1 [SIMPLE_EDGE] - SHUFFLE [RS_6] - PartitionCols:_col0 - Select Operator [SEL_2] (rows=6144/10 width=2) - Output:["_col0"] - Filter Operator [FIL_17] (rows=6144/10 width=2) - predicate:cint BETWEEN 1000000 AND 3000000 - TableScan [TS_0] (rows=12288/12288 width=2) - default@alltypesorc,a,Tbl:COMPLETE,Col:COMPLETE,Output:["cint"] - <-Map 4 [SIMPLE_EDGE] - SHUFFLE [RS_7] - PartitionCols:_col0 - Select Operator [SEL_5] (rows=3424/10 width=8) - Output:["_col0"] - Filter Operator [FIL_18] (rows=3424/10 width=8) - predicate:(cint is not null and cbigint is not null and cint BETWEEN 1000000 AND 3000000) - TableScan [TS_3] (rows=12288/12288 width=8) - default@alltypesorc,b,Tbl:COMPLETE,Col:COMPLETE,Output:["cint","cbigint"] + SHUFFLE [RS_10] + Merge Join Operator [MERGEJOIN_18] (rows=1501/10 width=215) + Conds:RS_6._col0=RS_7._col0(Inner) + <-Map 1 [SIMPLE_EDGE] + SHUFFLE [RS_6] + PartitionCols:_col0 + Select Operator [SEL_2] (rows=1365/10 width=215) + Output:["_col0"] + Filter Operator [FIL_16] (rows=1365/10 width=215) + predicate:cint BETWEEN 1000000 AND 3000000 + TableScan [TS_0] (rows=12288/12288 width=215) + default@alltypesorc,a,Tbl:COMPLETE,Col:NONE,Output:["cint"] + <-Map 4 [SIMPLE_EDGE] + SHUFFLE [RS_7] + PartitionCols:_col0 + Select Operator [SEL_5] (rows=1365/10 width=215) + Output:["_col0"] + Filter Operator [FIL_17] (rows=1365/10 width=215) + predicate:(cint BETWEEN 1000000 AND 3000000 and cbigint is not null) + TableScan [TS_3] (rows=12288/12288 width=215) + default@alltypesorc,b,Tbl:COMPLETE,Col:NONE,Output:["cint","cbigint"] PREHOOK: query: select count(*) @@ -235,38 +231,36 @@ Stage-0 limit:-1 Stage-1 Reducer 4 - File Output Operator [FS_16] - Select Operator [SEL_15] (rows=2765/5 width=12) + File Output Operator [FS_15] + Select Operator [SEL_14] (rows=750/5 width=215) Output:["_col0","_col1"] <-Reducer 3 [SIMPLE_EDGE] - SHUFFLE [RS_14] - Group By Operator [GBY_12] (rows=2765/5 width=12) - Output:["_col0","_col1"],aggregations:["count(VALUE._col0)"],keys:KEY._col0 + SHUFFLE [RS_13] + Group By Operator [GBY_11] (rows=750/5 width=215) + Output:["_col0","_col1"],aggregations:["count()"],keys:KEY._col0 <-Reducer 2 [SIMPLE_EDGE] - SHUFFLE [RS_11] + SHUFFLE [RS_10] PartitionCols:_col0 - Group By Operator [GBY_10] (rows=2765/5 width=12) - Output:["_col0","_col1"],aggregations:["count()"],keys:_col0 - Merge Join Operator [MERGEJOIN_21] (rows=7286/10 width=4) - Conds:RS_6._col1=RS_7._col0(Inner),Output:["_col0"] - <-Map 1 [SIMPLE_EDGE] - SHUFFLE [RS_6] - PartitionCols:_col1 - Select Operator [SEL_2] (rows=6144/10 width=5) - Output:["_col0","_col1"] - Filter Operator [FIL_19] (rows=6144/10 width=5) - predicate:cint BETWEEN 1000000 AND 3000000 - TableScan [TS_0] (rows=12288/12288 width=5) - default@alltypesorc,a,Tbl:COMPLETE,Col:COMPLETE,Output:["csmallint","cint"] - <-Map 5 [SIMPLE_EDGE] - SHUFFLE [RS_7] - PartitionCols:_col0 - Select Operator [SEL_5] (rows=3424/10 width=8) - Output:["_col0"] - Filter Operator [FIL_20] (rows=3424/10 width=8) - predicate:(cint is not null and cbigint is not null and cint BETWEEN 1000000 AND 3000000) - TableScan [TS_3] (rows=12288/12288 width=8) - default@alltypesorc,b,Tbl:COMPLETE,Col:COMPLETE,Output:["cint","cbigint"] + Merge Join Operator [MERGEJOIN_20] (rows=1501/10 width=215) + Conds:RS_6._col1=RS_7._col0(Inner),Output:["_col0"] + <-Map 1 [SIMPLE_EDGE] + SHUFFLE [RS_6] + PartitionCols:_col1 + Select Operator [SEL_2] (rows=1365/10 width=215) + Output:["_col0","_col1"] + Filter Operator [FIL_18] (rows=1365/10 width=215) + predicate:cint BETWEEN 1000000 AND 3000000 + TableScan [TS_0] (rows=12288/12288 width=215) + default@alltypesorc,a,Tbl:COMPLETE,Col:NONE,Output:["csmallint","cint"] + <-Map 5 [SIMPLE_EDGE] + SHUFFLE [RS_7] + PartitionCols:_col0 + Select Operator [SEL_5] (rows=1365/10 width=215) + Output:["_col0"] + Filter Operator [FIL_19] (rows=1365/10 width=215) + predicate:(cint BETWEEN 1000000 AND 3000000 and cbigint is not null) + TableScan [TS_3] (rows=12288/12288 width=215) + default@alltypesorc,b,Tbl:COMPLETE,Col:NONE,Output:["cint","cbigint"] PREHOOK: query: select a.csmallint, count(*) c1 @@ -311,8 +305,7 @@ order by a.cint POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypesorc #### A masked pattern was here #### -PREHOOK: query: -- Try with dynamically partitioned hashjoin -explain analyze +PREHOOK: query: explain analyze select * from alltypesorc a join alltypesorc b on a.cint = b.cint @@ -320,8 +313,7 @@ where a.cint between 1000000 and 3000000 and b.cbigint is not null order by a.cint PREHOOK: type: QUERY -POSTHOOK: query: -- Try with dynamically partitioned hashjoin -explain analyze +POSTHOOK: query: explain analyze select * from alltypesorc a join alltypesorc b on a.cint = b.cint @@ -341,27 +333,27 @@ Stage-0 Stage-1 Reducer 3 File Output Operator [FS_12] - Select Operator [SEL_11] (rows=6758/10 width=215) + Select Operator [SEL_11] (rows=1501/10 width=215) Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8","_col9","_col10","_col11","_col12","_col13","_col14","_col15","_col16","_col17","_col18","_col19","_col20","_col21","_col22","_col23"] <-Reducer 2 [SIMPLE_EDGE] SHUFFLE [RS_10] - Map Join Operator [MAPJOIN_17] (rows=6758/10 width=215) + Map Join Operator [MAPJOIN_17] (rows=1501/10 width=215) Conds:RS_6.KEY.reducesinkkey0=RS_7.KEY.reducesinkkey0(Inner),HybridGraceHashJoin:true,Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8","_col9","_col10","_col11","_col12","_col13","_col14","_col15","_col16","_col17","_col18","_col19","_col20","_col21","_col22","_col23"] <-Map 4 [CUSTOM_SIMPLE_EDGE] PARTITION_ONLY_SHUFFLE [RS_7] PartitionCols:_col2 - Select Operator [SEL_5] (rows=6144/10 width=215) + Select Operator [SEL_5] (rows=1365/10 width=215) Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8","_col9","_col10","_col11"] - Filter Operator [FIL_16] (rows=6144/10 width=215) - predicate:(cint is not null and cbigint is not null and cint BETWEEN 1000000 AND 3000000) + Filter Operator [FIL_16] (rows=1365/10 width=215) + predicate:(cint BETWEEN 1000000 AND 3000000 and cbigint is not null) TableScan [TS_3] (rows=12288/12288 width=215) default@alltypesorc,b,Tbl:COMPLETE,Col:NONE,Output:["ctinyint","csmallint","cint","cbigint","cfloat","cdouble","cstring1","cstring2","ctimestamp1","ctimestamp2","cboolean1","cboolean2"] <-Map 1 [CUSTOM_SIMPLE_EDGE] PARTITION_ONLY_SHUFFLE [RS_6] PartitionCols:_col2 - Select Operator [SEL_2] (rows=6144/10 width=215) + Select Operator [SEL_2] (rows=1365/10 width=215) Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8","_col9","_col10","_col11"] - Filter Operator [FIL_15] (rows=6144/10 width=215) + Filter Operator [FIL_15] (rows=1365/10 width=215) predicate:cint BETWEEN 1000000 AND 3000000 TableScan [TS_0] (rows=12288/12288 width=215) default@alltypesorc,a,Tbl:COMPLETE,Col:NONE,Output:["ctinyint","csmallint","cint","cbigint","cfloat","cdouble","cstring1","cstring2","ctimestamp1","ctimestamp2","cboolean1","cboolean2"] @@ -435,33 +427,31 @@ Stage-0 limit:-1 Stage-1 Reducer 3 - File Output Operator [FS_14] - Group By Operator [GBY_12] (rows=1/1 width=8) - Output:["_col0"],aggregations:["count(VALUE._col0)"] + File Output Operator [FS_13] + Group By Operator [GBY_11] (rows=1/1 width=8) + Output:["_col0"],aggregations:["count()"] <-Reducer 2 [SIMPLE_EDGE] - SHUFFLE [RS_11] - Group By Operator [GBY_10] (rows=1/14 width=8) - Output:["_col0"],aggregations:["count()"] - Map Join Operator [MAPJOIN_19] (rows=6758/10 width=215) - Conds:RS_6.KEY.reducesinkkey0=RS_7.KEY.reducesinkkey0(Inner),HybridGraceHashJoin:true - <-Map 4 [CUSTOM_SIMPLE_EDGE] - PARTITION_ONLY_SHUFFLE [RS_7] - PartitionCols:_col0 - Select Operator [SEL_5] (rows=6144/10 width=215) - Output:["_col0"] - Filter Operator [FIL_18] (rows=6144/10 width=215) - predicate:(cint is not null and cbigint is not null and cint BETWEEN 1000000 AND 3000000) - TableScan [TS_3] (rows=12288/12288 width=215) - default@alltypesorc,b,Tbl:COMPLETE,Col:NONE,Output:["cint","cbigint"] - <-Map 1 [CUSTOM_SIMPLE_EDGE] - PARTITION_ONLY_SHUFFLE [RS_6] - PartitionCols:_col0 - Select Operator [SEL_2] (rows=6144/10 width=215) - Output:["_col0"] - Filter Operator [FIL_17] (rows=6144/10 width=215) - predicate:cint BETWEEN 1000000 AND 3000000 - TableScan [TS_0] (rows=12288/12288 width=215) - default@alltypesorc,a,Tbl:COMPLETE,Col:NONE,Output:["cint"] + SHUFFLE [RS_10] + Map Join Operator [MAPJOIN_18] (rows=1501/10 width=215) + Conds:RS_6.KEY.reducesinkkey0=RS_7.KEY.reducesinkkey0(Inner),HybridGraceHashJoin:true + <-Map 4 [CUSTOM_SIMPLE_EDGE] + PARTITION_ONLY_SHUFFLE [RS_7] + PartitionCols:_col0 + Select Operator [SEL_5] (rows=1365/10 width=215) + Output:["_col0"] + Filter Operator [FIL_17] (rows=1365/10 width=215) + predicate:(cint BETWEEN 1000000 AND 3000000 and cbigint is not null) + TableScan [TS_3] (rows=12288/12288 width=215) + default@alltypesorc,b,Tbl:COMPLETE,Col:NONE,Output:["cint","cbigint"] + <-Map 1 [CUSTOM_SIMPLE_EDGE] + PARTITION_ONLY_SHUFFLE [RS_6] + PartitionCols:_col0 + Select Operator [SEL_2] (rows=1365/10 width=215) + Output:["_col0"] + Filter Operator [FIL_16] (rows=1365/10 width=215) + predicate:cint BETWEEN 1000000 AND 3000000 + TableScan [TS_0] (rows=12288/12288 width=215) + default@alltypesorc,a,Tbl:COMPLETE,Col:NONE,Output:["cint"] PREHOOK: query: select count(*) @@ -530,38 +520,36 @@ Stage-0 limit:-1 Stage-1 Reducer 4 - File Output Operator [FS_16] - Select Operator [SEL_15] (rows=3379/5 width=215) + File Output Operator [FS_15] + Select Operator [SEL_14] (rows=750/5 width=215) Output:["_col0","_col1"] <-Reducer 3 [SIMPLE_EDGE] - SHUFFLE [RS_14] - Group By Operator [GBY_12] (rows=3379/5 width=215) - Output:["_col0","_col1"],aggregations:["count(VALUE._col0)"],keys:KEY._col0 + SHUFFLE [RS_13] + Group By Operator [GBY_11] (rows=750/5 width=215) + Output:["_col0","_col1"],aggregations:["count()"],keys:KEY._col0 <-Reducer 2 [SIMPLE_EDGE] - SHUFFLE [RS_11] + SHUFFLE [RS_10] PartitionCols:_col0 - Group By Operator [GBY_10] (rows=6758/9 width=215) - Output:["_col0","_col1"],aggregations:["count()"],keys:_col0 - Map Join Operator [MAPJOIN_21] (rows=6758/10 width=215) - Conds:RS_6.KEY.reducesinkkey0=RS_7.KEY.reducesinkkey0(Inner),HybridGraceHashJoin:true,Output:["_col0"] - <-Map 5 [CUSTOM_SIMPLE_EDGE] - PARTITION_ONLY_SHUFFLE [RS_7] - PartitionCols:_col0 - Select Operator [SEL_5] (rows=6144/10 width=215) - Output:["_col0"] - Filter Operator [FIL_20] (rows=6144/10 width=215) - predicate:(cint is not null and cbigint is not null and cint BETWEEN 1000000 AND 3000000) - TableScan [TS_3] (rows=12288/12288 width=215) - default@alltypesorc,b,Tbl:COMPLETE,Col:NONE,Output:["cint","cbigint"] - <-Map 1 [CUSTOM_SIMPLE_EDGE] - PARTITION_ONLY_SHUFFLE [RS_6] - PartitionCols:_col1 - Select Operator [SEL_2] (rows=6144/10 width=215) - Output:["_col0","_col1"] - Filter Operator [FIL_19] (rows=6144/10 width=215) - predicate:cint BETWEEN 1000000 AND 3000000 - TableScan [TS_0] (rows=12288/12288 width=215) - default@alltypesorc,a,Tbl:COMPLETE,Col:NONE,Output:["csmallint","cint"] + Map Join Operator [MAPJOIN_20] (rows=1501/10 width=215) + Conds:RS_6.KEY.reducesinkkey0=RS_7.KEY.reducesinkkey0(Inner),HybridGraceHashJoin:true,Output:["_col0"] + <-Map 5 [CUSTOM_SIMPLE_EDGE] + PARTITION_ONLY_SHUFFLE [RS_7] + PartitionCols:_col0 + Select Operator [SEL_5] (rows=1365/10 width=215) + Output:["_col0"] + Filter Operator [FIL_19] (rows=1365/10 width=215) + predicate:(cint BETWEEN 1000000 AND 3000000 and cbigint is not null) + TableScan [TS_3] (rows=12288/12288 width=215) + default@alltypesorc,b,Tbl:COMPLETE,Col:NONE,Output:["cint","cbigint"] + <-Map 1 [CUSTOM_SIMPLE_EDGE] + PARTITION_ONLY_SHUFFLE [RS_6] + PartitionCols:_col1 + Select Operator [SEL_2] (rows=1365/10 width=215) + Output:["_col0","_col1"] + Filter Operator [FIL_18] (rows=1365/10 width=215) + predicate:cint BETWEEN 1000000 AND 3000000 + TableScan [TS_0] (rows=12288/12288 width=215) + default@alltypesorc,a,Tbl:COMPLETE,Col:NONE,Output:["csmallint","cint"] PREHOOK: query: select a.csmallint, count(*) c1 @@ -583,8 +571,8 @@ order by c1 POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypesorc #### A masked pattern was here #### +-13036 1 -8915 1 -3799 1 10782 1 --13036 1 NULL 6
http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/tez/explainanalyze_5.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/explainanalyze_5.q.out b/ql/src/test/results/clientpositive/tez/explainanalyze_5.q.out index 39bc6f4..65f7b06 100644 --- a/ql/src/test/results/clientpositive/tez/explainanalyze_5.q.out +++ b/ql/src/test/results/clientpositive/tez/explainanalyze_5.q.out @@ -1,33 +1,49 @@ -PREHOOK: query: analyze table src compute statistics -PREHOOK: type: QUERY +PREHOOK: query: drop table src_stats +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table src_stats +POSTHOOK: type: DROPTABLE +PREHOOK: query: create table src_stats as select * from src +PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src -PREHOOK: Output: default@src -FAILED: Hive Internal Error: java.lang.RuntimeException(Cannot overwrite read-only table: src) -java.lang.RuntimeException: Cannot overwrite read-only table: src -#### A masked pattern was here #### - -PREHOOK: query: explain analyze analyze table src compute statistics +PREHOOK: Output: database:default +PREHOOK: Output: default@src_stats +POSTHOOK: query: create table src_stats as select * from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@src_stats +POSTHOOK: Lineage: src_stats.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: src_stats.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: analyze table src_stats compute statistics PREHOOK: type: QUERY -POSTHOOK: query: explain analyze analyze table src compute statistics +PREHOOK: Input: default@src_stats +PREHOOK: Output: default@src_stats +POSTHOOK: query: analyze table src_stats compute statistics +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_stats +POSTHOOK: Output: default@src_stats +PREHOOK: query: explain analyze analyze table src_stats compute statistics +PREHOOK: type: QUERY +POSTHOOK: query: explain analyze analyze table src_stats compute statistics POSTHOOK: type: QUERY Stage-2 Stats-Aggr Operator Stage-0 Map 1 - TableScan [TS_0] (rows=500/0 width=10) - default@src,src,Tbl:COMPLETE,Col:COMPLETE + TableScan [TS_0] (rows=500/500 width=10) + default@src_stats,src_stats,Tbl:COMPLETE,Col:COMPLETE -PREHOOK: query: analyze table src compute statistics for columns +PREHOOK: query: analyze table src_stats compute statistics for columns PREHOOK: type: QUERY -PREHOOK: Input: default@src +PREHOOK: Input: default@src_stats #### A masked pattern was here #### -POSTHOOK: query: analyze table src compute statistics for columns +POSTHOOK: query: analyze table src_stats compute statistics for columns POSTHOOK: type: QUERY -POSTHOOK: Input: default@src +POSTHOOK: Input: default@src_stats #### A masked pattern was here #### -PREHOOK: query: explain analyze analyze table src compute statistics for columns +PREHOOK: query: explain analyze analyze table src_stats compute statistics for columns PREHOOK: type: QUERY -POSTHOOK: query: explain analyze analyze table src compute statistics for columns +POSTHOOK: query: explain analyze analyze table src_stats compute statistics for columns POSTHOOK: type: QUERY Vertex dependency in root stage Reducer 2 <- Map 1 (SIMPLE_EDGE) @@ -36,17 +52,15 @@ Stage-2 Column Stats Work{} Stage-0 Reducer 2 - File Output Operator [FS_6] - Group By Operator [GBY_4] (rows=1/1 width=960) - Output:["_col0","_col1"],aggregations:["compute_stats(VALUE._col0)","compute_stats(VALUE._col1)"] + File Output Operator [FS_5] + Group By Operator [GBY_3] (rows=1/1 width=960) + Output:["_col0","_col1"],aggregations:["compute_stats(VALUE._col0, 16)","compute_stats(VALUE._col2, 16)"] <-Map 1 [SIMPLE_EDGE] - SHUFFLE [RS_3] - Group By Operator [GBY_2] (rows=1/1 width=984) - Output:["_col0","_col1"],aggregations:["compute_stats(key, 16)","compute_stats(value, 16)"] - Select Operator [SEL_1] (rows=500/500 width=178) - Output:["key","value"] - TableScan [TS_0] (rows=500/500 width=178) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"] + SHUFFLE [RS_2] + Select Operator [SEL_1] (rows=500/500 width=10) + Output:["key","value"] + TableScan [TS_0] (rows=500/500 width=10) + default@src_stats,src_stats,Tbl:COMPLETE,Col:NONE,Output:["key","value"] PREHOOK: query: drop table src_multi2 PREHOOK: type: DROPTABLE @@ -94,61 +108,55 @@ Stage-4 Dependency Collection{} Stage-1 Reducer 5 - File Output Operator [FS_6] - Group By Operator [GBY_4] (rows=1/1 width=960) - Output:["_col0","_col1"],aggregations:["compute_stats(VALUE._col0)","compute_stats(VALUE._col1)"] + File Output Operator [FS_5] + Group By Operator [GBY_3] (rows=1/1 width=960) + Output:["_col0","_col1"],aggregations:["compute_stats(VALUE._col0, 16)","compute_stats(VALUE._col2, 16)"] <-Reducer 4 [SIMPLE_EDGE] - File Output Operator [FS_20] + File Output Operator [FS_19] table:{"name:":"default.src_multi2"} - Select Operator [SEL_19] (rows=639/508 width=178) + Select Operator [SEL_18] (rows=550/508 width=10) Output:["_col0","_col1"] - Merge Join Operator [MERGEJOIN_27] (rows=639/508 width=178) - Conds:RS_16._col0=RS_17._col0(Inner),Output:["_col0","_col3"] + Merge Join Operator [MERGEJOIN_26] (rows=550/508 width=10) + Conds:RS_15._col0=RS_16._col0(Inner),Output:["_col0","_col3"] <-Map 7 [SIMPLE_EDGE] - SHUFFLE [RS_17] + SHUFFLE [RS_16] PartitionCols:_col0 - Select Operator [SEL_15] (rows=500/500 width=178) + Select Operator [SEL_14] (rows=500/500 width=10) Output:["_col0","_col1"] - Filter Operator [FIL_26] (rows=500/500 width=178) + Filter Operator [FIL_25] (rows=500/500 width=10) predicate:key is not null - TableScan [TS_13] (rows=500/500 width=178) - default@src,src,Tbl:COMPLETE,Col:COMPLETE,Output:["key","value"] + TableScan [TS_12] (rows=500/500 width=10) + default@src,src,Tbl:COMPLETE,Col:NONE,Output:["key","value"] <-Reducer 3 [SIMPLE_EDGE] - SHUFFLE [RS_16] + SHUFFLE [RS_15] PartitionCols:_col0 - Select Operator [SEL_12] (rows=262/319 width=178) + Select Operator [SEL_11] (rows=262/319 width=10) Output:["_col0"] - Group By Operator [GBY_11] (rows=262/319 width=178) + Group By Operator [GBY_10] (rows=262/319 width=10) Output:["_col0","_col1"],keys:KEY._col0, KEY._col1 <-Union 2 [SIMPLE_EDGE] <-Map 1 [CONTAINS] - Reduce Output Operator [RS_10] + Reduce Output Operator [RS_9] PartitionCols:_col0, _col1 - Group By Operator [GBY_9] (rows=262/331 width=178) - Output:["_col0","_col1"],keys:_col0, _col1 - Select Operator [SEL_2] (rows=500/500 width=178) - Output:["_col0","_col1"] - Filter Operator [FIL_24] (rows=500/500 width=178) - predicate:key is not null - TableScan [TS_0] (rows=500/500 width=178) - Output:["key","value"] + Select Operator [SEL_2] (rows=500/500 width=10) + Output:["_col0","_col1"] + Filter Operator [FIL_23] (rows=500/500 width=10) + predicate:key is not null + TableScan [TS_0] (rows=500/500 width=10) + Output:["key","value"] <-Map 6 [CONTAINS] - Reduce Output Operator [RS_10] + Reduce Output Operator [RS_9] PartitionCols:_col0, _col1 - Group By Operator [GBY_9] (rows=262/331 width=178) - Output:["_col0","_col1"],keys:_col0, _col1 - Select Operator [SEL_5] (rows=25/25 width=175) - Output:["_col0","_col1"] - Filter Operator [FIL_25] (rows=25/25 width=175) - predicate:key is not null - TableScan [TS_3] (rows=25/25 width=175) - Output:["key","value"] - SHUFFLE [RS_3] - Group By Operator [GBY_2] (rows=1/1 width=984) - Output:["_col0","_col1"],aggregations:["compute_stats(key, 16)","compute_stats(value, 16)"] - Select Operator [SEL_1] (rows=639/508 width=178) - Output:["key","value"] - Please refer to the previous Select Operator [SEL_19] + Select Operator [SEL_5] (rows=25/25 width=7) + Output:["_col0","_col1"] + Filter Operator [FIL_24] (rows=25/25 width=7) + predicate:key is not null + TableScan [TS_3] (rows=25/25 width=7) + Output:["key","value"] + SHUFFLE [RS_2] + Select Operator [SEL_1] (rows=550/508 width=10) + Output:["key","value"] + Please refer to the previous Select Operator [SEL_18] PREHOOK: query: select count(*) from (select * from src union select * from src1)subq PREHOOK: type: QUERY @@ -208,17 +216,13 @@ Bucket Columns: [] Sort Columns: [] Storage Desc Params: serialization.format 1 -PREHOOK: query: -- SORT_QUERY_RESULTS - -create table acid_uami(i int, +PREHOOK: query: create table acid_uami(i int, de decimal(5,2), vc varchar(128)) clustered by (i) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true') PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@acid_uami -POSTHOOK: query: -- SORT_QUERY_RESULTS - -create table acid_uami(i int, +POSTHOOK: query: create table acid_uami(i int, de decimal(5,2), vc varchar(128)) clustered by (i) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true') POSTHOOK: type: CREATETABLE @@ -292,17 +296,17 @@ Stage-3 Reducer 2 File Output Operator [FS_8] table:{"name:":"default.acid_uami"} - Select Operator [SEL_4] (rows=6/2 width=302) + Select Operator [SEL_4] (rows=8 width=245) Output:["_col0","_col1","_col2","_col3"] <-Map 1 [SIMPLE_EDGE] SHUFFLE [RS_3] PartitionCols:UDFToInteger(_col0) - Select Operator [SEL_2] (rows=6/2 width=302) + Select Operator [SEL_2] (rows=8 width=245) Output:["_col0","_col1","_col3"] - Filter Operator [FIL_9] (rows=6/2 width=226) + Filter Operator [FIL_9] (rows=8 width=245) predicate:((de = 109.23) or (de = 119.23)) - TableScan [TS_0] (rows=8/4 width=226) - default@acid_uami,acid_uami, ACID table,Tbl:COMPLETE,Col:COMPLETE,Output:["i","de","vc"] + TableScan [TS_0] (rows=8 width=245) + default@acid_uami,acid_uami, ACID table,Tbl:COMPLETE,Col:NONE,Output:["i","de","vc"] PREHOOK: query: select * from acid_uami order by de PREHOOK: type: QUERY @@ -405,16 +409,16 @@ Stage-3 Reducer 2 File Output Operator [FS_7] table:{"name:":"default.acid_dot"} - Select Operator [SEL_4] (rows=31436/8 width=4) + Select Operator [SEL_4] (rows=31436 width=4) Output:["_col0"] <-Map 1 [SIMPLE_EDGE] SHUFFLE [RS_3] PartitionCols:UDFToInteger(_col0) - Select Operator [SEL_2] (rows=31436/8 width=4) + Select Operator [SEL_2] (rows=31436 width=4) Output:["_col0"] - Filter Operator [FIL_8] (rows=31436/8 width=4) + Filter Operator [FIL_8] (rows=31436 width=4) predicate:(cint < -1070551679) - TableScan [TS_0] (rows=94309/12288 width=4) + TableScan [TS_0] (rows=94309 width=4) default@acid_dot,acid_dot, ACID table,Tbl:COMPLETE,Col:NONE,Output:["cint"] PREHOOK: query: select count(*) from acid_dot http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/tez/explainuser_2.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/explainuser_2.q.out b/ql/src/test/results/clientpositive/tez/explainuser_2.q.out index a4d538b..68b5ba9 100644 --- a/ql/src/test/results/clientpositive/tez/explainuser_2.q.out +++ b/ql/src/test/results/clientpositive/tez/explainuser_2.q.out @@ -3911,13 +3911,13 @@ Stage-4 PartitionCols:_col0 Group By Operator [GBY_17] (rows=1 width=280) Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, substr(_col1, 5) - Group By Operator [GBY_14] (rows=1 width=172) + Group By Operator [GBY_14] (rows=1 width=272) Output:["_col0","_col1"],keys:KEY._col0, KEY._col1 <-Union 3 [SIMPLE_EDGE] <-Map 6 [CONTAINS] Reduce Output Operator [RS_13] PartitionCols:_col0, _col1 - Group By Operator [GBY_12] (rows=1 width=172) + Group By Operator [GBY_12] (rows=1 width=272) Output:["_col0","_col1"],keys:_col0, _col1 Select Operator [SEL_8] (rows=500 width=10) Output:["_col0","_col1"] @@ -3926,9 +3926,9 @@ Stage-4 <-Reducer 2 [CONTAINS] Reduce Output Operator [RS_13] PartitionCols:_col0, _col1 - Group By Operator [GBY_12] (rows=1 width=172) + Group By Operator [GBY_12] (rows=1 width=272) Output:["_col0","_col1"],keys:_col0, _col1 - Select Operator [SEL_6] (rows=1 width=172) + Select Operator [SEL_6] (rows=1 width=272) Output:["_col0","_col1"] Group By Operator [GBY_5] (rows=1 width=8) Output:["_col0"],aggregations:["count(VALUE._col0)"] @@ -3941,9 +3941,9 @@ Stage-4 default@src,s1,Tbl:COMPLETE,Col:COMPLETE File Output Operator [FS_27] table:{"name:":"default.dest2"} - Select Operator [SEL_26] (rows=1 width=356) + Select Operator [SEL_26] (rows=1 width=456) Output:["_col0","_col1","_col2"] - Group By Operator [GBY_25] (rows=1 width=364) + Group By Operator [GBY_25] (rows=1 width=464) Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, _col1 Please refer to the previous Group By Operator [GBY_14] Stage-5 @@ -4090,7 +4090,7 @@ Stage-4 Output:["key","value"] Reduce Output Operator [RS_22] PartitionCols:_col0, _col1 - Group By Operator [GBY_21] (rows=1 width=364) + Group By Operator [GBY_21] (rows=1 width=464) Output:["_col0","_col1","_col2","_col3"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, _col1, substr(_col1, 5) Please refer to the previous Select Operator [SEL_8] <-Map 7 [CONTAINS] @@ -4104,7 +4104,7 @@ Stage-4 Output:["key","value"] Reduce Output Operator [RS_22] PartitionCols:_col0, _col1 - Group By Operator [GBY_21] (rows=1 width=364) + Group By Operator [GBY_21] (rows=1 width=464) Output:["_col0","_col1","_col2","_col3"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, _col1, substr(_col1, 5) Please refer to the previous Select Operator [SEL_12] <-Reducer 2 [CONTAINS] @@ -4112,7 +4112,7 @@ Stage-4 PartitionCols:_col0 Group By Operator [GBY_15] (rows=1 width=280) Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, substr(_col1, 5) - Select Operator [SEL_6] (rows=1 width=172) + Select Operator [SEL_6] (rows=1 width=272) Output:["_col0","_col1"] Group By Operator [GBY_5] (rows=1 width=8) Output:["_col0"],aggregations:["count(VALUE._col0)"] @@ -4125,13 +4125,13 @@ Stage-4 default@src,s1,Tbl:COMPLETE,Col:COMPLETE Reduce Output Operator [RS_22] PartitionCols:_col0, _col1 - Group By Operator [GBY_21] (rows=1 width=364) + Group By Operator [GBY_21] (rows=1 width=464) Output:["_col0","_col1","_col2","_col3"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, _col1, substr(_col1, 5) Please refer to the previous Select Operator [SEL_6] Reducer 5 File Output Operator [FS_25] table:{"name:":"default.dest2"} - Group By Operator [GBY_23] (rows=1 width=180) + Group By Operator [GBY_23] (rows=1 width=280) Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1 <- Please refer to the previous Union 3 [SIMPLE_EDGE] Stage-5 @@ -4194,7 +4194,7 @@ Stage-4 Output:["key","value"] Reduce Output Operator [RS_19] PartitionCols:_col0, _col1 - Group By Operator [GBY_18] (rows=1 width=364) + Group By Operator [GBY_18] (rows=1 width=464) Output:["_col0","_col1","_col2","_col3"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, _col1, substr(_col1, 5) Please refer to the previous Select Operator [SEL_8] <-Reducer 2 [CONTAINS] @@ -4202,7 +4202,7 @@ Stage-4 PartitionCols:_col0 Group By Operator [GBY_12] (rows=1 width=280) Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, substr(_col1, 5) - Select Operator [SEL_6] (rows=1 width=172) + Select Operator [SEL_6] (rows=1 width=272) Output:["_col0","_col1"] Group By Operator [GBY_5] (rows=1 width=8) Output:["_col0"],aggregations:["count(VALUE._col0)"] @@ -4215,13 +4215,13 @@ Stage-4 default@src,s1,Tbl:COMPLETE,Col:COMPLETE Reduce Output Operator [RS_19] PartitionCols:_col0, _col1 - Group By Operator [GBY_18] (rows=1 width=364) + Group By Operator [GBY_18] (rows=1 width=464) Output:["_col0","_col1","_col2","_col3"],aggregations:["count(DISTINCT substr(_col1, 5))"],keys:_col0, _col1, substr(_col1, 5) Please refer to the previous Select Operator [SEL_6] Reducer 5 File Output Operator [FS_22] table:{"name:":"default.dest2"} - Group By Operator [GBY_20] (rows=1 width=180) + Group By Operator [GBY_20] (rows=1 width=280) Output:["_col0","_col1","_col2"],aggregations:["count(DISTINCT KEY._col2:0._col0)"],keys:KEY._col0, KEY._col1 <- Please refer to the previous Union 3 [SIMPLE_EDGE] Stage-5 http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/tez/mergejoin.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/mergejoin.q.out b/ql/src/test/results/clientpositive/tez/mergejoin.q.out index e6b81ca..5853b2b 100644 --- a/ql/src/test/results/clientpositive/tez/mergejoin.q.out +++ b/ql/src/test/results/clientpositive/tez/mergejoin.q.out @@ -64,10 +64,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=25) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Reducer 2 Reduce Operator Tree: @@ -93,10 +93,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=25) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 @@ -1860,10 +1860,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=25) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Reducer 2 Reduce Operator Tree: @@ -1932,10 +1932,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=25) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 @@ -2006,10 +2006,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=242) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Execution mode: vectorized Map 5 @@ -2071,10 +2071,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=242) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 @@ -2343,10 +2343,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=242) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Execution mode: vectorized Map 5 @@ -2408,10 +2408,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=242) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 @@ -2535,10 +2535,10 @@ STAGE PLANS: aggregations: min(_col0), max(_col0), bloom_filter(_col0, expectedEntries=25) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Reducer 2 Reduce Operator Tree: @@ -2607,10 +2607,10 @@ STAGE PLANS: aggregations: min(VALUE._col0), max(VALUE._col1), bloom_filter(VALUE._col2, expectedEntries=25) mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator sort order: - Statistics: Num rows: 1 Data size: 252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: binary) Stage: Stage-0 http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/tez/tez_union_multiinsert.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/tez_union_multiinsert.q.out b/ql/src/test/results/clientpositive/tez/tez_union_multiinsert.q.out index ddc56ce..c69c346 100644 --- a/ql/src/test/results/clientpositive/tez/tez_union_multiinsert.q.out +++ b/ql/src/test/results/clientpositive/tez/tez_union_multiinsert.q.out @@ -101,12 +101,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Map 7 Map Operator Tree: TableScan @@ -132,12 +132,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reducer 2 Reduce Operator Tree: Group By Operator @@ -148,7 +148,7 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: count(DISTINCT substr(_col1, 5)) keys: _col0 (type: string), substr(_col1, 5) (type: string) @@ -165,12 +165,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reducer 4 Reduce Operator Tree: Group By Operator @@ -194,10 +194,10 @@ STAGE PLANS: keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -975,12 +975,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Map 5 Map Operator Tree: TableScan @@ -1022,12 +1022,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reducer 3 Reduce Operator Tree: Group By Operator @@ -1051,10 +1051,10 @@ STAGE PLANS: keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -1070,7 +1070,7 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: count(DISTINCT substr(_col1, 5)) keys: _col0 (type: string), substr(_col1, 5) (type: string) @@ -1087,12 +1087,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Union 2 Vertex: Union 2 @@ -1861,12 +1861,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Map 5 Map Operator Tree: TableScan @@ -1908,12 +1908,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reducer 3 Reduce Operator Tree: Group By Operator @@ -1937,10 +1937,10 @@ STAGE PLANS: keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -1956,7 +1956,7 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: count(DISTINCT substr(_col1, 5)) keys: _col0 (type: string), substr(_col1, 5) (type: string) @@ -1973,12 +1973,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Union 2 Vertex: Union 2 @@ -2754,12 +2754,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reducer 2 Reduce Operator Tree: Group By Operator @@ -2770,7 +2770,7 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator aggregations: count(DISTINCT substr(_col1, 5)) keys: _col0 (type: string), substr(_col1, 5) (type: string) @@ -2787,12 +2787,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reducer 4 Reduce Operator Tree: Group By Operator @@ -2816,10 +2816,10 @@ STAGE PLANS: keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -3582,12 +3582,12 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL Reducer 2 Reduce Operator Tree: Group By Operator @@ -3598,24 +3598,24 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE Group By Operator keys: _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL Reducer 4 Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: PARTIAL Group By Operator aggregations: count(DISTINCT substr(_col1, 5)) keys: _col0 (type: string), substr(_col1, 5) (type: string) @@ -3632,14 +3632,14 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string) mode: complete outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Select Operator expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 356 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 356 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/tez/union6.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/tez/union6.q.out b/ql/src/test/results/clientpositive/tez/union6.q.out index 3d8a5ca..dc99bde 100644 --- a/ql/src/test/results/clientpositive/tez/union6.q.out +++ b/ql/src/test/results/clientpositive/tez/union6.q.out @@ -60,7 +60,7 @@ STAGE PLANS: Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 26 Data size: 363 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 26 Data size: 463 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -76,10 +76,10 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 26 Data size: 363 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 26 Data size: 463 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/union17.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/union17.q.out b/ql/src/test/results/clientpositive/union17.q.out index 91ff95f..5c2e6ac 100644 --- a/ql/src/test/results/clientpositive/union17.q.out +++ b/ql/src/test/results/clientpositive/union17.q.out @@ -64,7 +64,7 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false table: @@ -77,7 +77,7 @@ STAGE PLANS: Map Operator Tree: TableScan Union - Statistics: Num rows: 501 Data size: 5484 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 5584 Basic stats: COMPLETE Column stats: PARTIAL Group By Operator aggregations: count(DISTINCT substr(_col1, 5)) keys: _col0 (type: string), substr(_col1, 5) (type: string) @@ -94,7 +94,7 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false table: @@ -109,7 +109,7 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Union - Statistics: Num rows: 501 Data size: 5484 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 5584 Basic stats: COMPLETE Column stats: PARTIAL Group By Operator aggregations: count(DISTINCT substr(_col1, 5)) keys: _col0 (type: string), substr(_col1, 5) (type: string) @@ -126,7 +126,7 @@ STAGE PLANS: keys: _col0 (type: string), _col1 (type: string), substr(_col1, 5) (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false table: @@ -170,17 +170,17 @@ STAGE PLANS: key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 364 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 464 Basic stats: COMPLETE Column stats: PARTIAL Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col2:0._col0) keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 180 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/9f5bea3b/ql/src/test/results/clientpositive/union18.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/union18.q.out b/ql/src/test/results/clientpositive/union18.q.out index 146387f..caa664c 100644 --- a/ql/src/test/results/clientpositive/union18.q.out +++ b/ql/src/test/results/clientpositive/union18.q.out @@ -73,7 +73,7 @@ STAGE PLANS: Select Operator expressions: 'tst1' (type: string), UDFToString(_col0) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 272 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false table: @@ -86,10 +86,10 @@ STAGE PLANS: Map Operator Tree: TableScan Union - Statistics: Num rows: 501 Data size: 5484 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 5584 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 501 Data size: 5484 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 5584 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -98,10 +98,10 @@ STAGE PLANS: Select Operator expressions: _col0 (type: string), _col1 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 501 Data size: 128256 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 228456 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 501 Data size: 128256 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 228456 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -115,10 +115,10 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Union - Statistics: Num rows: 501 Data size: 5484 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 5584 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 501 Data size: 5484 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 5584 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -127,10 +127,10 @@ STAGE PLANS: Select Operator expressions: _col0 (type: string), _col1 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 501 Data size: 128256 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 228456 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 501 Data size: 128256 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 501 Data size: 228456 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat