Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_decimal.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_decimal.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_decimal.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_decimal.q.out Fri Nov 7 20:42:43 2014 @@ -1,45 +1,3 @@ -PREHOOK: query: DROP TABLE IF EXISTS part -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE IF EXISTS part -POSTHOOK: type: DROPTABLE -PREHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DECIMAL(6,2), - p_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part -POSTHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DECIMAL(6,2), - p_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@part -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@part PREHOOK: query: -- 1. aggregate functions with decimal type select p_mfgr, p_retailprice,
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_general_queries.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_general_queries.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_general_queries.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_general_queries.q.out Fri Nov 7 20:42:43 2014 @@ -1,45 +1,3 @@ -PREHOOK: query: DROP TABLE part -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE part -POSTHOOK: type: DROPTABLE -PREHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part -POSTHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@part -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@part PREHOOK: query: -- 1. testNoPTFNoWindowing select p_mfgr, p_name, p_size from part Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_streaming.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_streaming.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_streaming.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/ptf_streaming.q.out Fri Nov 7 20:42:43 2014 @@ -1,45 +1,3 @@ -PREHOOK: query: DROP TABLE part -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE part -POSTHOOK: type: DROPTABLE -PREHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part -POSTHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@part -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@part PREHOOK: query: create temporary function noopstreaming as 'org.apache.hadoop.hive.ql.udf.ptf.NoopStreaming$NoopStreamingResolver' PREHOOK: type: CREATEFUNCTION PREHOOK: Output: noopstreaming Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/subquery_in.q.out Fri Nov 7 20:42:43 2014 @@ -1,110 +1,14 @@ PREHOOK: query: -- SORT_QUERY_RESULTS -DROP TABLE part -PREHOOK: type: DROPTABLE -POSTHOOK: query: -- SORT_QUERY_RESULTS - -DROP TABLE part -POSTHOOK: type: DROPTABLE -PREHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part -POSTHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@part -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@part -PREHOOK: query: DROP TABLE lineitem -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE lineitem -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT, - L_PARTKEY INT, - L_SUPPKEY INT, - L_LINENUMBER INT, - L_QUANTITY DOUBLE, - L_EXTENDEDPRICE DOUBLE, - L_DISCOUNT DOUBLE, - L_TAX DOUBLE, - L_RETURNFLAG STRING, - L_LINESTATUS STRING, - l_shipdate STRING, - L_COMMITDATE STRING, - L_RECEIPTDATE STRING, - L_SHIPINSTRUCT STRING, - L_SHIPMODE STRING, - L_COMMENT STRING) -ROW FORMAT DELIMITED -FIELDS TERMINATED BY '|' -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@lineitem -POSTHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT, - L_PARTKEY INT, - L_SUPPKEY INT, - L_LINENUMBER INT, - L_QUANTITY DOUBLE, - L_EXTENDEDPRICE DOUBLE, - L_DISCOUNT DOUBLE, - L_TAX DOUBLE, - L_RETURNFLAG STRING, - L_LINESTATUS STRING, - l_shipdate STRING, - L_COMMITDATE STRING, - L_RECEIPTDATE STRING, - L_SHIPINSTRUCT STRING, - L_SHIPMODE STRING, - L_COMMENT STRING) -ROW FORMAT DELIMITED -FIELDS TERMINATED BY '|' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lineitem -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@lineitem -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@lineitem -PREHOOK: query: -- non agg, non corr +-- non agg, non corr explain select * from src where src.key in (select key from src s1 where s1.key > '9') PREHOOK: type: QUERY -POSTHOOK: query: -- non agg, non corr +POSTHOOK: query: -- SORT_QUERY_RESULTS + +-- non agg, non corr explain select * from src @@ -370,40 +274,40 @@ STAGE PLANS: Map Operator Tree: TableScan alias: part - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: p_mfgr (type: string), p_size (type: int) sort order: ++ Map-reduce partition columns: p_mfgr (type: string) - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE value expressions: p_mfgr (type: string), p_size (type: int) Map 5 Map Operator Tree: TableScan alias: part - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: UDFToDouble(p_size) is not null (type: boolean) - Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: UDFToDouble(p_size) (type: double) sort order: + Map-reduce partition columns: UDFToDouble(p_size) (type: double) - Statistics: Num rows: 15 Data size: 1586 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 13 Data size: 1573 Basic stats: COMPLETE Column stats: NONE value expressions: p_name (type: string), p_size (type: int) Reducer 2 Reduce Operator Tree: Extract - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE PTF Operator - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (_wcol0 <= 2) (type: boolean) - Statistics: Num rows: 10 Data size: 1057 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col5 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 10 Data size: 1057 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 8 Data size: 968 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: avg(_col0) mode: hash @@ -446,14 +350,14 @@ STAGE PLANS: 0 {VALUE._col1} {VALUE._col5} 1 outputColumnNames: _col1, _col5 - Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col5 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 16 Data size: 1744 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 14 Data size: 1730 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -522,26 +426,26 @@ STAGE PLANS: Map Operator Tree: TableScan alias: b - Statistics: Num rows: 15 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (p_size is not null and p_mfgr is not null) (type: boolean) - Statistics: Num rows: 4 Data size: 846 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: p_size (type: int), p_mfgr (type: string) sort order: ++ Map-reduce partition columns: p_size (type: int), p_mfgr (type: string) - Statistics: Num rows: 4 Data size: 846 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 7 Data size: 847 Basic stats: COMPLETE Column stats: NONE value expressions: p_name (type: string) Map 3 Map Operator Tree: TableScan alias: part - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: p_mfgr (type: string), p_size (type: int) sort order: ++ Map-reduce partition columns: p_mfgr (type: string) - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE value expressions: p_mfgr (type: string), p_size (type: int) Reducer 2 Reduce Operator Tree: @@ -552,14 +456,14 @@ STAGE PLANS: 0 {VALUE._col1} {KEY.reducesinkkey1} {KEY.reducesinkkey0} 1 outputColumnNames: _col1, _col2, _col5 - Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col2 (type: string), _col1 (type: string), _col5 (type: int) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 4 Data size: 930 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 7 Data size: 931 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -567,27 +471,27 @@ STAGE PLANS: Reducer 4 Reduce Operator Tree: Extract - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE PTF Operator - Statistics: Num rows: 30 Data size: 3173 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 26 Data size: 3147 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((_wcol0 <= 2) and _col2 is not null) (type: boolean) - Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col2 (type: string), _col5 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: min(_col1) keys: _col0 (type: string) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 5 Data size: 528 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 484 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: int) Reducer 5 Reduce Operator Tree: @@ -596,24 +500,24 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: _col1 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: int), _col0 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: int), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 2 Data size: 211 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 242 Basic stats: COMPLETE Column stats: NONE Stage: Stage-0 Fetch Operator @@ -845,60 +749,60 @@ STAGE PLANS: Map Operator Tree: TableScan alias: li - Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((l_partkey is not null and l_orderkey is not null) and (l_linenumber = 1)) (type: boolean) - Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: l_partkey (type: int) sort order: + Map-reduce partition columns: l_partkey (type: int) - Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE value expressions: l_orderkey (type: int), l_suppkey (type: int) Map 4 Map Operator Tree: TableScan alias: lineitem - Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: l_partkey is not null (type: boolean) - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: l_partkey (type: int) outputColumnNames: l_partkey - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: l_partkey (type: int) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Map 6 Map Operator Tree: TableScan alias: lineitem - Statistics: Num rows: 1728 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((l_shipmode = 'AIR') and l_orderkey is not null) (type: boolean) - Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: l_orderkey (type: int) outputColumnNames: _col0 - Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: int) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Join Operator @@ -908,12 +812,12 @@ STAGE PLANS: 0 {KEY.reducesinkkey0} 1 {VALUE._col0} {VALUE._col1} outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col1 (type: int) sort order: + Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: int), _col3 (type: int) Reducer 3 Reduce Operator Tree: @@ -924,14 +828,14 @@ STAGE PLANS: 0 {VALUE._col0} {VALUE._col2} 1 outputColumnNames: _col0, _col3 - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col3 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -942,16 +846,16 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Stage: Stage-0 Fetch Operator Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/temp_table.q.out Fri Nov 7 20:42:43 2014 @@ -355,9 +355,15 @@ alltypesorc bar bay baz +cbo_t1 +cbo_t2 +cbo_t3 foo +lineitem +part src src1 +src_cbo src_json src_sequencefile src_thrift Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union11.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union11.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union11.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union11.q.out Fri Nov 7 20:42:43 2014 @@ -110,14 +110,14 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -191,6 +191,6 @@ POSTHOOK: query: select unionsrc.key, co POSTHOOK: type: QUERY POSTHOOK: Input: default@src #### A masked pattern was here #### -tst3 1 -tst1 1 tst2 1 +tst1 1 +tst3 1 Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union14.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union14.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union14.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union14.q.out Fri Nov 7 20:42:43 2014 @@ -72,14 +72,14 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -131,20 +131,20 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Input: default@src1 #### A masked pattern was here #### -278 1 -273 1 -128 1 255 1 -tst1 1 +213 1 238 1 -146 1 150 1 - 10 -406 1 -369 1 -213 1 311 1 -224 1 +146 1 +369 1 66 1 + 10 +278 1 401 1 +406 1 +tst1 1 +128 1 98 1 +273 1 +224 1 Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union15.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union15.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union15.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union15.q.out Fri Nov 7 20:42:43 2014 @@ -116,14 +116,14 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL Select Operator expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out Fri Nov 7 20:42:43 2014 @@ -134,14 +134,14 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL Select Operator expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union5.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union5.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union5.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union5.q.out Fri Nov 7 20:42:43 2014 @@ -89,14 +89,14 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -144,5 +144,5 @@ POSTHOOK: query: select unionsrc.key, co POSTHOOK: type: QUERY POSTHOOK: Input: default@src #### A masked pattern was here #### -tst1 1 tst2 1 +tst1 1 Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union7.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union7.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union7.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union7.q.out Fri Nov 7 20:42:43 2014 @@ -92,14 +92,14 @@ STAGE PLANS: keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL Select Operator expressions: _col0 (type: string), _col1 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 100 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 1 Data size: 96 Basic stats: COMPLETE Column stats: PARTIAL table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out Fri Nov 7 20:42:43 2014 @@ -718,15 +718,15 @@ POSTHOOK: Input: default@decimal_date_te 14.9324324324 19.1135135135 20.3081081081 -22.1000000000 +22.1 24.4891891892 33.4486486486 34.6432432432 40.0189189189 42.4081081081 43.0054054054 -44.2000000000 -44.2000000000 +44.2 +44.2 44.7972972973 45.9918918919 PREHOOK: query: SELECT COUNT(*) FROM decimal_date_test WHERE cdecimal1 NOT BETWEEN -2000 AND 4390.1351351351 Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out Fri Nov 7 20:42:43 2014 @@ -113,14 +113,14 @@ POSTHOOK: query: SELECT cint, POSTHOOK: type: QUERY POSTHOOK: Input: default@decimal_vgby #### A masked pattern was here #### -762 2 5831542.2692483780 1531.2194054054 5833073.4886537834 2 6984454.21109769200000 1833.9456923076925 6986288.15678999969250 -253665376 1024 9767.0054054054 -9779.5486486487 -347484.0818378374 1024 11697.96923076923100 -11712.99230769231000 -416182.64030769233089 --3728 6 5831542.2692483780 -3367.6517567568 5817556.0411483778 6 6984454.21109769200000 -4033.445769230769 6967702.86724384584710 --563 2 -515.6210729730 -3367.6517567568 -3883.2728297298 2 -617.56077692307690 -4033.445769230769 -4651.00654615384590 -6981 3 5831542.269248378 -515.6210729730 5830511.0271024320 3 6984454.211097692 -617.56077692307690 6983219.08954384584620 -626923679 1024 9723.4027027027 -9778.9513513514 10541.0525297287 1024 11645.74615384615400 -11712.27692307692300 12625.04759999997746 -528534767 1024 5831542.2692483780 -9777.1594594595 11646372.8607481068 1024 6984454.21109769200000 -11710.13076923077100 13948892.79980307629003 -NULL 3072 9318.4351351351 -4298.1513513514 5018444.1081079808 3072 11160.71538461538500 -5147.90769230769300 6010604.30769230735360 +762 2 5831542.269248378 1531.2194054054 5833073.4886537834 2 6984454.211097692 1833.9456923076925 6986288.1567899996925 +253665376 1024 9767.0054054054 -9779.5486486487 -347484.0818378374 1024 11697.969230769231 -11712.99230769231 -416182.64030769233089 +-3728 6 5831542.269248378 -3367.6517567568 5817556.0411483778 6 6984454.211097692 -4033.445769230769 6967702.8672438458471 +-563 2 -515.6210729730 -3367.6517567568 -3883.2728297298 2 -617.5607769230769 -4033.445769230769 -4651.0065461538459 +6981 3 5831542.269248378 -515.6210729730 5830511.0271024320 3 6984454.211097692 -617.5607769230769 6983219.0895438458462 +626923679 1024 9723.4027027027 -9778.9513513514 10541.0525297287 1024 11645.746153846154 -11712.276923076923 12625.04759999997746 +528534767 1024 5831542.269248378 -9777.1594594595 11646372.8607481068 1024 6984454.211097692 -11710.130769230771 13948892.79980307629003 +NULL 3072 9318.4351351351 -4298.1513513514 5018444.1081079808 3072 11160.715384615385 -5147.907692307693 6010604.3076923073536 PREHOOK: query: -- Now add the others... EXPLAIN SELECT cint, COUNT(cdecimal1), MAX(cdecimal1), MIN(cdecimal1), SUM(cdecimal1), AVG(cdecimal1), STDDEV_POP(cdecimal1), STDDEV_SAMP(cdecimal1), @@ -217,11 +217,11 @@ POSTHOOK: query: SELECT cint, POSTHOOK: type: QUERY POSTHOOK: Input: default@decimal_vgby #### A masked pattern was here #### -762 2 5831542.2692483780 1531.2194054054 5833073.4886537834 2916536.7443268917 2915005.5249214866 4122440.3477364695 2 6984454.21109769200000 1833.9456923076925 6986288.15678999969250 3493144.07839499984625 3491310.1327026924 4937458.140118758 -253665376 1024 9767.0054054054 -9779.5486486487 -347484.0818378374 -339.33992366976309 5708.9563478862 5711.745967572779 1024 11697.96923076923100 -11712.99230769231000 -416182.64030769233089 -406.428359675480791885 6837.632716002934 6840.973851172274 --3728 6 5831542.2692483780 -3367.6517567568 5817556.0411483778 969592.67352472963333 2174330.2092403853 2381859.406131774 6 6984454.21109769200000 -4033.445769230769 6967702.86724384584710 1161283.811207307641183333 2604201.2704476737 2852759.5602156054 --563 2 -515.6210729730 -3367.6517567568 -3883.2728297298 -1941.6364148649 1426.0153418918999 2016.6902366556308 2 -617.56077692307690 -4033.445769230769 -4651.00654615384590 -2325.50327307692295 1707.9424961538462 2415.395441814127 -6981 3 5831542.269248378 -515.6210729730 5830511.0271024320 1943503.67570081066667 2749258.455012492 3367140.1929065133 3 6984454.211097692 -617.56077692307690 6983219.08954384584620 2327739.696514615282066667 3292794.4113115156 4032833.0678006653 -626923679 1024 9723.4027027027 -9778.9513513514 10541.0525297287 10.29399661106318 5742.09145323734 5744.897264034267 1024 11645.74615384615400 -11712.27692307692300 12625.04759999997746 12.329148046874977988 6877.318722794877 6880.679250101603 -528534767 1024 5831542.2692483780 -9777.1594594595 11646372.8607481068 11373.41099682432305 257528.92988206653 257654.7686043977 1024 6984454.21109769200000 -11710.13076923077100 13948892.79980307629003 13621.965624807691689482 308443.1074570801 308593.82484083984 -NULL 3072 9318.4351351351 -4298.1513513514 5018444.1081079808 1633.60810810806667 5695.483082135364 5696.4103077145055 3072 11160.71538461538500 -5147.90769230769300 6010604.30769230735360 1956.576923076922966667 6821.495748565159 6822.606289190924 +762 2 5831542.269248378 1531.2194054054 5833073.4886537834 2916536.7443268917 2915005.5249214866 4122440.3477364695 2 6984454.211097692 1833.9456923076925 6986288.1567899996925 3493144.07839499984625 3491310.1327026924 4937458.140118758 +253665376 1024 9767.0054054054 -9779.5486486487 -347484.0818378374 -339.33992366976309 5708.9563478862 5711.745967572779 1024 11697.969230769231 -11712.99230769231 -416182.64030769233089 -406.428359675480791885 6837.632716002934 6840.973851172274 +-3728 6 5831542.269248378 -3367.6517567568 5817556.0411483778 969592.67352472963333 2174330.2092403853 2381859.406131774 6 6984454.211097692 -4033.445769230769 6967702.8672438458471 1161283.811207307641183333 2604201.2704476737 2852759.5602156054 +-563 2 -515.6210729730 -3367.6517567568 -3883.2728297298 -1941.6364148649 1426.0153418918999 2016.6902366556308 2 -617.5607769230769 -4033.445769230769 -4651.0065461538459 -2325.50327307692295 1707.9424961538462 2415.395441814127 +6981 3 5831542.269248378 -515.6210729730 5830511.0271024320 1943503.67570081066667 2749258.455012492 3367140.1929065133 3 6984454.211097692 -617.5607769230769 6983219.0895438458462 2327739.696514615282066667 3292794.4113115156 4032833.0678006653 +626923679 1024 9723.4027027027 -9778.9513513514 10541.0525297287 10.29399661106318 5742.09145323734 5744.897264034267 1024 11645.746153846154 -11712.276923076923 12625.04759999997746 12.329148046874977988 6877.318722794877 6880.679250101603 +528534767 1024 5831542.269248378 -9777.1594594595 11646372.8607481068 11373.41099682432305 257528.92988206653 257654.7686043977 1024 6984454.211097692 -11710.130769230771 13948892.79980307629003 13621.965624807691689482 308443.1074570801 308593.82484083984 +NULL 3072 9318.4351351351 -4298.1513513514 5018444.1081079808 1633.60810810806667 5695.483082135364 5696.4103077145055 3072 11160.715384615385 -5147.907692307693 6010604.3076923073536 1956.576923076922966667 6821.495748565159 6822.606289190924 Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out?rev=1637445&r1=1637444&r2=1637445&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_mapjoin_reduce.q.out Fri Nov 7 20:42:43 2014 @@ -1,104 +1,6 @@ PREHOOK: query: -- SORT_QUERY_RESULTS -DROP TABLE part -PREHOOK: type: DROPTABLE -POSTHOOK: query: -- SORT_QUERY_RESULTS - -DROP TABLE part -POSTHOOK: type: DROPTABLE -PREHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part -POSTHOOK: query: -- data setup -CREATE TABLE part( - p_partkey INT, - p_name STRING, - p_mfgr STRING, - p_brand STRING, - p_type STRING, - p_size INT, - p_container STRING, - p_retailprice DOUBLE, - p_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@part -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/part_tiny.txt' overwrite into table part -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@part -PREHOOK: query: DROP TABLE lineitem -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE lineitem -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT, - L_PARTKEY INT, - L_SUPPKEY INT, - L_LINENUMBER INT, - L_QUANTITY DOUBLE, - L_EXTENDEDPRICE DOUBLE, - L_DISCOUNT DOUBLE, - L_TAX DOUBLE, - L_RETURNFLAG STRING, - L_LINESTATUS STRING, - l_shipdate STRING, - L_COMMITDATE STRING, - L_RECEIPTDATE STRING, - L_SHIPINSTRUCT STRING, - L_SHIPMODE STRING, - L_COMMENT STRING) -ROW FORMAT DELIMITED -FIELDS TERMINATED BY '|' -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@lineitem -POSTHOOK: query: CREATE TABLE lineitem (L_ORDERKEY INT, - L_PARTKEY INT, - L_SUPPKEY INT, - L_LINENUMBER INT, - L_QUANTITY DOUBLE, - L_EXTENDEDPRICE DOUBLE, - L_DISCOUNT DOUBLE, - L_TAX DOUBLE, - L_RETURNFLAG STRING, - L_LINESTATUS STRING, - l_shipdate STRING, - L_COMMITDATE STRING, - L_RECEIPTDATE STRING, - L_SHIPINSTRUCT STRING, - L_SHIPMODE STRING, - L_COMMENT STRING) -ROW FORMAT DELIMITED -FIELDS TERMINATED BY '|' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lineitem -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@lineitem -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE lineitem -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@lineitem -PREHOOK: query: -- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer. +-- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer. -- Query copied from subquery_in.q -- non agg, non corr, with join in Parent Query @@ -108,7 +10,9 @@ from (select distinct l_partkey as p_par where li.l_linenumber = 1 and li.l_orderkey in (select l_orderkey from lineitem where l_shipmode = 'AIR') PREHOOK: type: QUERY -POSTHOOK: query: -- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer. +POSTHOOK: query: -- SORT_QUERY_RESULTS + +-- Verify HIVE-8097 with a query that has a Vectorized MapJoin in the Reducer. -- Query copied from subquery_in.q -- non agg, non corr, with join in Parent Query @@ -135,60 +39,60 @@ STAGE PLANS: Map Operator Tree: TableScan alias: li - Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((l_partkey is not null and l_orderkey is not null) and (l_linenumber = 1)) (type: boolean) - Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: l_partkey (type: int) sort order: + Map-reduce partition columns: l_partkey (type: int) - Statistics: Num rows: 94 Data size: 1504 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12 Data size: 1439 Basic stats: COMPLETE Column stats: NONE value expressions: l_orderkey (type: int), l_suppkey (type: int) Map 4 Map Operator Tree: TableScan alias: lineitem - Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: l_partkey is not null (type: boolean) - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: l_partkey (type: int) outputColumnNames: l_partkey - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: l_partkey (type: int) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Map 6 Map Operator Tree: TableScan alias: lineitem - Statistics: Num rows: 1728 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((l_shipmode = 'AIR') and l_orderkey is not null) (type: boolean) - Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: l_orderkey (type: int) outputColumnNames: _col0 - Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: int) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 432 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Join Operator @@ -198,12 +102,12 @@ STAGE PLANS: 0 {KEY.reducesinkkey0} 1 {VALUE._col0} {VALUE._col1} outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col1 (type: int) sort order: + Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: int), _col3 (type: int) Reducer 3 Reduce Operator Tree: @@ -214,14 +118,14 @@ STAGE PLANS: 0 {VALUE._col0} {VALUE._col2} 1 outputColumnNames: _col0, _col3 - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col3 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -232,16 +136,16 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Stage: Stage-0 @@ -305,60 +209,60 @@ STAGE PLANS: Map Operator Tree: TableScan alias: li - Statistics: Num rows: 756 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (((l_partkey is not null and l_orderkey is not null) and l_linenumber is not null) and (l_linenumber = 1)) (type: boolean) - Statistics: Num rows: 47 Data size: 752 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: l_partkey (type: int) sort order: + Map-reduce partition columns: l_partkey (type: int) - Statistics: Num rows: 47 Data size: 752 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE value expressions: l_orderkey (type: int), l_suppkey (type: int) Map 4 Map Operator Tree: TableScan alias: lineitem - Statistics: Num rows: 3024 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: l_partkey is not null (type: boolean) - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: l_partkey (type: int) outputColumnNames: l_partkey - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: l_partkey (type: int) mode: hash outputColumnNames: _col0 - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1512 Data size: 6049 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 50 Data size: 5999 Basic stats: COMPLETE Column stats: NONE Map 6 Map Operator Tree: TableScan alias: lineitem - Statistics: Num rows: 1099 Data size: 12099 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 100 Data size: 11999 Basic stats: COMPLETE Column stats: NONE Filter Operator - predicate: (((l_shipmode = 'AIR') and l_orderkey is not null) and l_linenumber is not null) (type: boolean) - Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE Column stats: NONE + predicate: ((((l_shipmode = 'AIR') and l_orderkey is not null) and l_linenumber is not null) and (l_linenumber = 1)) (type: boolean) + Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: l_orderkey (type: int), l_linenumber (type: int) + expressions: l_orderkey (type: int), 1 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: int), _col1 (type: int) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: int) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 138 Data size: 1519 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6 Data size: 719 Basic stats: COMPLETE Column stats: NONE Reducer 2 Reduce Operator Tree: Join Operator @@ -368,12 +272,12 @@ STAGE PLANS: 0 {KEY.reducesinkkey0} 1 {VALUE._col0} {VALUE._col1} outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col1 (type: int), 1 (type: int) sort order: ++ Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 831 Data size: 3326 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27 Data size: 3298 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: int), _col3 (type: int) Reducer 3 Reduce Operator Tree: @@ -384,14 +288,14 @@ STAGE PLANS: 0 {VALUE._col0} {VALUE._col2} 1 outputColumnNames: _col0, _col3 - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col3 (type: int) outputColumnNames: _col0, _col1 - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 914 Data size: 3658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3627 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -402,16 +306,16 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 756 Data size: 3024 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 25 Data size: 2999 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Stage: Stage-0