HIVE-18759: Remove unconnected q.out-s (Zoltan Haindrich reviewed by Ashutosh Chauhan)
Signed-off-by: Zoltan Haindrich <k...@rxd.hu> Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/99380fbd Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/99380fbd Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/99380fbd Branch: refs/heads/master Commit: 99380fbd2682df2c02f375b96cd272dd7d913fde Parents: 987d5ff Author: Zoltan Haindrich <k...@rxd.hu> Authored: Tue Feb 27 07:47:07 2018 +0100 Committer: Zoltan Haindrich <k...@rxd.hu> Committed: Tue Feb 27 07:47:07 2018 +0100 ---------------------------------------------------------------------- .../alter_concatenate_indexed_table.q.out | 80 -- .../authorization_create_index.q.out | 9 - .../authorization_drop_index.q.out | 16 - .../authorization_uri_index.q.out | 9 - .../results/clientnegative/bad_indextype.q.out | 1 - .../clientnegative/drop_index_failure.q.out | 1 - .../clientnegative/merge_negative_1.q.out | 16 - .../show_create_table_index.q.out | 16 - .../clientnegative/temp_table_index.q.out | 12 - .../truncate_column_indexed_table.q.out | 26 - .../alter_concatenate_indexed_table.q.out | 271 ---- .../results/clientpositive/alter_index.q.out | 67 - .../clientpositive/authorization_index.q.out | 62 - .../results/clientpositive/drop_index.q.out | 10 - .../drop_index_removes_partition_dirs.q.out | 47 - .../clientpositive/drop_table_with_index.q.out | 153 --- .../clientpositive/llap/vector_const.q.out | 66 - .../llap/vector_empty_where.q.out | 652 --------- .../llap/vector_gather_stats.q.out | 108 -- .../clientpositive/llap/vector_join.q.out | 104 -- .../llap/vector_join_part_col_char.q.out | 175 --- .../llap/vector_non_constant_in_expr.q.out | 51 - .../llap/vector_non_string_partition.q.out | 274 ---- .../vector_orc_string_reader_empty_dict.q.out | 62 - .../llap/vector_outer_join_no_keys.q.out | 408 ------ .../llap/vector_tablesample_rows.q.out | 400 ------ .../llap/vectorization_limit.q.out | 943 ------------- .../llap/vectorization_parquet_projection.q.out | 684 ---------- .../llap/vectorized_mapjoin2.q.out | 214 --- .../show_indexes_edge_cases.q.out | 175 --- .../clientpositive/show_indexes_syntax.q.out | 117 -- .../spark/index_auto_self_join.q.out | 225 ---- .../clientpositive/spark/index_bitmap3.q.out | 1260 ----------------- .../spark/index_bitmap_auto.q.out | 1271 ------------------ .../results/clientpositive/udf_bitmap_and.q.out | 68 - .../clientpositive/udf_bitmap_empty.q.out | 18 - .../results/clientpositive/udf_bitmap_or.q.out | 68 - .../clientpositive/vector_string_decimal.q.out | 137 -- .../clientpositive/vector_udf_adaptor_1.q.out | 192 --- .../vector_udf_string_to_boolean.q.out | 189 --- .../clientpositive/vectorization_div0.q.out | 831 ------------ 41 files changed, 9488 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out b/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out deleted file mode 100644 index 19fbf51..0000000 --- a/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out +++ /dev/null @@ -1,80 +0,0 @@ -PREHOOK: query: create table src_rc_concatenate_test(key int, value string) stored as rcfile -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: create table src_rc_concatenate_test(key int, value string) stored as rcfile -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into table src_rc_concatenate_test -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into table src_rc_concatenate_test -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into table src_rc_concatenate_test -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into table src_rc_concatenate_test -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into table src_rc_concatenate_test -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into table src_rc_concatenate_test -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: show table extended like `src_rc_concatenate_test` -PREHOOK: type: SHOW_TABLESTATUS -POSTHOOK: query: show table extended like `src_rc_concatenate_test` -POSTHOOK: type: SHOW_TABLESTATUS -tableName:src_rc_concatenate_test -#### A masked pattern was here #### -inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat -outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat -columns:struct columns { i32 key, string value} -partitioned:false -partitionColumns: -totalNumberFiles:3 -totalFileSize:636 -maxFileSize:222 -minFileSize:206 -#### A masked pattern was here #### - -PREHOOK: query: select count(1) from src_rc_concatenate_test -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from src_rc_concatenate_test -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -15 -PREHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -214 -7678496319 -PREHOOK: query: create index src_rc_concatenate_test_index on table src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@src_rc_concatenate_test -POSTHOOK: query: create index src_rc_concatenate_test_index on table src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@src_rc_concatenate_test -POSTHOOK: Output: default@default__src_rc_concatenate_test_src_rc_concatenate_test_index__ -PREHOOK: query: show indexes on src_rc_concatenate_test -PREHOOK: type: SHOWINDEXES -POSTHOOK: query: show indexes on src_rc_concatenate_test -POSTHOOK: type: SHOWINDEXES -src_rc_concatenate_test_index src_rc_concatenate_test key default__src_rc_concatenate_test_src_rc_concatenate_test_index__ compact -FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: can not do merge because source table default.src_rc_concatenate_test is indexed. http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/authorization_create_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/authorization_create_index.q.out b/ql/src/test/results/clientnegative/authorization_create_index.q.out deleted file mode 100644 index 0e1f41e..0000000 --- a/ql/src/test/results/clientnegative/authorization_create_index.q.out +++ /dev/null @@ -1,9 +0,0 @@ -PREHOOK: query: create table t1 (a int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t1 -POSTHOOK: query: create table t1 (a int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t1 -FAILED: HiveAccessControlException Permission denied: Principal [name=user2, type=USER] does not have following privileges for operation CREATEINDEX [[OBJECT OWNERSHIP] on Object [type=TABLE_OR_VIEW, name=default.t1]] http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/authorization_drop_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/authorization_drop_index.q.out b/ql/src/test/results/clientnegative/authorization_drop_index.q.out deleted file mode 100644 index 72d782d..0000000 --- a/ql/src/test/results/clientnegative/authorization_drop_index.q.out +++ /dev/null @@ -1,16 +0,0 @@ -PREHOOK: query: create table t1 (a int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t1 -POSTHOOK: query: create table t1 (a int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t1 -PREHOOK: query: create index t1_index on table t1(a) as 'COMPACT' WITH DEFERRED REBUILD -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@t1 -POSTHOOK: query: create index t1_index on table t1(a) as 'COMPACT' WITH DEFERRED REBUILD -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@t1 -POSTHOOK: Output: default@default__t1_t1_index__ -FAILED: HiveAccessControlException Permission denied: Principal [name=user2, type=USER] does not have following privileges for operation DROPINDEX [[OBJECT OWNERSHIP] on Object [type=TABLE_OR_VIEW, name=default.t1]] http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/authorization_uri_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/authorization_uri_index.q.out b/ql/src/test/results/clientnegative/authorization_uri_index.q.out deleted file mode 100644 index 97b81b1..0000000 --- a/ql/src/test/results/clientnegative/authorization_uri_index.q.out +++ /dev/null @@ -1,9 +0,0 @@ -PREHOOK: query: create table t1(i int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t1 -POSTHOOK: query: create table t1(i int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t1 -#### A masked pattern was here #### http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/bad_indextype.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/bad_indextype.q.out b/ql/src/test/results/clientnegative/bad_indextype.q.out deleted file mode 100644 index 1ec59a7..0000000 --- a/ql/src/test/results/clientnegative/bad_indextype.q.out +++ /dev/null @@ -1 +0,0 @@ -FAILED: SemanticException class name provided for index handler not found. http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/drop_index_failure.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/drop_index_failure.q.out b/ql/src/test/results/clientnegative/drop_index_failure.q.out deleted file mode 100644 index f64ff5b..0000000 --- a/ql/src/test/results/clientnegative/drop_index_failure.q.out +++ /dev/null @@ -1 +0,0 @@ -FAILED: SemanticException [Error 10003]: Invalid index UnknownIndex http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/merge_negative_1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/merge_negative_1.q.out b/ql/src/test/results/clientnegative/merge_negative_1.q.out deleted file mode 100644 index 039a953..0000000 --- a/ql/src/test/results/clientnegative/merge_negative_1.q.out +++ /dev/null @@ -1,16 +0,0 @@ -PREHOOK: query: create table src2 like src -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@src2 -POSTHOOK: query: create table src2 like src -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@src2 -PREHOOK: query: CREATE INDEX src_index_merge_test ON TABLE src2(key) as 'COMPACT' WITH DEFERRED REBUILD -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@src2 -POSTHOOK: query: CREATE INDEX src_index_merge_test ON TABLE src2(key) as 'COMPACT' WITH DEFERRED REBUILD -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@src2 -POSTHOOK: Output: default@default__src2_src_index_merge_test__ -FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: can not do merge because source table default.src2 is indexed. http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/show_create_table_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/show_create_table_index.q.out b/ql/src/test/results/clientnegative/show_create_table_index.q.out deleted file mode 100644 index 7f03efd..0000000 --- a/ql/src/test/results/clientnegative/show_create_table_index.q.out +++ /dev/null @@ -1,16 +0,0 @@ -PREHOOK: query: CREATE TABLE tmp_showcrt (key int, value string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tmp_showcrt -POSTHOOK: query: CREATE TABLE tmp_showcrt (key int, value string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tmp_showcrt -PREHOOK: query: CREATE INDEX tmp_index on table tmp_showcrt(key) as 'compact' WITH DEFERRED REBUILD -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@tmp_showcrt -POSTHOOK: query: CREATE INDEX tmp_index on table tmp_showcrt(key) as 'compact' WITH DEFERRED REBUILD -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@tmp_showcrt -POSTHOOK: Output: default@default__tmp_showcrt_tmp_index__ -FAILED: SemanticException [Error 10144]: SHOW CREATE TABLE does not support tables of type INDEX_TABLE. default__tmp_showcrt_tmp_index__ has table type INDEX_TABLE http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/temp_table_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/temp_table_index.q.out b/ql/src/test/results/clientnegative/temp_table_index.q.out deleted file mode 100644 index 643d592..0000000 --- a/ql/src/test/results/clientnegative/temp_table_index.q.out +++ /dev/null @@ -1,12 +0,0 @@ -PREHOOK: query: create temporary table tmp1 (c1 string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tmp1 -POSTHOOK: query: create temporary table tmp1 (c1 string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tmp1 -PREHOOK: query: create index tmp1_idx on table tmp1 (c1) as 'COMPACT' with deferred rebuild -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@tmp1 -FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. org.apache.hadoop.hive.ql.metadata.HiveException: tableName=default.tmp1 is a TEMPORARY TABLE. Index on TEMPORARY TABLE is not supported. http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/truncate_column_indexed_table.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientnegative/truncate_column_indexed_table.q.out b/ql/src/test/results/clientnegative/truncate_column_indexed_table.q.out deleted file mode 100644 index 451fdba..0000000 --- a/ql/src/test/results/clientnegative/truncate_column_indexed_table.q.out +++ /dev/null @@ -1,26 +0,0 @@ -PREHOOK: query: CREATE TABLE test_tab (key STRING, value STRING) STORED AS RCFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@test_tab -POSTHOOK: query: CREATE TABLE test_tab (key STRING, value STRING) STORED AS RCFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@test_tab -PREHOOK: query: INSERT OVERWRITE TABLE test_tab SELECT * FROM src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@test_tab -POSTHOOK: query: INSERT OVERWRITE TABLE test_tab SELECT * FROM src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@test_tab -POSTHOOK: Lineage: test_tab.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: test_tab.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: CREATE INDEX test_tab_index ON TABLE test_tab (key) as 'COMPACT' WITH DEFERRED REBUILD -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@test_tab -POSTHOOK: query: CREATE INDEX test_tab_index ON TABLE test_tab (key) as 'COMPACT' WITH DEFERRED REBUILD -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@test_tab -POSTHOOK: Output: default@default__test_tab_test_tab_index__ -FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: Can not truncate columns from table with indexes http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/alter_concatenate_indexed_table.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/alter_concatenate_indexed_table.q.out b/ql/src/test/results/clientpositive/alter_concatenate_indexed_table.q.out deleted file mode 100644 index ffcbcf9..0000000 --- a/ql/src/test/results/clientpositive/alter_concatenate_indexed_table.q.out +++ /dev/null @@ -1,271 +0,0 @@ -PREHOOK: query: create table src_rc_concatenate_test(key int, value string) stored as rcfile -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: create table src_rc_concatenate_test(key int, value string) stored as rcfile -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into table src_rc_concatenate_test -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into table src_rc_concatenate_test -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into table src_rc_concatenate_test -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into table src_rc_concatenate_test -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into table src_rc_concatenate_test -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into table src_rc_concatenate_test -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: show table extended like `src_rc_concatenate_test` -PREHOOK: type: SHOW_TABLESTATUS -POSTHOOK: query: show table extended like `src_rc_concatenate_test` -POSTHOOK: type: SHOW_TABLESTATUS -tableName:src_rc_concatenate_test -#### A masked pattern was here #### -inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat -outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat -columns:struct columns { i32 key, string value} -partitioned:false -partitionColumns: -totalNumberFiles:3 -totalFileSize:636 -maxFileSize:222 -minFileSize:206 -#### A masked pattern was here #### - -PREHOOK: query: select count(1) from src_rc_concatenate_test -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from src_rc_concatenate_test -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -15 -PREHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -214 -7678496319 -PREHOOK: query: create index src_rc_concatenate_test_index on table src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@src_rc_concatenate_test -POSTHOOK: query: create index src_rc_concatenate_test_index on table src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@src_rc_concatenate_test -POSTHOOK: Output: default@default__src_rc_concatenate_test_src_rc_concatenate_test_index__ -PREHOOK: query: show indexes on src_rc_concatenate_test -PREHOOK: type: SHOWINDEXES -POSTHOOK: query: show indexes on src_rc_concatenate_test -POSTHOOK: type: SHOWINDEXES -src_rc_concatenate_test_index src_rc_concatenate_test key default__src_rc_concatenate_test_src_rc_concatenate_test_index__ compact -PREHOOK: query: alter table src_rc_concatenate_test concatenate -PREHOOK: type: ALTER_TABLE_MERGE -PREHOOK: Input: default@src_rc_concatenate_test -PREHOOK: Output: default@src_rc_concatenate_test -POSTHOOK: query: alter table src_rc_concatenate_test concatenate -POSTHOOK: type: ALTER_TABLE_MERGE -POSTHOOK: Input: default@src_rc_concatenate_test -POSTHOOK: Output: default@src_rc_concatenate_test -PREHOOK: query: show table extended like `src_rc_concatenate_test` -PREHOOK: type: SHOW_TABLESTATUS -POSTHOOK: query: show table extended like `src_rc_concatenate_test` -POSTHOOK: type: SHOW_TABLESTATUS -tableName:src_rc_concatenate_test -#### A masked pattern was here #### -inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat -outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat -columns:struct columns { i32 key, string value} -partitioned:false -partitionColumns: -totalNumberFiles:1 -totalFileSize:239 -maxFileSize:239 -minFileSize:239 -#### A masked pattern was here #### - -PREHOOK: query: select count(1) from src_rc_concatenate_test -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from src_rc_concatenate_test -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -15 -PREHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test -#### A masked pattern was here #### -214 -7678496319 -PREHOOK: query: drop index src_rc_concatenate_test_index on src_rc_concatenate_test -PREHOOK: type: DROPINDEX -PREHOOK: Input: default@src_rc_concatenate_test -POSTHOOK: query: drop index src_rc_concatenate_test_index on src_rc_concatenate_test -POSTHOOK: type: DROPINDEX -POSTHOOK: Input: default@src_rc_concatenate_test -PREHOOK: query: create table src_rc_concatenate_test_part(key int, value string) partitioned by (ds string) stored as rcfile -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@src_rc_concatenate_test_part -POSTHOOK: query: create table src_rc_concatenate_test_part(key int, value string) partitioned by (ds string) stored as rcfile -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@src_rc_concatenate_test_part -PREHOOK: query: alter table src_rc_concatenate_test_part add partition (ds='2011') -PREHOOK: type: ALTERTABLE_ADDPARTS -PREHOOK: Output: default@src_rc_concatenate_test_part -POSTHOOK: query: alter table src_rc_concatenate_test_part add partition (ds='2011') -POSTHOOK: type: ALTERTABLE_ADDPARTS -POSTHOOK: Output: default@src_rc_concatenate_test_part -POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -PREHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into table src_rc_concatenate_test_part partition (ds='2011') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into table src_rc_concatenate_test_part partition (ds='2011') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -PREHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into table src_rc_concatenate_test_part partition (ds='2011') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into table src_rc_concatenate_test_part partition (ds='2011') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -PREHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into table src_rc_concatenate_test_part partition (ds='2011') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -POSTHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into table src_rc_concatenate_test_part partition (ds='2011') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -PREHOOK: query: show table extended like `src_rc_concatenate_test_part` partition (ds='2011') -PREHOOK: type: SHOW_TABLESTATUS -POSTHOOK: query: show table extended like `src_rc_concatenate_test_part` partition (ds='2011') -POSTHOOK: type: SHOW_TABLESTATUS -tableName:src_rc_concatenate_test_part -#### A masked pattern was here #### -inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat -outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat -columns:struct columns { i32 key, string value} -partitioned:true -partitionColumns:struct partition_columns { string ds} -totalNumberFiles:3 -totalFileSize:636 -maxFileSize:222 -minFileSize:206 -#### A masked pattern was here #### - -PREHOOK: query: select count(1) from src_rc_concatenate_test_part -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test_part -PREHOOK: Input: default@src_rc_concatenate_test_part@ds=2011 -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from src_rc_concatenate_test_part -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test_part -POSTHOOK: Input: default@src_rc_concatenate_test_part@ds=2011 -#### A masked pattern was here #### -15 -PREHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test_part -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test_part -PREHOOK: Input: default@src_rc_concatenate_test_part@ds=2011 -#### A masked pattern was here #### -POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test_part -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test_part -POSTHOOK: Input: default@src_rc_concatenate_test_part@ds=2011 -#### A masked pattern was here #### -214 -7678496319 -PREHOOK: query: create index src_rc_concatenate_test_part_index on table src_rc_concatenate_test_part(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@src_rc_concatenate_test_part -POSTHOOK: query: create index src_rc_concatenate_test_part_index on table src_rc_concatenate_test_part(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@src_rc_concatenate_test_part -POSTHOOK: Output: default@default__src_rc_concatenate_test_part_src_rc_concatenate_test_part_index__ -PREHOOK: query: show indexes on src_rc_concatenate_test_part -PREHOOK: type: SHOWINDEXES -POSTHOOK: query: show indexes on src_rc_concatenate_test_part -POSTHOOK: type: SHOWINDEXES -src_rc_concatenate_test_part_index src_rc_concatenate_test_part key default__src_rc_concatenate_test_part_src_rc_concatenate_test_part_index__ compact -PREHOOK: query: alter table src_rc_concatenate_test_part partition (ds='2011') concatenate -PREHOOK: type: ALTER_PARTITION_MERGE -PREHOOK: Input: default@src_rc_concatenate_test_part -PREHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -POSTHOOK: query: alter table src_rc_concatenate_test_part partition (ds='2011') concatenate -POSTHOOK: type: ALTER_PARTITION_MERGE -POSTHOOK: Input: default@src_rc_concatenate_test_part -POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011 -PREHOOK: query: show table extended like `src_rc_concatenate_test_part` partition (ds='2011') -PREHOOK: type: SHOW_TABLESTATUS -POSTHOOK: query: show table extended like `src_rc_concatenate_test_part` partition (ds='2011') -POSTHOOK: type: SHOW_TABLESTATUS -tableName:src_rc_concatenate_test_part -#### A masked pattern was here #### -inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat -outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat -columns:struct columns { i32 key, string value} -partitioned:true -partitionColumns:struct partition_columns { string ds} -totalNumberFiles:1 -totalFileSize:239 -maxFileSize:239 -minFileSize:239 -#### A masked pattern was here #### - -PREHOOK: query: select count(1) from src_rc_concatenate_test_part -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test_part -PREHOOK: Input: default@src_rc_concatenate_test_part@ds=2011 -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from src_rc_concatenate_test_part -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test_part -POSTHOOK: Input: default@src_rc_concatenate_test_part@ds=2011 -#### A masked pattern was here #### -15 -PREHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test_part -PREHOOK: type: QUERY -PREHOOK: Input: default@src_rc_concatenate_test_part -PREHOOK: Input: default@src_rc_concatenate_test_part@ds=2011 -#### A masked pattern was here #### -POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from src_rc_concatenate_test_part -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_rc_concatenate_test_part -POSTHOOK: Input: default@src_rc_concatenate_test_part@ds=2011 -#### A masked pattern was here #### -214 -7678496319 -PREHOOK: query: drop index src_rc_concatenate_test_part_index on src_rc_concatenate_test_part -PREHOOK: type: DROPINDEX -PREHOOK: Input: default@src_rc_concatenate_test_part -POSTHOOK: query: drop index src_rc_concatenate_test_part_index on src_rc_concatenate_test_part -POSTHOOK: type: DROPINDEX -POSTHOOK: Input: default@src_rc_concatenate_test_part http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/alter_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/alter_index.q.out b/ql/src/test/results/clientpositive/alter_index.q.out deleted file mode 100644 index 262ad07..0000000 --- a/ql/src/test/results/clientpositive/alter_index.q.out +++ /dev/null @@ -1,67 +0,0 @@ -PREHOOK: query: drop index src_index_8 on src -PREHOOK: type: DROPINDEX -PREHOOK: Input: default@src -POSTHOOK: query: drop index src_index_8 on src -POSTHOOK: type: DROPINDEX -POSTHOOK: Input: default@src -PREHOOK: query: create index src_index_8 on table default.src(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@src -POSTHOOK: query: create index src_index_8 on table default.src(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@src -POSTHOOK: Output: default@default__src_src_index_8__ -PREHOOK: query: desc extended default__src_src_index_8__ -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@default__src_src_index_8__ -POSTHOOK: query: desc extended default__src_src_index_8__ -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@default__src_src_index_8__ -key string default -_bucketname string -_offsets array<bigint> - -#### A masked pattern was here #### -PREHOOK: query: alter index src_index_8 on default.src set IDXPROPERTIES ("prop1"="val1_new", "prop3"="val3") -PREHOOK: type: ALTERINDEX_PROPS -POSTHOOK: query: alter index src_index_8 on default.src set IDXPROPERTIES ("prop1"="val1_new", "prop3"="val3") -POSTHOOK: type: ALTERINDEX_PROPS -PREHOOK: query: desc extended default__src_src_index_8__ -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@default__src_src_index_8__ -POSTHOOK: query: desc extended default__src_src_index_8__ -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@default__src_src_index_8__ -key string default -_bucketname string -_offsets array<bigint> - -#### A masked pattern was here #### -PREHOOK: query: drop index src_index_8 on default.src -PREHOOK: type: DROPINDEX -PREHOOK: Input: default@src -POSTHOOK: query: drop index src_index_8 on default.src -POSTHOOK: type: DROPINDEX -POSTHOOK: Input: default@src -PREHOOK: query: show tables -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:default -POSTHOOK: query: show tables -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:default -alltypesorc -alltypesparquet -cbo_t1 -cbo_t2 -cbo_t3 -lineitem -part -src -src1 -src_cbo -src_json -src_sequencefile -src_thrift -srcbucket -srcbucket2 -srcpart http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/authorization_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/authorization_index.q.out b/ql/src/test/results/clientpositive/authorization_index.q.out deleted file mode 100644 index 87486b3..0000000 --- a/ql/src/test/results/clientpositive/authorization_index.q.out +++ /dev/null @@ -1,62 +0,0 @@ -PREHOOK: query: create table t1 (a int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t1 -POSTHOOK: query: create table t1 (a int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t1 -PREHOOK: query: create index t1_index on table t1(a) as 'COMPACT' WITH DEFERRED REBUILD -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@t1 -POSTHOOK: query: create index t1_index on table t1(a) as 'COMPACT' WITH DEFERRED REBUILD -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@t1 -POSTHOOK: Output: default@default__t1_t1_index__ -PREHOOK: query: desc formatted default__t1_t1_index__ -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@default__t1_t1_index__ -POSTHOOK: query: desc formatted default__t1_t1_index__ -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@default__t1_t1_index__ -# col_name data_type comment -a int -_bucketname string -_offsets array<bigint> - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Retention: 0 -#### A masked pattern was here #### -Table Type: INDEX_TABLE -Table Parameters: -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [Order(col:a, order:1)] -PREHOOK: query: alter index t1_index on t1 rebuild -PREHOOK: type: ALTERINDEX_REBUILD -PREHOOK: Input: default@t1 -PREHOOK: Output: default@default__t1_t1_index__ -POSTHOOK: query: alter index t1_index on t1 rebuild -POSTHOOK: type: ALTERINDEX_REBUILD -POSTHOOK: Input: default@t1 -POSTHOOK: Output: default@default__t1_t1_index__ -POSTHOOK: Lineage: default__t1_t1_index__._bucketname SIMPLE [(t1)t1.FieldSchema(name:INPUT__FILE__NAME, type:string, comment:), ] -POSTHOOK: Lineage: default__t1_t1_index__._offsets EXPRESSION [(t1)t1.FieldSchema(name:BLOCK__OFFSET__INSIDE__FILE, type:bigint, comment:), ] -POSTHOOK: Lineage: default__t1_t1_index__.a SIMPLE [(t1)t1.FieldSchema(name:a, type:int, comment:null), ] -PREHOOK: query: drop table t1 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@t1 -PREHOOK: Output: default@t1 -POSTHOOK: query: drop table t1 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@t1 -POSTHOOK: Output: default@t1 http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/drop_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/drop_index.q.out b/ql/src/test/results/clientpositive/drop_index.q.out deleted file mode 100644 index dc154d3..0000000 --- a/ql/src/test/results/clientpositive/drop_index.q.out +++ /dev/null @@ -1,10 +0,0 @@ -PREHOOK: query: DROP INDEX IF EXISTS UnknownIndex ON src -PREHOOK: type: DROPINDEX -PREHOOK: Input: default@src -POSTHOOK: query: DROP INDEX IF EXISTS UnknownIndex ON src -POSTHOOK: type: DROPINDEX -POSTHOOK: Input: default@src -PREHOOK: query: DROP INDEX IF EXISTS UnknownIndex ON UnknownTable -PREHOOK: type: DROPINDEX -POSTHOOK: query: DROP INDEX IF EXISTS UnknownIndex ON UnknownTable -POSTHOOK: type: DROPINDEX http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/drop_index_removes_partition_dirs.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/drop_index_removes_partition_dirs.q.out b/ql/src/test/results/clientpositive/drop_index_removes_partition_dirs.q.out deleted file mode 100644 index bbd86b4..0000000 --- a/ql/src/test/results/clientpositive/drop_index_removes_partition_dirs.q.out +++ /dev/null @@ -1,47 +0,0 @@ -PREHOOK: query: CREATE TABLE test_table (key STRING, value STRING) -PARTITIONED BY (part STRING) -STORED AS RCFILE -#### A masked pattern was here #### -PREHOOK: type: CREATETABLE -#### A masked pattern was here #### -PREHOOK: Output: database:default -PREHOOK: Output: default@test_table -POSTHOOK: query: CREATE TABLE test_table (key STRING, value STRING) -PARTITIONED BY (part STRING) -STORED AS RCFILE -#### A masked pattern was here #### -POSTHOOK: type: CREATETABLE -#### A masked pattern was here #### -POSTHOOK: Output: database:default -POSTHOOK: Output: default@test_table -PREHOOK: query: CREATE INDEX test_index ON -TABLE test_table(key) AS 'compact' WITH DEFERRED REBUILD -IN TABLE test_index_table -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@test_table -POSTHOOK: query: CREATE INDEX test_index ON -TABLE test_table(key) AS 'compact' WITH DEFERRED REBUILD -IN TABLE test_index_table -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@test_table -POSTHOOK: Output: default@test_index_table -PREHOOK: query: ALTER TABLE test_index_table ADD PARTITION (part = '1') -#### A masked pattern was here #### -PREHOOK: type: ALTERTABLE_ADDPARTS -#### A masked pattern was here #### -PREHOOK: Output: default@test_index_table -POSTHOOK: query: ALTER TABLE test_index_table ADD PARTITION (part = '1') -#### A masked pattern was here #### -POSTHOOK: type: ALTERTABLE_ADDPARTS -#### A masked pattern was here #### -POSTHOOK: Output: default@test_index_table -POSTHOOK: Output: default@test_index_table@part=1 -Found 1 items -#### A masked pattern was here #### -PREHOOK: query: DROP INDEX test_index ON test_table -PREHOOK: type: DROPINDEX -PREHOOK: Input: default@test_table -POSTHOOK: query: DROP INDEX test_index ON test_table -POSTHOOK: type: DROPINDEX -POSTHOOK: Input: default@test_table -#### A masked pattern was here #### http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/drop_table_with_index.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/drop_table_with_index.q.out b/ql/src/test/results/clientpositive/drop_table_with_index.q.out deleted file mode 100644 index 2312eee..0000000 --- a/ql/src/test/results/clientpositive/drop_table_with_index.q.out +++ /dev/null @@ -1,153 +0,0 @@ -PREHOOK: query: DROP TABLE IF EXISTS aa -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE IF EXISTS aa -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE aa (L_ORDERKEY INT, - L_PARTKEY INT, - L_SUPPKEY INT, - L_LINENUMBER INT, - L_QUANTITY DOUBLE, - L_EXTENDEDPRICE DOUBLE, - L_DISCOUNT DOUBLE, - L_TAX DOUBLE, - L_RETURNFLAG STRING, - L_LINESTATUS STRING, - l_shipdate STRING, - L_COMMITDATE STRING, - L_RECEIPTDATE STRING, - L_SHIPINSTRUCT STRING, - L_SHIPMODE STRING, - L_COMMENT STRING) -ROW FORMAT DELIMITED -FIELDS TERMINATED BY '|' -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@aa -POSTHOOK: query: CREATE TABLE aa (L_ORDERKEY INT, - L_PARTKEY INT, - L_SUPPKEY INT, - L_LINENUMBER INT, - L_QUANTITY DOUBLE, - L_EXTENDEDPRICE DOUBLE, - L_DISCOUNT DOUBLE, - L_TAX DOUBLE, - L_RETURNFLAG STRING, - L_LINESTATUS STRING, - l_shipdate STRING, - L_COMMITDATE STRING, - L_RECEIPTDATE STRING, - L_SHIPINSTRUCT STRING, - L_SHIPMODE STRING, - L_COMMENT STRING) -ROW FORMAT DELIMITED -FIELDS TERMINATED BY '|' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@aa -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE aa -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@aa -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' OVERWRITE INTO TABLE aa -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@aa -PREHOOK: query: CREATE INDEX aa_lshipdate_idx ON TABLE aa(l_shipdate) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(l_shipdate)") -PREHOOK: type: CREATEINDEX -PREHOOK: Input: default@aa -POSTHOOK: query: CREATE INDEX aa_lshipdate_idx ON TABLE aa(l_shipdate) AS 'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD IDXPROPERTIES("AGGREGATES"="count(l_shipdate)") -POSTHOOK: type: CREATEINDEX -POSTHOOK: Input: default@aa -POSTHOOK: Output: default@default__aa_aa_lshipdate_idx__ -PREHOOK: query: ALTER INDEX aa_lshipdate_idx ON aa REBUILD -PREHOOK: type: ALTERINDEX_REBUILD -PREHOOK: Input: default@aa -PREHOOK: Output: default@default__aa_aa_lshipdate_idx__ -POSTHOOK: query: ALTER INDEX aa_lshipdate_idx ON aa REBUILD -POSTHOOK: type: ALTERINDEX_REBUILD -POSTHOOK: Input: default@aa -POSTHOOK: Output: default@default__aa_aa_lshipdate_idx__ -POSTHOOK: Lineage: default__aa_aa_lshipdate_idx__._bucketname SIMPLE [(aa)aa.FieldSchema(name:INPUT__FILE__NAME, type:string, comment:), ] -POSTHOOK: Lineage: default__aa_aa_lshipdate_idx__._count_of_l_shipdate EXPRESSION [(aa)aa.FieldSchema(name:l_shipdate, type:string, comment:null), ] -POSTHOOK: Lineage: default__aa_aa_lshipdate_idx__._offsets EXPRESSION [(aa)aa.FieldSchema(name:BLOCK__OFFSET__INSIDE__FILE, type:bigint, comment:), ] -POSTHOOK: Lineage: default__aa_aa_lshipdate_idx__.l_shipdate SIMPLE [(aa)aa.FieldSchema(name:l_shipdate, type:string, comment:null), ] -PREHOOK: query: show tables -PREHOOK: type: SHOWTABLES -PREHOOK: Input: database:default -POSTHOOK: query: show tables -POSTHOOK: type: SHOWTABLES -POSTHOOK: Input: database:default -aa -alltypesorc -alltypesparquet -cbo_t1 -cbo_t2 -cbo_t3 -default__aa_aa_lshipdate_idx__ -lineitem -part -src -src1 -src_cbo -src_json -src_sequencefile -src_thrift -srcbucket -srcbucket2 -srcpart -PREHOOK: query: explain select l_shipdate, count(l_shipdate) -from aa -group by l_shipdate -PREHOOK: type: QUERY -POSTHOOK: query: explain select l_shipdate, count(l_shipdate) -from aa -group by l_shipdate -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: aa - Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: l_shipdate (type: string) - outputColumnNames: l_shipdate - Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(l_shipdate) - keys: l_shipdate (type: string) - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_const.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/vector_const.q.out b/ql/src/test/results/clientpositive/llap/vector_const.q.out deleted file mode 100644 index 964ddcc..0000000 --- a/ql/src/test/results/clientpositive/llap/vector_const.q.out +++ /dev/null @@ -1,66 +0,0 @@ -PREHOOK: query: CREATE TEMPORARY TABLE varchar_const_1 (c1 int) STORED AS ORC -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@varchar_const_1 -POSTHOOK: query: CREATE TEMPORARY TABLE varchar_const_1 (c1 int) STORED AS ORC -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@varchar_const_1 -PREHOOK: query: INSERT INTO varchar_const_1 values(42) -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@varchar_const_1 -POSTHOOK: query: INSERT INTO varchar_const_1 values(42) -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@varchar_const_1 -POSTHOOK: Lineage: varchar_const_1.c1 SCRIPT [] -PREHOOK: query: EXPLAIN -SELECT CONCAT(CAST('F' AS CHAR(2)), CAST('F' AS VARCHAR(2))) FROM VARCHAR_CONST_1 -PREHOOK: type: QUERY -POSTHOOK: query: EXPLAIN -SELECT CONCAT(CAST('F' AS CHAR(2)), CAST('F' AS VARCHAR(2))) FROM VARCHAR_CONST_1 -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: varchar_const_1 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'FF' (type: varchar(4)) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 86 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 86 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized, llap - LLAP IO: all inputs - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT CONCAT(CAST('F' AS CHAR(2)), CAST('F' AS VARCHAR(2))) FROM VARCHAR_CONST_1 -PREHOOK: type: QUERY -PREHOOK: Input: default@varchar_const_1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT CONCAT(CAST('F' AS CHAR(2)), CAST('F' AS VARCHAR(2))) FROM VARCHAR_CONST_1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@varchar_const_1 -#### A masked pattern was here #### -FF http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_empty_where.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/llap/vector_empty_where.q.out b/ql/src/test/results/clientpositive/llap/vector_empty_where.q.out deleted file mode 100644 index 494c5c9..0000000 --- a/ql/src/test/results/clientpositive/llap/vector_empty_where.q.out +++ /dev/null @@ -1,652 +0,0 @@ -PREHOOK: query: explain vectorization expression -select count (distinct cint) from alltypesorc where cstring1 -PREHOOK: type: QUERY -POSTHOOK: query: explain vectorization expression -select count (distinct cint) from alltypesorc where cstring1 -POSTHOOK: type: QUERY -PLAN VECTORIZATION: - enabled: true - enabledConditionsMet: [hive.vectorized.execution.enabled IS true] - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: alltypesorc - Statistics: Num rows: 12288 Data size: 899146 Basic stats: COMPLETE Column stats: COMPLETE - TableScan Vectorization: - native: true - Filter Operator - Filter Vectorization: - className: VectorFilterOperator - native: true - predicateExpression: SelectColumnIsTrue(col 13:boolean)(children: CastStringToBoolean(col 6) -> 13:boolean) - predicate: cstring1 (type: string) - Statistics: Num rows: 6144 Data size: 449620 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: cint (type: int) - outputColumnNames: cint - Select Vectorization: - className: VectorSelectOperator - native: true - projectedOutputColumnNums: [2] - Statistics: Num rows: 6144 Data size: 449620 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - Group By Vectorization: - className: VectorGroupByOperator - groupByMode: HASH - keyExpressions: col 2:int - native: false - vectorProcessingMode: HASH - projectedOutputColumnNums: [] - keys: cint (type: int) - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - sort order: + - Map-reduce partition columns: _col0 (type: int) - Reduce Sink Vectorization: - className: VectorReduceSinkLongOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized, llap - LLAP IO: all inputs - Map Vectorization: - enabled: true - enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true - inputFormatFeatureSupport: [] - featureSupportInUse: [] - inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reducer 2 - Execution mode: vectorized, llap - Reduce Vectorization: - enabled: true - enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reduce Operator Tree: - Group By Operator - Group By Vectorization: - className: VectorGroupByOperator - groupByMode: MERGEPARTIAL - keyExpressions: col 0:int - native: false - vectorProcessingMode: MERGE_PARTIAL - projectedOutputColumnNums: [] - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col0) - Group By Vectorization: - aggregators: VectorUDAFCount(col 0:int) -> bigint - className: VectorGroupByOperator - groupByMode: HASH - native: false - vectorProcessingMode: HASH - projectedOutputColumnNums: [0] - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - sort order: - Reduce Sink Vectorization: - className: VectorReduceSinkEmptyKeyOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Reducer 3 - Execution mode: vectorized, llap - Reduce Vectorization: - enabled: true - enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - Group By Vectorization: - aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint - className: VectorGroupByOperator - groupByMode: MERGEPARTIAL - native: false - vectorProcessingMode: GLOBAL - projectedOutputColumnNums: [0] - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - File Sink Vectorization: - className: VectorFileSinkOperator - native: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count (distinct cint) from alltypesorc where cstring1 -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: select count (distinct cint) from alltypesorc where cstring1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -6041 -PREHOOK: query: explain vectorization expression -select count (distinct cint) from alltypesorc where cint -PREHOOK: type: QUERY -POSTHOOK: query: explain vectorization expression -select count (distinct cint) from alltypesorc where cint -POSTHOOK: type: QUERY -PLAN VECTORIZATION: - enabled: true - enabledConditionsMet: [hive.vectorized.execution.enabled IS true] - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: alltypesorc - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - TableScan Vectorization: - native: true - Filter Operator - Filter Vectorization: - className: VectorFilterOperator - native: true - predicateExpression: SelectColumnIsTrue(col 13:boolean)(children: CastLongToBooleanViaLongToLong(col 2:int) -> 13:boolean) - predicate: cint (type: int) - Statistics: Num rows: 6144 Data size: 18348 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - Group By Vectorization: - className: VectorGroupByOperator - groupByMode: HASH - keyExpressions: col 2:int - native: false - vectorProcessingMode: HASH - projectedOutputColumnNums: [] - keys: cint (type: int) - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - sort order: + - Map-reduce partition columns: _col0 (type: int) - Reduce Sink Vectorization: - className: VectorReduceSinkLongOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized, llap - LLAP IO: all inputs - Map Vectorization: - enabled: true - enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true - inputFormatFeatureSupport: [] - featureSupportInUse: [] - inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reducer 2 - Execution mode: vectorized, llap - Reduce Vectorization: - enabled: true - enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reduce Operator Tree: - Group By Operator - Group By Vectorization: - className: VectorGroupByOperator - groupByMode: MERGEPARTIAL - keyExpressions: col 0:int - native: false - vectorProcessingMode: MERGE_PARTIAL - projectedOutputColumnNums: [] - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col0) - Group By Vectorization: - aggregators: VectorUDAFCount(col 0:int) -> bigint - className: VectorGroupByOperator - groupByMode: HASH - native: false - vectorProcessingMode: HASH - projectedOutputColumnNums: [0] - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - sort order: - Reduce Sink Vectorization: - className: VectorReduceSinkEmptyKeyOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Reducer 3 - Execution mode: vectorized, llap - Reduce Vectorization: - enabled: true - enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - Group By Vectorization: - aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint - className: VectorGroupByOperator - groupByMode: MERGEPARTIAL - native: false - vectorProcessingMode: GLOBAL - projectedOutputColumnNums: [0] - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - File Sink Vectorization: - className: VectorFileSinkOperator - native: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count (distinct cint) from alltypesorc where cint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: select count (distinct cint) from alltypesorc where cint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -6082 -PREHOOK: query: explain vectorization expression -select count (distinct cint) from alltypesorc where cfloat -PREHOOK: type: QUERY -POSTHOOK: query: explain vectorization expression -select count (distinct cint) from alltypesorc where cfloat -POSTHOOK: type: QUERY -PLAN VECTORIZATION: - enabled: true - enabledConditionsMet: [hive.vectorized.execution.enabled IS true] - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: alltypesorc - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - TableScan Vectorization: - native: true - Filter Operator - Filter Vectorization: - className: VectorFilterOperator - native: true - predicateExpression: SelectColumnIsTrue(col 13:boolean)(children: CastDoubleToBooleanViaDoubleToLong(col 4:float) -> 13:boolean) - predicate: cfloat (type: float) - Statistics: Num rows: 6144 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: cint (type: int) - outputColumnNames: cint - Select Vectorization: - className: VectorSelectOperator - native: true - projectedOutputColumnNums: [2] - Statistics: Num rows: 6144 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - Group By Vectorization: - className: VectorGroupByOperator - groupByMode: HASH - keyExpressions: col 2:int - native: false - vectorProcessingMode: HASH - projectedOutputColumnNums: [] - keys: cint (type: int) - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - sort order: + - Map-reduce partition columns: _col0 (type: int) - Reduce Sink Vectorization: - className: VectorReduceSinkLongOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized, llap - LLAP IO: all inputs - Map Vectorization: - enabled: true - enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true - inputFormatFeatureSupport: [] - featureSupportInUse: [] - inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reducer 2 - Execution mode: vectorized, llap - Reduce Vectorization: - enabled: true - enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reduce Operator Tree: - Group By Operator - Group By Vectorization: - className: VectorGroupByOperator - groupByMode: MERGEPARTIAL - keyExpressions: col 0:int - native: false - vectorProcessingMode: MERGE_PARTIAL - projectedOutputColumnNums: [] - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col0) - Group By Vectorization: - aggregators: VectorUDAFCount(col 0:int) -> bigint - className: VectorGroupByOperator - groupByMode: HASH - native: false - vectorProcessingMode: HASH - projectedOutputColumnNums: [0] - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - sort order: - Reduce Sink Vectorization: - className: VectorReduceSinkEmptyKeyOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Reducer 3 - Execution mode: vectorized, llap - Reduce Vectorization: - enabled: true - enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - Group By Vectorization: - aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint - className: VectorGroupByOperator - groupByMode: MERGEPARTIAL - native: false - vectorProcessingMode: GLOBAL - projectedOutputColumnNums: [0] - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - File Sink Vectorization: - className: VectorFileSinkOperator - native: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count (distinct cint) from alltypesorc where cfloat -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: select count (distinct cint) from alltypesorc where cfloat -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -3022 -PREHOOK: query: explain vectorization expression -select count (distinct cint) from alltypesorc where ctimestamp1 -PREHOOK: type: QUERY -POSTHOOK: query: explain vectorization expression -select count (distinct cint) from alltypesorc where ctimestamp1 -POSTHOOK: type: QUERY -PLAN VECTORIZATION: - enabled: true - enabledConditionsMet: [hive.vectorized.execution.enabled IS true] - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Tez -#### A masked pattern was here #### - Edges: - Reducer 2 <- Map 1 (SIMPLE_EDGE) - Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) -#### A masked pattern was here #### - Vertices: - Map 1 - Map Operator Tree: - TableScan - alias: alltypesorc - Statistics: Num rows: 12288 Data size: 528216 Basic stats: COMPLETE Column stats: COMPLETE - TableScan Vectorization: - native: true - Filter Operator - Filter Vectorization: - className: VectorFilterOperator - native: true - predicateExpression: SelectColumnIsTrue(col 13:boolean)(children: CastTimestampToBoolean(col 8:timestamp) -> 13:boolean) - predicate: ctimestamp1 (type: timestamp) - Statistics: Num rows: 6144 Data size: 264108 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: cint (type: int) - outputColumnNames: cint - Select Vectorization: - className: VectorSelectOperator - native: true - projectedOutputColumnNums: [2] - Statistics: Num rows: 6144 Data size: 264108 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - Group By Vectorization: - className: VectorGroupByOperator - groupByMode: HASH - keyExpressions: col 2:int - native: false - vectorProcessingMode: HASH - projectedOutputColumnNums: [] - keys: cint (type: int) - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - sort order: + - Map-reduce partition columns: _col0 (type: int) - Reduce Sink Vectorization: - className: VectorReduceSinkLongOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized, llap - LLAP IO: all inputs - Map Vectorization: - enabled: true - enabledConditionsMet: hive.vectorized.use.vectorized.input.format IS true - inputFormatFeatureSupport: [] - featureSupportInUse: [] - inputFileFormats: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reducer 2 - Execution mode: vectorized, llap - Reduce Vectorization: - enabled: true - enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reduce Operator Tree: - Group By Operator - Group By Vectorization: - className: VectorGroupByOperator - groupByMode: MERGEPARTIAL - keyExpressions: col 0:int - native: false - vectorProcessingMode: MERGE_PARTIAL - projectedOutputColumnNums: [] - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 3016 Data size: 9008 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col0) - Group By Vectorization: - aggregators: VectorUDAFCount(col 0:int) -> bigint - className: VectorGroupByOperator - groupByMode: HASH - native: false - vectorProcessingMode: HASH - projectedOutputColumnNums: [0] - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - sort order: - Reduce Sink Vectorization: - className: VectorReduceSinkEmptyKeyOperator - native: true - nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Reducer 3 - Execution mode: vectorized, llap - Reduce Vectorization: - enabled: true - enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - allNative: false - usesVectorUDFAdaptor: false - vectorized: true - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - Group By Vectorization: - aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint - className: VectorGroupByOperator - groupByMode: MERGEPARTIAL - native: false - vectorProcessingMode: GLOBAL - projectedOutputColumnNums: [0] - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - File Sink Vectorization: - className: VectorFileSinkOperator - native: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count (distinct cint) from alltypesorc where ctimestamp1 -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: select count (distinct cint) from alltypesorc where ctimestamp1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -3022