HIVE-18759: Remove unconnected q.out-s (Zoltan Haindrich reviewed by Ashutosh 
Chauhan)

Signed-off-by: Zoltan Haindrich <k...@rxd.hu>


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/99380fbd
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/99380fbd
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/99380fbd

Branch: refs/heads/master
Commit: 99380fbd2682df2c02f375b96cd272dd7d913fde
Parents: 987d5ff
Author: Zoltan Haindrich <k...@rxd.hu>
Authored: Tue Feb 27 07:47:07 2018 +0100
Committer: Zoltan Haindrich <k...@rxd.hu>
Committed: Tue Feb 27 07:47:07 2018 +0100

----------------------------------------------------------------------
 .../alter_concatenate_indexed_table.q.out       |   80 --
 .../authorization_create_index.q.out            |    9 -
 .../authorization_drop_index.q.out              |   16 -
 .../authorization_uri_index.q.out               |    9 -
 .../results/clientnegative/bad_indextype.q.out  |    1 -
 .../clientnegative/drop_index_failure.q.out     |    1 -
 .../clientnegative/merge_negative_1.q.out       |   16 -
 .../show_create_table_index.q.out               |   16 -
 .../clientnegative/temp_table_index.q.out       |   12 -
 .../truncate_column_indexed_table.q.out         |   26 -
 .../alter_concatenate_indexed_table.q.out       |  271 ----
 .../results/clientpositive/alter_index.q.out    |   67 -
 .../clientpositive/authorization_index.q.out    |   62 -
 .../results/clientpositive/drop_index.q.out     |   10 -
 .../drop_index_removes_partition_dirs.q.out     |   47 -
 .../clientpositive/drop_table_with_index.q.out  |  153 ---
 .../clientpositive/llap/vector_const.q.out      |   66 -
 .../llap/vector_empty_where.q.out               |  652 ---------
 .../llap/vector_gather_stats.q.out              |  108 --
 .../clientpositive/llap/vector_join.q.out       |  104 --
 .../llap/vector_join_part_col_char.q.out        |  175 ---
 .../llap/vector_non_constant_in_expr.q.out      |   51 -
 .../llap/vector_non_string_partition.q.out      |  274 ----
 .../vector_orc_string_reader_empty_dict.q.out   |   62 -
 .../llap/vector_outer_join_no_keys.q.out        |  408 ------
 .../llap/vector_tablesample_rows.q.out          |  400 ------
 .../llap/vectorization_limit.q.out              |  943 -------------
 .../llap/vectorization_parquet_projection.q.out |  684 ----------
 .../llap/vectorized_mapjoin2.q.out              |  214 ---
 .../show_indexes_edge_cases.q.out               |  175 ---
 .../clientpositive/show_indexes_syntax.q.out    |  117 --
 .../spark/index_auto_self_join.q.out            |  225 ----
 .../clientpositive/spark/index_bitmap3.q.out    | 1260 -----------------
 .../spark/index_bitmap_auto.q.out               | 1271 ------------------
 .../results/clientpositive/udf_bitmap_and.q.out |   68 -
 .../clientpositive/udf_bitmap_empty.q.out       |   18 -
 .../results/clientpositive/udf_bitmap_or.q.out  |   68 -
 .../clientpositive/vector_string_decimal.q.out  |  137 --
 .../clientpositive/vector_udf_adaptor_1.q.out   |  192 ---
 .../vector_udf_string_to_boolean.q.out          |  189 ---
 .../clientpositive/vectorization_div0.q.out     |  831 ------------
 41 files changed, 9488 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out 
b/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out
deleted file mode 100644
index 19fbf51..0000000
--- a/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out
+++ /dev/null
@@ -1,80 +0,0 @@
-PREHOOK: query: create table src_rc_concatenate_test(key int, value string) 
stored as rcfile
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: create table src_rc_concatenate_test(key int, value string) 
stored as rcfile
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into 
table src_rc_concatenate_test
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into 
table src_rc_concatenate_test
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into 
table src_rc_concatenate_test
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into 
table src_rc_concatenate_test
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into 
table src_rc_concatenate_test
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into 
table src_rc_concatenate_test
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: show table extended like `src_rc_concatenate_test`
-PREHOOK: type: SHOW_TABLESTATUS
-POSTHOOK: query: show table extended like `src_rc_concatenate_test`
-POSTHOOK: type: SHOW_TABLESTATUS
-tableName:src_rc_concatenate_test
-#### A masked pattern was here ####
-inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat
-outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-columns:struct columns { i32 key, string value}
-partitioned:false
-partitionColumns:
-totalNumberFiles:3
-totalFileSize:636
-maxFileSize:222
-minFileSize:206
-#### A masked pattern was here ####
-
-PREHOOK: query: select count(1) from src_rc_concatenate_test
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-POSTHOOK: query: select count(1) from src_rc_concatenate_test
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-15
-PREHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-214    -7678496319
-PREHOOK: query: create index src_rc_concatenate_test_index on table 
src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES 
("prop1"="val1", "prop2"="val2")
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@src_rc_concatenate_test
-POSTHOOK: query: create index src_rc_concatenate_test_index on table 
src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES 
("prop1"="val1", "prop2"="val2")
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@src_rc_concatenate_test
-POSTHOOK: Output: 
default@default__src_rc_concatenate_test_src_rc_concatenate_test_index__
-PREHOOK: query: show indexes on src_rc_concatenate_test
-PREHOOK: type: SHOWINDEXES
-POSTHOOK: query: show indexes on src_rc_concatenate_test
-POSTHOOK: type: SHOWINDEXES
-src_rc_concatenate_test_index  src_rc_concatenate_test key                     
default__src_rc_concatenate_test_src_rc_concatenate_test_index__        compact 
                
-FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: 
can not do merge because source table default.src_rc_concatenate_test is 
indexed.

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/authorization_create_index.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientnegative/authorization_create_index.q.out 
b/ql/src/test/results/clientnegative/authorization_create_index.q.out
deleted file mode 100644
index 0e1f41e..0000000
--- a/ql/src/test/results/clientnegative/authorization_create_index.q.out
+++ /dev/null
@@ -1,9 +0,0 @@
-PREHOOK: query: create table t1 (a int)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@t1
-POSTHOOK: query: create table t1 (a int)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@t1
-FAILED: HiveAccessControlException Permission denied: Principal [name=user2, 
type=USER] does not have following privileges for operation CREATEINDEX 
[[OBJECT OWNERSHIP] on Object [type=TABLE_OR_VIEW, name=default.t1]]

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/authorization_drop_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/authorization_drop_index.q.out 
b/ql/src/test/results/clientnegative/authorization_drop_index.q.out
deleted file mode 100644
index 72d782d..0000000
--- a/ql/src/test/results/clientnegative/authorization_drop_index.q.out
+++ /dev/null
@@ -1,16 +0,0 @@
-PREHOOK: query: create table t1 (a int)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@t1
-POSTHOOK: query: create table t1 (a int)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@t1
-PREHOOK: query: create index t1_index on table t1(a) as 'COMPACT' WITH 
DEFERRED REBUILD
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@t1
-POSTHOOK: query: create index t1_index on table t1(a) as 'COMPACT' WITH 
DEFERRED REBUILD
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@t1
-POSTHOOK: Output: default@default__t1_t1_index__
-FAILED: HiveAccessControlException Permission denied: Principal [name=user2, 
type=USER] does not have following privileges for operation DROPINDEX [[OBJECT 
OWNERSHIP] on Object [type=TABLE_OR_VIEW, name=default.t1]]

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/authorization_uri_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/authorization_uri_index.q.out 
b/ql/src/test/results/clientnegative/authorization_uri_index.q.out
deleted file mode 100644
index 97b81b1..0000000
--- a/ql/src/test/results/clientnegative/authorization_uri_index.q.out
+++ /dev/null
@@ -1,9 +0,0 @@
-PREHOOK: query: create table t1(i int)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@t1
-POSTHOOK: query: create table t1(i int)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@t1
-#### A masked pattern was here ####

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/bad_indextype.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/bad_indextype.q.out 
b/ql/src/test/results/clientnegative/bad_indextype.q.out
deleted file mode 100644
index 1ec59a7..0000000
--- a/ql/src/test/results/clientnegative/bad_indextype.q.out
+++ /dev/null
@@ -1 +0,0 @@
-FAILED: SemanticException class name provided for index handler not found.

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/drop_index_failure.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/drop_index_failure.q.out 
b/ql/src/test/results/clientnegative/drop_index_failure.q.out
deleted file mode 100644
index f64ff5b..0000000
--- a/ql/src/test/results/clientnegative/drop_index_failure.q.out
+++ /dev/null
@@ -1 +0,0 @@
-FAILED: SemanticException [Error 10003]: Invalid index UnknownIndex

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/merge_negative_1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/merge_negative_1.q.out 
b/ql/src/test/results/clientnegative/merge_negative_1.q.out
deleted file mode 100644
index 039a953..0000000
--- a/ql/src/test/results/clientnegative/merge_negative_1.q.out
+++ /dev/null
@@ -1,16 +0,0 @@
-PREHOOK: query: create table src2 like src
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@src2
-POSTHOOK: query: create table src2 like src
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@src2
-PREHOOK: query: CREATE INDEX src_index_merge_test ON TABLE src2(key) as 
'COMPACT' WITH DEFERRED REBUILD
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@src2
-POSTHOOK: query: CREATE INDEX src_index_merge_test ON TABLE src2(key) as 
'COMPACT' WITH DEFERRED REBUILD
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@src2
-POSTHOOK: Output: default@default__src2_src_index_merge_test__
-FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: 
can not do merge because source table default.src2 is indexed.

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/show_create_table_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/show_create_table_index.q.out 
b/ql/src/test/results/clientnegative/show_create_table_index.q.out
deleted file mode 100644
index 7f03efd..0000000
--- a/ql/src/test/results/clientnegative/show_create_table_index.q.out
+++ /dev/null
@@ -1,16 +0,0 @@
-PREHOOK: query: CREATE TABLE tmp_showcrt (key int, value string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tmp_showcrt
-POSTHOOK: query: CREATE TABLE tmp_showcrt (key int, value string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tmp_showcrt
-PREHOOK: query: CREATE INDEX tmp_index on table tmp_showcrt(key) as 'compact' 
WITH DEFERRED REBUILD
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@tmp_showcrt
-POSTHOOK: query: CREATE INDEX tmp_index on table tmp_showcrt(key) as 'compact' 
WITH DEFERRED REBUILD
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@tmp_showcrt
-POSTHOOK: Output: default@default__tmp_showcrt_tmp_index__
-FAILED: SemanticException [Error 10144]: SHOW CREATE TABLE does not support 
tables of type INDEX_TABLE. default__tmp_showcrt_tmp_index__ has table type 
INDEX_TABLE

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/temp_table_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientnegative/temp_table_index.q.out 
b/ql/src/test/results/clientnegative/temp_table_index.q.out
deleted file mode 100644
index 643d592..0000000
--- a/ql/src/test/results/clientnegative/temp_table_index.q.out
+++ /dev/null
@@ -1,12 +0,0 @@
-PREHOOK: query: create temporary table tmp1 (c1 string)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@tmp1
-POSTHOOK: query: create temporary table tmp1 (c1 string)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@tmp1
-PREHOOK: query: create index tmp1_idx on table tmp1 (c1) as 'COMPACT' with 
deferred rebuild
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@tmp1
-FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.DDLTask. 
org.apache.hadoop.hive.ql.metadata.HiveException: tableName=default.tmp1 is a 
TEMPORARY TABLE. Index on TEMPORARY TABLE is not supported.

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientnegative/truncate_column_indexed_table.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientnegative/truncate_column_indexed_table.q.out 
b/ql/src/test/results/clientnegative/truncate_column_indexed_table.q.out
deleted file mode 100644
index 451fdba..0000000
--- a/ql/src/test/results/clientnegative/truncate_column_indexed_table.q.out
+++ /dev/null
@@ -1,26 +0,0 @@
-PREHOOK: query: CREATE TABLE test_tab (key STRING, value STRING) STORED AS 
RCFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@test_tab
-POSTHOOK: query: CREATE TABLE test_tab (key STRING, value STRING) STORED AS 
RCFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@test_tab
-PREHOOK: query: INSERT OVERWRITE TABLE test_tab SELECT * FROM src
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-PREHOOK: Output: default@test_tab
-POSTHOOK: query: INSERT OVERWRITE TABLE test_tab SELECT * FROM src
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@test_tab
-POSTHOOK: Lineage: test_tab.key SIMPLE [(src)src.FieldSchema(name:key, 
type:string, comment:default), ]
-POSTHOOK: Lineage: test_tab.value SIMPLE [(src)src.FieldSchema(name:value, 
type:string, comment:default), ]
-PREHOOK: query: CREATE INDEX test_tab_index ON TABLE test_tab (key) as 
'COMPACT' WITH DEFERRED REBUILD
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@test_tab
-POSTHOOK: query: CREATE INDEX test_tab_index ON TABLE test_tab (key) as 
'COMPACT' WITH DEFERRED REBUILD
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@test_tab
-POSTHOOK: Output: default@default__test_tab_test_tab_index__
-FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: 
Can not truncate columns from table with indexes

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/alter_concatenate_indexed_table.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/alter_concatenate_indexed_table.q.out 
b/ql/src/test/results/clientpositive/alter_concatenate_indexed_table.q.out
deleted file mode 100644
index ffcbcf9..0000000
--- a/ql/src/test/results/clientpositive/alter_concatenate_indexed_table.q.out
+++ /dev/null
@@ -1,271 +0,0 @@
-PREHOOK: query: create table src_rc_concatenate_test(key int, value string) 
stored as rcfile
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: create table src_rc_concatenate_test(key int, value string) 
stored as rcfile
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into 
table src_rc_concatenate_test
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into 
table src_rc_concatenate_test
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into 
table src_rc_concatenate_test
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into 
table src_rc_concatenate_test
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into 
table src_rc_concatenate_test
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into 
table src_rc_concatenate_test
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: show table extended like `src_rc_concatenate_test`
-PREHOOK: type: SHOW_TABLESTATUS
-POSTHOOK: query: show table extended like `src_rc_concatenate_test`
-POSTHOOK: type: SHOW_TABLESTATUS
-tableName:src_rc_concatenate_test
-#### A masked pattern was here ####
-inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat
-outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-columns:struct columns { i32 key, string value}
-partitioned:false
-partitionColumns:
-totalNumberFiles:3
-totalFileSize:636
-maxFileSize:222
-minFileSize:206
-#### A masked pattern was here ####
-
-PREHOOK: query: select count(1) from src_rc_concatenate_test
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-POSTHOOK: query: select count(1) from src_rc_concatenate_test
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-15
-PREHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-214    -7678496319
-PREHOOK: query: create index src_rc_concatenate_test_index on table 
src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES 
("prop1"="val1", "prop2"="val2")
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@src_rc_concatenate_test
-POSTHOOK: query: create index src_rc_concatenate_test_index on table 
src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES 
("prop1"="val1", "prop2"="val2")
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@src_rc_concatenate_test
-POSTHOOK: Output: 
default@default__src_rc_concatenate_test_src_rc_concatenate_test_index__
-PREHOOK: query: show indexes on src_rc_concatenate_test
-PREHOOK: type: SHOWINDEXES
-POSTHOOK: query: show indexes on src_rc_concatenate_test
-POSTHOOK: type: SHOWINDEXES
-src_rc_concatenate_test_index  src_rc_concatenate_test key                     
default__src_rc_concatenate_test_src_rc_concatenate_test_index__        compact 
                
-PREHOOK: query: alter table src_rc_concatenate_test concatenate
-PREHOOK: type: ALTER_TABLE_MERGE
-PREHOOK: Input: default@src_rc_concatenate_test
-PREHOOK: Output: default@src_rc_concatenate_test
-POSTHOOK: query: alter table src_rc_concatenate_test concatenate
-POSTHOOK: type: ALTER_TABLE_MERGE
-POSTHOOK: Input: default@src_rc_concatenate_test
-POSTHOOK: Output: default@src_rc_concatenate_test
-PREHOOK: query: show table extended like `src_rc_concatenate_test`
-PREHOOK: type: SHOW_TABLESTATUS
-POSTHOOK: query: show table extended like `src_rc_concatenate_test`
-POSTHOOK: type: SHOW_TABLESTATUS
-tableName:src_rc_concatenate_test
-#### A masked pattern was here ####
-inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat
-outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-columns:struct columns { i32 key, string value}
-partitioned:false
-partitionColumns:
-totalNumberFiles:1
-totalFileSize:239
-maxFileSize:239
-minFileSize:239
-#### A masked pattern was here ####
-
-PREHOOK: query: select count(1) from src_rc_concatenate_test
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-POSTHOOK: query: select count(1) from src_rc_concatenate_test
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-15
-PREHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test
-#### A masked pattern was here ####
-214    -7678496319
-PREHOOK: query: drop index src_rc_concatenate_test_index on 
src_rc_concatenate_test
-PREHOOK: type: DROPINDEX
-PREHOOK: Input: default@src_rc_concatenate_test
-POSTHOOK: query: drop index src_rc_concatenate_test_index on 
src_rc_concatenate_test
-POSTHOOK: type: DROPINDEX
-POSTHOOK: Input: default@src_rc_concatenate_test
-PREHOOK: query: create table src_rc_concatenate_test_part(key int, value 
string) partitioned by (ds string) stored as rcfile
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@src_rc_concatenate_test_part
-POSTHOOK: query: create table src_rc_concatenate_test_part(key int, value 
string) partitioned by (ds string) stored as rcfile
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@src_rc_concatenate_test_part
-PREHOOK: query: alter table src_rc_concatenate_test_part add partition 
(ds='2011')
-PREHOOK: type: ALTERTABLE_ADDPARTS
-PREHOOK: Output: default@src_rc_concatenate_test_part
-POSTHOOK: query: alter table src_rc_concatenate_test_part add partition 
(ds='2011')
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-POSTHOOK: Output: default@src_rc_concatenate_test_part
-POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into 
table src_rc_concatenate_test_part partition (ds='2011')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_1.rc' into 
table src_rc_concatenate_test_part partition (ds='2011')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into 
table src_rc_concatenate_test_part partition (ds='2011')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_2.rc' into 
table src_rc_concatenate_test_part partition (ds='2011')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-PREHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into 
table src_rc_concatenate_test_part partition (ds='2011')
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-POSTHOOK: query: load data local inpath '../../data/files/smbbucket_3.rc' into 
table src_rc_concatenate_test_part partition (ds='2011')
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-PREHOOK: query: show table extended like `src_rc_concatenate_test_part` 
partition (ds='2011')
-PREHOOK: type: SHOW_TABLESTATUS
-POSTHOOK: query: show table extended like `src_rc_concatenate_test_part` 
partition (ds='2011')
-POSTHOOK: type: SHOW_TABLESTATUS
-tableName:src_rc_concatenate_test_part
-#### A masked pattern was here ####
-inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat
-outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-columns:struct columns { i32 key, string value}
-partitioned:true
-partitionColumns:struct partition_columns { string ds}
-totalNumberFiles:3
-totalFileSize:636
-maxFileSize:222
-minFileSize:206
-#### A masked pattern was here ####
-
-PREHOOK: query: select count(1) from src_rc_concatenate_test_part
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test_part
-PREHOOK: Input: default@src_rc_concatenate_test_part@ds=2011
-#### A masked pattern was here ####
-POSTHOOK: query: select count(1) from src_rc_concatenate_test_part
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test_part
-POSTHOOK: Input: default@src_rc_concatenate_test_part@ds=2011
-#### A masked pattern was here ####
-15
-PREHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test_part
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test_part
-PREHOOK: Input: default@src_rc_concatenate_test_part@ds=2011
-#### A masked pattern was here ####
-POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test_part
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test_part
-POSTHOOK: Input: default@src_rc_concatenate_test_part@ds=2011
-#### A masked pattern was here ####
-214    -7678496319
-PREHOOK: query: create index src_rc_concatenate_test_part_index on table 
src_rc_concatenate_test_part(key) as 'compact' WITH DEFERRED REBUILD 
IDXPROPERTIES ("prop1"="val1", "prop2"="val2")
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@src_rc_concatenate_test_part
-POSTHOOK: query: create index src_rc_concatenate_test_part_index on table 
src_rc_concatenate_test_part(key) as 'compact' WITH DEFERRED REBUILD 
IDXPROPERTIES ("prop1"="val1", "prop2"="val2")
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@src_rc_concatenate_test_part
-POSTHOOK: Output: 
default@default__src_rc_concatenate_test_part_src_rc_concatenate_test_part_index__
-PREHOOK: query: show indexes on src_rc_concatenate_test_part
-PREHOOK: type: SHOWINDEXES
-POSTHOOK: query: show indexes on src_rc_concatenate_test_part
-POSTHOOK: type: SHOWINDEXES
-src_rc_concatenate_test_part_index     src_rc_concatenate_test_part    key     
                
default__src_rc_concatenate_test_part_src_rc_concatenate_test_part_index__      
compact                 
-PREHOOK: query: alter table src_rc_concatenate_test_part partition (ds='2011') 
concatenate
-PREHOOK: type: ALTER_PARTITION_MERGE
-PREHOOK: Input: default@src_rc_concatenate_test_part
-PREHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-POSTHOOK: query: alter table src_rc_concatenate_test_part partition 
(ds='2011') concatenate
-POSTHOOK: type: ALTER_PARTITION_MERGE
-POSTHOOK: Input: default@src_rc_concatenate_test_part
-POSTHOOK: Output: default@src_rc_concatenate_test_part@ds=2011
-PREHOOK: query: show table extended like `src_rc_concatenate_test_part` 
partition (ds='2011')
-PREHOOK: type: SHOW_TABLESTATUS
-POSTHOOK: query: show table extended like `src_rc_concatenate_test_part` 
partition (ds='2011')
-POSTHOOK: type: SHOW_TABLESTATUS
-tableName:src_rc_concatenate_test_part
-#### A masked pattern was here ####
-inputformat:org.apache.hadoop.hive.ql.io.RCFileInputFormat
-outputformat:org.apache.hadoop.hive.ql.io.RCFileOutputFormat
-columns:struct columns { i32 key, string value}
-partitioned:true
-partitionColumns:struct partition_columns { string ds}
-totalNumberFiles:1
-totalFileSize:239
-maxFileSize:239
-minFileSize:239
-#### A masked pattern was here ####
-
-PREHOOK: query: select count(1) from src_rc_concatenate_test_part
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test_part
-PREHOOK: Input: default@src_rc_concatenate_test_part@ds=2011
-#### A masked pattern was here ####
-POSTHOOK: query: select count(1) from src_rc_concatenate_test_part
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test_part
-POSTHOOK: Input: default@src_rc_concatenate_test_part@ds=2011
-#### A masked pattern was here ####
-15
-PREHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test_part
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src_rc_concatenate_test_part
-PREHOOK: Input: default@src_rc_concatenate_test_part@ds=2011
-#### A masked pattern was here ####
-POSTHOOK: query: select sum(hash(key)), sum(hash(value)) from 
src_rc_concatenate_test_part
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src_rc_concatenate_test_part
-POSTHOOK: Input: default@src_rc_concatenate_test_part@ds=2011
-#### A masked pattern was here ####
-214    -7678496319
-PREHOOK: query: drop index src_rc_concatenate_test_part_index on 
src_rc_concatenate_test_part
-PREHOOK: type: DROPINDEX
-PREHOOK: Input: default@src_rc_concatenate_test_part
-POSTHOOK: query: drop index src_rc_concatenate_test_part_index on 
src_rc_concatenate_test_part
-POSTHOOK: type: DROPINDEX
-POSTHOOK: Input: default@src_rc_concatenate_test_part

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/alter_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/alter_index.q.out 
b/ql/src/test/results/clientpositive/alter_index.q.out
deleted file mode 100644
index 262ad07..0000000
--- a/ql/src/test/results/clientpositive/alter_index.q.out
+++ /dev/null
@@ -1,67 +0,0 @@
-PREHOOK: query: drop index src_index_8 on src
-PREHOOK: type: DROPINDEX
-PREHOOK: Input: default@src
-POSTHOOK: query: drop index src_index_8 on src
-POSTHOOK: type: DROPINDEX
-POSTHOOK: Input: default@src
-PREHOOK: query: create index src_index_8 on table default.src(key) as 
'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2")
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@src
-POSTHOOK: query: create index src_index_8 on table default.src(key) as 
'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2")
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@src
-POSTHOOK: Output: default@default__src_src_index_8__
-PREHOOK: query: desc extended default__src_src_index_8__
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@default__src_src_index_8__
-POSTHOOK: query: desc extended default__src_src_index_8__
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@default__src_src_index_8__
-key                    string                  default             
-_bucketname            string                                      
-_offsets               array<bigint>                               
-                
-#### A masked pattern was here ####
-PREHOOK: query: alter index src_index_8 on default.src set IDXPROPERTIES 
("prop1"="val1_new", "prop3"="val3")
-PREHOOK: type: ALTERINDEX_PROPS
-POSTHOOK: query: alter index src_index_8 on default.src set IDXPROPERTIES 
("prop1"="val1_new", "prop3"="val3")
-POSTHOOK: type: ALTERINDEX_PROPS
-PREHOOK: query: desc extended default__src_src_index_8__
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@default__src_src_index_8__
-POSTHOOK: query: desc extended default__src_src_index_8__
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@default__src_src_index_8__
-key                    string                  default             
-_bucketname            string                                      
-_offsets               array<bigint>                               
-                
-#### A masked pattern was here ####
-PREHOOK: query: drop index src_index_8 on default.src
-PREHOOK: type: DROPINDEX
-PREHOOK: Input: default@src
-POSTHOOK: query: drop index src_index_8 on default.src
-POSTHOOK: type: DROPINDEX
-POSTHOOK: Input: default@src
-PREHOOK: query: show tables
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:default
-POSTHOOK: query: show tables
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:default
-alltypesorc
-alltypesparquet
-cbo_t1
-cbo_t2
-cbo_t3
-lineitem
-part
-src
-src1
-src_cbo
-src_json
-src_sequencefile
-src_thrift
-srcbucket
-srcbucket2
-srcpart

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/authorization_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/authorization_index.q.out 
b/ql/src/test/results/clientpositive/authorization_index.q.out
deleted file mode 100644
index 87486b3..0000000
--- a/ql/src/test/results/clientpositive/authorization_index.q.out
+++ /dev/null
@@ -1,62 +0,0 @@
-PREHOOK: query: create table t1 (a int)
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@t1
-POSTHOOK: query: create table t1 (a int)
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@t1
-PREHOOK: query: create index t1_index on table t1(a) as 'COMPACT' WITH 
DEFERRED REBUILD
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@t1
-POSTHOOK: query: create index t1_index on table t1(a) as 'COMPACT' WITH 
DEFERRED REBUILD
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@t1
-POSTHOOK: Output: default@default__t1_t1_index__
-PREHOOK: query: desc formatted default__t1_t1_index__
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@default__t1_t1_index__
-POSTHOOK: query: desc formatted default__t1_t1_index__
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@default__t1_t1_index__
-# col_name             data_type               comment             
-a                      int                                         
-_bucketname            string                                      
-_offsets               array<bigint>                               
-                
-# Detailed Table Information            
-Database:              default                  
-#### A masked pattern was here ####
-Retention:             0                        
-#### A masked pattern was here ####
-Table Type:            INDEX_TABLE              
-Table Parameters:               
-#### A masked pattern was here ####
-                
-# Storage Information           
-SerDe Library:         org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe      
 
-InputFormat:           org.apache.hadoop.mapred.TextInputFormat         
-OutputFormat:          org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat  
 
-Compressed:            No                       
-Num Buckets:           -1                       
-Bucket Columns:        []                       
-Sort Columns:          [Order(col:a, order:1)]  
-PREHOOK: query: alter index t1_index on t1 rebuild
-PREHOOK: type: ALTERINDEX_REBUILD
-PREHOOK: Input: default@t1
-PREHOOK: Output: default@default__t1_t1_index__
-POSTHOOK: query: alter index t1_index on t1 rebuild
-POSTHOOK: type: ALTERINDEX_REBUILD
-POSTHOOK: Input: default@t1
-POSTHOOK: Output: default@default__t1_t1_index__
-POSTHOOK: Lineage: default__t1_t1_index__._bucketname SIMPLE 
[(t1)t1.FieldSchema(name:INPUT__FILE__NAME, type:string, comment:), ]
-POSTHOOK: Lineage: default__t1_t1_index__._offsets EXPRESSION 
[(t1)t1.FieldSchema(name:BLOCK__OFFSET__INSIDE__FILE, type:bigint, comment:), ]
-POSTHOOK: Lineage: default__t1_t1_index__.a SIMPLE [(t1)t1.FieldSchema(name:a, 
type:int, comment:null), ]
-PREHOOK: query: drop table t1
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@t1
-PREHOOK: Output: default@t1
-POSTHOOK: query: drop table t1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@t1
-POSTHOOK: Output: default@t1

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/drop_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/drop_index.q.out 
b/ql/src/test/results/clientpositive/drop_index.q.out
deleted file mode 100644
index dc154d3..0000000
--- a/ql/src/test/results/clientpositive/drop_index.q.out
+++ /dev/null
@@ -1,10 +0,0 @@
-PREHOOK: query: DROP INDEX IF EXISTS UnknownIndex ON src
-PREHOOK: type: DROPINDEX
-PREHOOK: Input: default@src
-POSTHOOK: query: DROP INDEX IF EXISTS UnknownIndex ON src
-POSTHOOK: type: DROPINDEX
-POSTHOOK: Input: default@src
-PREHOOK: query: DROP INDEX IF EXISTS UnknownIndex ON UnknownTable
-PREHOOK: type: DROPINDEX
-POSTHOOK: query: DROP INDEX IF EXISTS UnknownIndex ON UnknownTable
-POSTHOOK: type: DROPINDEX

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/drop_index_removes_partition_dirs.q.out
----------------------------------------------------------------------
diff --git 
a/ql/src/test/results/clientpositive/drop_index_removes_partition_dirs.q.out 
b/ql/src/test/results/clientpositive/drop_index_removes_partition_dirs.q.out
deleted file mode 100644
index bbd86b4..0000000
--- a/ql/src/test/results/clientpositive/drop_index_removes_partition_dirs.q.out
+++ /dev/null
@@ -1,47 +0,0 @@
-PREHOOK: query: CREATE TABLE test_table (key STRING, value STRING)
-PARTITIONED BY (part STRING)
-STORED AS RCFILE
-#### A masked pattern was here ####
-PREHOOK: type: CREATETABLE
-#### A masked pattern was here ####
-PREHOOK: Output: database:default
-PREHOOK: Output: default@test_table
-POSTHOOK: query: CREATE TABLE test_table (key STRING, value STRING)
-PARTITIONED BY (part STRING)
-STORED AS RCFILE
-#### A masked pattern was here ####
-POSTHOOK: type: CREATETABLE
-#### A masked pattern was here ####
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@test_table
-PREHOOK: query: CREATE INDEX test_index ON 
-TABLE test_table(key) AS 'compact' WITH DEFERRED REBUILD
-IN TABLE test_index_table
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@test_table
-POSTHOOK: query: CREATE INDEX test_index ON 
-TABLE test_table(key) AS 'compact' WITH DEFERRED REBUILD
-IN TABLE test_index_table
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@test_table
-POSTHOOK: Output: default@test_index_table
-PREHOOK: query: ALTER TABLE test_index_table ADD PARTITION (part = '1')
-#### A masked pattern was here ####
-PREHOOK: type: ALTERTABLE_ADDPARTS
-#### A masked pattern was here ####
-PREHOOK: Output: default@test_index_table
-POSTHOOK: query: ALTER TABLE test_index_table ADD PARTITION (part = '1')
-#### A masked pattern was here ####
-POSTHOOK: type: ALTERTABLE_ADDPARTS
-#### A masked pattern was here ####
-POSTHOOK: Output: default@test_index_table
-POSTHOOK: Output: default@test_index_table@part=1
-Found 1 items
-#### A masked pattern was here ####
-PREHOOK: query: DROP INDEX test_index ON test_table
-PREHOOK: type: DROPINDEX
-PREHOOK: Input: default@test_table
-POSTHOOK: query: DROP INDEX test_index ON test_table
-POSTHOOK: type: DROPINDEX
-POSTHOOK: Input: default@test_table
-#### A masked pattern was here ####

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/drop_table_with_index.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/drop_table_with_index.q.out 
b/ql/src/test/results/clientpositive/drop_table_with_index.q.out
deleted file mode 100644
index 2312eee..0000000
--- a/ql/src/test/results/clientpositive/drop_table_with_index.q.out
+++ /dev/null
@@ -1,153 +0,0 @@
-PREHOOK: query: DROP TABLE IF EXISTS aa
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE IF EXISTS aa
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE aa (L_ORDERKEY      INT,
-                                L_PARTKEY       INT,
-                                L_SUPPKEY       INT,
-                                L_LINENUMBER    INT,
-                                L_QUANTITY      DOUBLE,
-                                L_EXTENDEDPRICE DOUBLE,
-                                L_DISCOUNT      DOUBLE,
-                                L_TAX           DOUBLE,
-                                L_RETURNFLAG    STRING,
-                                L_LINESTATUS    STRING,
-                                l_shipdate      STRING,
-                                L_COMMITDATE    STRING,
-                                L_RECEIPTDATE   STRING,
-                                L_SHIPINSTRUCT  STRING,
-                                L_SHIPMODE      STRING,
-                                L_COMMENT       STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@aa
-POSTHOOK: query: CREATE TABLE aa (L_ORDERKEY      INT,
-                                L_PARTKEY       INT,
-                                L_SUPPKEY       INT,
-                                L_LINENUMBER    INT,
-                                L_QUANTITY      DOUBLE,
-                                L_EXTENDEDPRICE DOUBLE,
-                                L_DISCOUNT      DOUBLE,
-                                L_TAX           DOUBLE,
-                                L_RETURNFLAG    STRING,
-                                L_LINESTATUS    STRING,
-                                l_shipdate      STRING,
-                                L_COMMITDATE    STRING,
-                                L_RECEIPTDATE   STRING,
-                                L_SHIPINSTRUCT  STRING,
-                                L_SHIPMODE      STRING,
-                                L_COMMENT       STRING)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY '|'
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@aa
-PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' 
OVERWRITE INTO TABLE aa
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@aa
-POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/lineitem.txt' 
OVERWRITE INTO TABLE aa
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@aa
-PREHOOK: query: CREATE INDEX aa_lshipdate_idx ON TABLE aa(l_shipdate) AS 
'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD 
IDXPROPERTIES("AGGREGATES"="count(l_shipdate)")
-PREHOOK: type: CREATEINDEX
-PREHOOK: Input: default@aa
-POSTHOOK: query: CREATE INDEX aa_lshipdate_idx ON TABLE aa(l_shipdate) AS 
'org.apache.hadoop.hive.ql.index.AggregateIndexHandler' WITH DEFERRED REBUILD 
IDXPROPERTIES("AGGREGATES"="count(l_shipdate)")
-POSTHOOK: type: CREATEINDEX
-POSTHOOK: Input: default@aa
-POSTHOOK: Output: default@default__aa_aa_lshipdate_idx__
-PREHOOK: query: ALTER INDEX aa_lshipdate_idx ON aa REBUILD
-PREHOOK: type: ALTERINDEX_REBUILD
-PREHOOK: Input: default@aa
-PREHOOK: Output: default@default__aa_aa_lshipdate_idx__
-POSTHOOK: query: ALTER INDEX aa_lshipdate_idx ON aa REBUILD
-POSTHOOK: type: ALTERINDEX_REBUILD
-POSTHOOK: Input: default@aa
-POSTHOOK: Output: default@default__aa_aa_lshipdate_idx__
-POSTHOOK: Lineage: default__aa_aa_lshipdate_idx__._bucketname SIMPLE 
[(aa)aa.FieldSchema(name:INPUT__FILE__NAME, type:string, comment:), ]
-POSTHOOK: Lineage: default__aa_aa_lshipdate_idx__._count_of_l_shipdate 
EXPRESSION [(aa)aa.FieldSchema(name:l_shipdate, type:string, comment:null), ]
-POSTHOOK: Lineage: default__aa_aa_lshipdate_idx__._offsets EXPRESSION 
[(aa)aa.FieldSchema(name:BLOCK__OFFSET__INSIDE__FILE, type:bigint, comment:), ]
-POSTHOOK: Lineage: default__aa_aa_lshipdate_idx__.l_shipdate SIMPLE 
[(aa)aa.FieldSchema(name:l_shipdate, type:string, comment:null), ]
-PREHOOK: query: show tables
-PREHOOK: type: SHOWTABLES
-PREHOOK: Input: database:default
-POSTHOOK: query: show tables
-POSTHOOK: type: SHOWTABLES
-POSTHOOK: Input: database:default
-aa
-alltypesorc
-alltypesparquet
-cbo_t1
-cbo_t2
-cbo_t3
-default__aa_aa_lshipdate_idx__
-lineitem
-part
-src
-src1
-src_cbo
-src_json
-src_sequencefile
-src_thrift
-srcbucket
-srcbucket2
-srcpart
-PREHOOK: query: explain select l_shipdate, count(l_shipdate)
-from aa
-group by l_shipdate
-PREHOOK: type: QUERY
-POSTHOOK: query: explain select l_shipdate, count(l_shipdate)
-from aa
-group by l_shipdate
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Map Reduce
-      Map Operator Tree:
-          TableScan
-            alias: aa
-            Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE 
Column stats: NONE
-            Select Operator
-              expressions: l_shipdate (type: string)
-              outputColumnNames: l_shipdate
-              Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE 
Column stats: NONE
-              Group By Operator
-                aggregations: count(l_shipdate)
-                keys: l_shipdate (type: string)
-                mode: hash
-                outputColumnNames: _col0, _col1
-                Statistics: Num rows: 1 Data size: 120990 Basic stats: 
COMPLETE Column stats: NONE
-                Reduce Output Operator
-                  key expressions: _col0 (type: string)
-                  sort order: +
-                  Map-reduce partition columns: _col0 (type: string)
-                  Statistics: Num rows: 1 Data size: 120990 Basic stats: 
COMPLETE Column stats: NONE
-                  value expressions: _col1 (type: bigint)
-      Reduce Operator Tree:
-        Group By Operator
-          aggregations: count(VALUE._col0)
-          keys: KEY._col0 (type: string)
-          mode: mergepartial
-          outputColumnNames: _col0, _col1
-          Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE 
Column stats: NONE
-          File Output Operator
-            compressed: false
-            Statistics: Num rows: 1 Data size: 120990 Basic stats: COMPLETE 
Column stats: NONE
-            table:
-                input format: org.apache.hadoop.mapred.SequenceFileInputFormat
-                output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_const.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_const.q.out 
b/ql/src/test/results/clientpositive/llap/vector_const.q.out
deleted file mode 100644
index 964ddcc..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_const.q.out
+++ /dev/null
@@ -1,66 +0,0 @@
-PREHOOK: query: CREATE TEMPORARY TABLE varchar_const_1 (c1 int) STORED AS ORC
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@varchar_const_1
-POSTHOOK: query: CREATE TEMPORARY TABLE varchar_const_1 (c1 int) STORED AS ORC
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@varchar_const_1
-PREHOOK: query: INSERT INTO varchar_const_1 values(42)
-PREHOOK: type: QUERY
-PREHOOK: Input: _dummy_database@_dummy_table
-PREHOOK: Output: default@varchar_const_1
-POSTHOOK: query: INSERT INTO varchar_const_1 values(42)
-POSTHOOK: type: QUERY
-POSTHOOK: Input: _dummy_database@_dummy_table
-POSTHOOK: Output: default@varchar_const_1
-POSTHOOK: Lineage: varchar_const_1.c1 SCRIPT []
-PREHOOK: query: EXPLAIN
-SELECT CONCAT(CAST('F' AS CHAR(2)), CAST('F' AS VARCHAR(2))) FROM 
VARCHAR_CONST_1
-PREHOOK: type: QUERY
-POSTHOOK: query: EXPLAIN
-SELECT CONCAT(CAST('F' AS CHAR(2)), CAST('F' AS VARCHAR(2))) FROM 
VARCHAR_CONST_1
-POSTHOOK: type: QUERY
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: varchar_const_1
-                  Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  Select Operator
-                    expressions: 'FF' (type: varchar(4))
-                    outputColumnNames: _col0
-                    Statistics: Num rows: 1 Data size: 86 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    File Output Operator
-                      compressed: false
-                      Statistics: Num rows: 1 Data size: 86 Basic stats: 
COMPLETE Column stats: COMPLETE
-                      table:
-                          input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                          output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                          serde: 
org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: SELECT CONCAT(CAST('F' AS CHAR(2)), CAST('F' AS VARCHAR(2))) 
FROM VARCHAR_CONST_1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@varchar_const_1
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT CONCAT(CAST('F' AS CHAR(2)), CAST('F' AS VARCHAR(2))) 
FROM VARCHAR_CONST_1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@varchar_const_1
-#### A masked pattern was here ####
-FF

http://git-wip-us.apache.org/repos/asf/hive/blob/99380fbd/ql/src/test/results/clientpositive/llap/vector_empty_where.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/vector_empty_where.q.out 
b/ql/src/test/results/clientpositive/llap/vector_empty_where.q.out
deleted file mode 100644
index 494c5c9..0000000
--- a/ql/src/test/results/clientpositive/llap/vector_empty_where.q.out
+++ /dev/null
@@ -1,652 +0,0 @@
-PREHOOK: query: explain vectorization expression
-select count (distinct cint) from alltypesorc where cstring1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
-select count (distinct cint) from alltypesorc where cstring1
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: alltypesorc
-                  Statistics: Num rows: 12288 Data size: 899146 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  TableScan Vectorization:
-                      native: true
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: SelectColumnIsTrue(col 
13:boolean)(children: CastStringToBoolean(col 6) -> 13:boolean)
-                    predicate: cstring1 (type: string)
-                    Statistics: Num rows: 6144 Data size: 449620 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Select Operator
-                      expressions: cint (type: int)
-                      outputColumnNames: cint
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumnNums: [2]
-                      Statistics: Num rows: 6144 Data size: 449620 Basic 
stats: COMPLETE Column stats: COMPLETE
-                      Group By Operator
-                        Group By Vectorization:
-                            className: VectorGroupByOperator
-                            groupByMode: HASH
-                            keyExpressions: col 2:int
-                            native: false
-                            vectorProcessingMode: HASH
-                            projectedOutputColumnNums: []
-                        keys: cint (type: int)
-                        mode: hash
-                        outputColumnNames: _col0
-                        Statistics: Num rows: 3016 Data size: 9008 Basic 
stats: COMPLETE Column stats: COMPLETE
-                        Reduce Output Operator
-                          key expressions: _col0 (type: int)
-                          sort order: +
-                          Map-reduce partition columns: _col0 (type: int)
-                          Reduce Sink Vectorization:
-                              className: VectorReduceSinkLongOperator
-                              native: true
-                              nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                          Statistics: Num rows: 3016 Data size: 9008 Basic 
stats: COMPLETE Column stats: COMPLETE
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-        Reducer 2 
-            Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-            Reduce Operator Tree:
-              Group By Operator
-                Group By Vectorization:
-                    className: VectorGroupByOperator
-                    groupByMode: MERGEPARTIAL
-                    keyExpressions: col 0:int
-                    native: false
-                    vectorProcessingMode: MERGE_PARTIAL
-                    projectedOutputColumnNums: []
-                keys: KEY._col0 (type: int)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 3016 Data size: 9008 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Group By Operator
-                  aggregations: count(_col0)
-                  Group By Vectorization:
-                      aggregators: VectorUDAFCount(col 0:int) -> bigint
-                      className: VectorGroupByOperator
-                      groupByMode: HASH
-                      native: false
-                      vectorProcessingMode: HASH
-                      projectedOutputColumnNums: [0]
-                  mode: hash
-                  outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  Reduce Output Operator
-                    sort order: 
-                    Reduce Sink Vectorization:
-                        className: VectorReduceSinkEmptyKeyOperator
-                        native: true
-                        nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                    value expressions: _col0 (type: bigint)
-        Reducer 3 
-            Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(VALUE._col0)
-                Group By Vectorization:
-                    aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint
-                    className: VectorGroupByOperator
-                    groupByMode: MERGEPARTIAL
-                    native: false
-                    vectorProcessingMode: GLOBAL
-                    projectedOutputColumnNums: [0]
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select count (distinct cint) from alltypesorc where cstring1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-POSTHOOK: query: select count (distinct cint) from alltypesorc where cstring1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-6041
-PREHOOK: query: explain vectorization expression
-select count (distinct cint) from alltypesorc where cint
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
-select count (distinct cint) from alltypesorc where cint
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: alltypesorc
-                  Statistics: Num rows: 12288 Data size: 36696 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  TableScan Vectorization:
-                      native: true
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: SelectColumnIsTrue(col 
13:boolean)(children: CastLongToBooleanViaLongToLong(col 2:int) -> 13:boolean)
-                    predicate: cint (type: int)
-                    Statistics: Num rows: 6144 Data size: 18348 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Group By Operator
-                      Group By Vectorization:
-                          className: VectorGroupByOperator
-                          groupByMode: HASH
-                          keyExpressions: col 2:int
-                          native: false
-                          vectorProcessingMode: HASH
-                          projectedOutputColumnNums: []
-                      keys: cint (type: int)
-                      mode: hash
-                      outputColumnNames: _col0
-                      Statistics: Num rows: 3016 Data size: 9008 Basic stats: 
COMPLETE Column stats: COMPLETE
-                      Reduce Output Operator
-                        key expressions: _col0 (type: int)
-                        sort order: +
-                        Map-reduce partition columns: _col0 (type: int)
-                        Reduce Sink Vectorization:
-                            className: VectorReduceSinkLongOperator
-                            native: true
-                            nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                        Statistics: Num rows: 3016 Data size: 9008 Basic 
stats: COMPLETE Column stats: COMPLETE
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-        Reducer 2 
-            Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-            Reduce Operator Tree:
-              Group By Operator
-                Group By Vectorization:
-                    className: VectorGroupByOperator
-                    groupByMode: MERGEPARTIAL
-                    keyExpressions: col 0:int
-                    native: false
-                    vectorProcessingMode: MERGE_PARTIAL
-                    projectedOutputColumnNums: []
-                keys: KEY._col0 (type: int)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 3016 Data size: 9008 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Group By Operator
-                  aggregations: count(_col0)
-                  Group By Vectorization:
-                      aggregators: VectorUDAFCount(col 0:int) -> bigint
-                      className: VectorGroupByOperator
-                      groupByMode: HASH
-                      native: false
-                      vectorProcessingMode: HASH
-                      projectedOutputColumnNums: [0]
-                  mode: hash
-                  outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  Reduce Output Operator
-                    sort order: 
-                    Reduce Sink Vectorization:
-                        className: VectorReduceSinkEmptyKeyOperator
-                        native: true
-                        nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                    value expressions: _col0 (type: bigint)
-        Reducer 3 
-            Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(VALUE._col0)
-                Group By Vectorization:
-                    aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint
-                    className: VectorGroupByOperator
-                    groupByMode: MERGEPARTIAL
-                    native: false
-                    vectorProcessingMode: GLOBAL
-                    projectedOutputColumnNums: [0]
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select count (distinct cint) from alltypesorc where cint
-PREHOOK: type: QUERY
-PREHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-POSTHOOK: query: select count (distinct cint) from alltypesorc where cint
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-6082
-PREHOOK: query: explain vectorization expression
-select count (distinct cint) from alltypesorc where cfloat
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
-select count (distinct cint) from alltypesorc where cfloat
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: alltypesorc
-                  Statistics: Num rows: 12288 Data size: 73392 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  TableScan Vectorization:
-                      native: true
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: SelectColumnIsTrue(col 
13:boolean)(children: CastDoubleToBooleanViaDoubleToLong(col 4:float) -> 
13:boolean)
-                    predicate: cfloat (type: float)
-                    Statistics: Num rows: 6144 Data size: 36696 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Select Operator
-                      expressions: cint (type: int)
-                      outputColumnNames: cint
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumnNums: [2]
-                      Statistics: Num rows: 6144 Data size: 36696 Basic stats: 
COMPLETE Column stats: COMPLETE
-                      Group By Operator
-                        Group By Vectorization:
-                            className: VectorGroupByOperator
-                            groupByMode: HASH
-                            keyExpressions: col 2:int
-                            native: false
-                            vectorProcessingMode: HASH
-                            projectedOutputColumnNums: []
-                        keys: cint (type: int)
-                        mode: hash
-                        outputColumnNames: _col0
-                        Statistics: Num rows: 3016 Data size: 9008 Basic 
stats: COMPLETE Column stats: COMPLETE
-                        Reduce Output Operator
-                          key expressions: _col0 (type: int)
-                          sort order: +
-                          Map-reduce partition columns: _col0 (type: int)
-                          Reduce Sink Vectorization:
-                              className: VectorReduceSinkLongOperator
-                              native: true
-                              nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                          Statistics: Num rows: 3016 Data size: 9008 Basic 
stats: COMPLETE Column stats: COMPLETE
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-        Reducer 2 
-            Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-            Reduce Operator Tree:
-              Group By Operator
-                Group By Vectorization:
-                    className: VectorGroupByOperator
-                    groupByMode: MERGEPARTIAL
-                    keyExpressions: col 0:int
-                    native: false
-                    vectorProcessingMode: MERGE_PARTIAL
-                    projectedOutputColumnNums: []
-                keys: KEY._col0 (type: int)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 3016 Data size: 9008 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Group By Operator
-                  aggregations: count(_col0)
-                  Group By Vectorization:
-                      aggregators: VectorUDAFCount(col 0:int) -> bigint
-                      className: VectorGroupByOperator
-                      groupByMode: HASH
-                      native: false
-                      vectorProcessingMode: HASH
-                      projectedOutputColumnNums: [0]
-                  mode: hash
-                  outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  Reduce Output Operator
-                    sort order: 
-                    Reduce Sink Vectorization:
-                        className: VectorReduceSinkEmptyKeyOperator
-                        native: true
-                        nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                    value expressions: _col0 (type: bigint)
-        Reducer 3 
-            Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(VALUE._col0)
-                Group By Vectorization:
-                    aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint
-                    className: VectorGroupByOperator
-                    groupByMode: MERGEPARTIAL
-                    native: false
-                    vectorProcessingMode: GLOBAL
-                    projectedOutputColumnNums: [0]
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select count (distinct cint) from alltypesorc where cfloat
-PREHOOK: type: QUERY
-PREHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-POSTHOOK: query: select count (distinct cint) from alltypesorc where cfloat
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-3022
-PREHOOK: query: explain vectorization expression
-select count (distinct cint) from alltypesorc where ctimestamp1
-PREHOOK: type: QUERY
-POSTHOOK: query: explain vectorization expression
-select count (distinct cint) from alltypesorc where ctimestamp1
-POSTHOOK: type: QUERY
-PLAN VECTORIZATION:
-  enabled: true
-  enabledConditionsMet: [hive.vectorized.execution.enabled IS true]
-
-STAGE DEPENDENCIES:
-  Stage-1 is a root stage
-  Stage-0 depends on stages: Stage-1
-
-STAGE PLANS:
-  Stage: Stage-1
-    Tez
-#### A masked pattern was here ####
-      Edges:
-        Reducer 2 <- Map 1 (SIMPLE_EDGE)
-        Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE)
-#### A masked pattern was here ####
-      Vertices:
-        Map 1 
-            Map Operator Tree:
-                TableScan
-                  alias: alltypesorc
-                  Statistics: Num rows: 12288 Data size: 528216 Basic stats: 
COMPLETE Column stats: COMPLETE
-                  TableScan Vectorization:
-                      native: true
-                  Filter Operator
-                    Filter Vectorization:
-                        className: VectorFilterOperator
-                        native: true
-                        predicateExpression: SelectColumnIsTrue(col 
13:boolean)(children: CastTimestampToBoolean(col 8:timestamp) -> 13:boolean)
-                    predicate: ctimestamp1 (type: timestamp)
-                    Statistics: Num rows: 6144 Data size: 264108 Basic stats: 
COMPLETE Column stats: COMPLETE
-                    Select Operator
-                      expressions: cint (type: int)
-                      outputColumnNames: cint
-                      Select Vectorization:
-                          className: VectorSelectOperator
-                          native: true
-                          projectedOutputColumnNums: [2]
-                      Statistics: Num rows: 6144 Data size: 264108 Basic 
stats: COMPLETE Column stats: COMPLETE
-                      Group By Operator
-                        Group By Vectorization:
-                            className: VectorGroupByOperator
-                            groupByMode: HASH
-                            keyExpressions: col 2:int
-                            native: false
-                            vectorProcessingMode: HASH
-                            projectedOutputColumnNums: []
-                        keys: cint (type: int)
-                        mode: hash
-                        outputColumnNames: _col0
-                        Statistics: Num rows: 3016 Data size: 9008 Basic 
stats: COMPLETE Column stats: COMPLETE
-                        Reduce Output Operator
-                          key expressions: _col0 (type: int)
-                          sort order: +
-                          Map-reduce partition columns: _col0 (type: int)
-                          Reduce Sink Vectorization:
-                              className: VectorReduceSinkLongOperator
-                              native: true
-                              nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                          Statistics: Num rows: 3016 Data size: 9008 Basic 
stats: COMPLETE Column stats: COMPLETE
-            Execution mode: vectorized, llap
-            LLAP IO: all inputs
-            Map Vectorization:
-                enabled: true
-                enabledConditionsMet: 
hive.vectorized.use.vectorized.input.format IS true
-                inputFormatFeatureSupport: []
-                featureSupportInUse: []
-                inputFileFormats: 
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-        Reducer 2 
-            Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-            Reduce Operator Tree:
-              Group By Operator
-                Group By Vectorization:
-                    className: VectorGroupByOperator
-                    groupByMode: MERGEPARTIAL
-                    keyExpressions: col 0:int
-                    native: false
-                    vectorProcessingMode: MERGE_PARTIAL
-                    projectedOutputColumnNums: []
-                keys: KEY._col0 (type: int)
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 3016 Data size: 9008 Basic stats: 
COMPLETE Column stats: COMPLETE
-                Group By Operator
-                  aggregations: count(_col0)
-                  Group By Vectorization:
-                      aggregators: VectorUDAFCount(col 0:int) -> bigint
-                      className: VectorGroupByOperator
-                      groupByMode: HASH
-                      native: false
-                      vectorProcessingMode: HASH
-                      projectedOutputColumnNums: [0]
-                  mode: hash
-                  outputColumnNames: _col0
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  Reduce Output Operator
-                    sort order: 
-                    Reduce Sink Vectorization:
-                        className: VectorReduceSinkEmptyKeyOperator
-                        native: true
-                        nativeConditionsMet: 
hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine 
tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, 
BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true
-                    Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                    value expressions: _col0 (type: bigint)
-        Reducer 3 
-            Execution mode: vectorized, llap
-            Reduce Vectorization:
-                enabled: true
-                enableConditionsMet: hive.vectorized.execution.reduce.enabled 
IS true, hive.execution.engine tez IN [tez, spark] IS true
-                allNative: false
-                usesVectorUDFAdaptor: false
-                vectorized: true
-            Reduce Operator Tree:
-              Group By Operator
-                aggregations: count(VALUE._col0)
-                Group By Vectorization:
-                    aggregators: VectorUDAFCountMerge(col 0:bigint) -> bigint
-                    className: VectorGroupByOperator
-                    groupByMode: MERGEPARTIAL
-                    native: false
-                    vectorProcessingMode: GLOBAL
-                    projectedOutputColumnNums: [0]
-                mode: mergepartial
-                outputColumnNames: _col0
-                Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                File Output Operator
-                  compressed: false
-                  File Sink Vectorization:
-                      className: VectorFileSinkOperator
-                      native: false
-                  Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE 
Column stats: COMPLETE
-                  table:
-                      input format: 
org.apache.hadoop.mapred.SequenceFileInputFormat
-                      output format: 
org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
-                      serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
-
-  Stage: Stage-0
-    Fetch Operator
-      limit: -1
-      Processor Tree:
-        ListSink
-
-PREHOOK: query: select count (distinct cint) from alltypesorc where ctimestamp1
-PREHOOK: type: QUERY
-PREHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-POSTHOOK: query: select count (distinct cint) from alltypesorc where 
ctimestamp1
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@alltypesorc
-#### A masked pattern was here ####
-3022

Reply via email to