http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/schema_evol_text_vecrow_mapwork_table.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/schema_evol_text_vecrow_mapwork_table.q.out b/ql/src/test/results/clientpositive/schema_evol_text_vecrow_mapwork_table.q.out index 50790f4..561c128 100644 --- a/ql/src/test/results/clientpositive/schema_evol_text_vecrow_mapwork_table.q.out +++ b/ql/src/test/results/clientpositive/schema_evol_text_vecrow_mapwork_table.q.out @@ -52,6 +52,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information @@ -187,25 +192,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_add_int_permute_select - Statistics: Num rows: 10 Data size: 111 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), a (type: int), b (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 10 Data size: 111 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 111 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: int), _col2 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: int), VALUE._col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 10 Data size: 111 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 111 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -318,6 +323,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information @@ -456,25 +466,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_add_int_string_permute_select - Statistics: Num rows: 10 Data size: 155 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), a (type: int), b (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 10 Data size: 155 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 155 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: int), _col2 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: int), VALUE._col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 10 Data size: 155 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 155 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -740,25 +750,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_string_group_double - Statistics: Num rows: 10 Data size: 383 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 151 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: double), c2 (type: double), c3 (type: double), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 383 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 151 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 383 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 151 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: double), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: double), VALUE._col1 (type: double), VALUE._col2 (type: double), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 383 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 151 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 383 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 151 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -912,25 +922,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_date_group_string_group_timestamp - Statistics: Num rows: 9 Data size: 1194 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 450 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: string), c2 (type: char(50)), c3 (type: char(15)), c4 (type: varchar(50)), c5 (type: varchar(15)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 9 Data size: 1194 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 450 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 9 Data size: 1194 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 450 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: string), _col2 (type: char(50)), _col3 (type: char(15)), _col4 (type: varchar(50)), _col5 (type: varchar(15)), _col6 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string), VALUE._col1 (type: char(50)), VALUE._col2 (type: char(15)), VALUE._col3 (type: varchar(50)), VALUE._col4 (type: varchar(15)), VALUE._col5 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 9 Data size: 1194 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 450 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 9 Data size: 1194 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 450 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1100,25 +1110,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_date_group_string_group_date - Statistics: Num rows: 9 Data size: 555 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 195 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: string), c2 (type: char(50)), c3 (type: char(15)), c4 (type: varchar(50)), c5 (type: varchar(15)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 9 Data size: 555 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 195 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 9 Data size: 555 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 195 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: string), _col2 (type: char(50)), _col3 (type: char(15)), _col4 (type: varchar(50)), _col5 (type: varchar(15)), _col6 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string), VALUE._col1 (type: char(50)), VALUE._col2 (type: char(15)), VALUE._col3 (type: varchar(50)), VALUE._col4 (type: varchar(15)), VALUE._col5 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 9 Data size: 555 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 195 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 9 Data size: 555 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 195 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1268,25 +1278,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_multi_ints_string - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: string), c2 (type: string), c3 (type: string), c4 (type: string), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string), VALUE._col3 (type: string), VALUE._col4 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1429,25 +1439,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_multi_ints_char - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: char(50)), c2 (type: char(50)), c3 (type: char(50)), c4 (type: char(50)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: char(50)), _col2 (type: char(50)), _col3 (type: char(50)), _col4 (type: char(50)), _col5 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: char(50)), VALUE._col1 (type: char(50)), VALUE._col2 (type: char(50)), VALUE._col3 (type: char(50)), VALUE._col4 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1590,25 +1600,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_multi_ints_char_trunc - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: char(5)), c2 (type: char(5)), c3 (type: char(5)), c4 (type: char(5)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: char(5)), _col2 (type: char(5)), _col3 (type: char(5)), _col4 (type: char(5)), _col5 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: char(5)), VALUE._col1 (type: char(5)), VALUE._col2 (type: char(5)), VALUE._col3 (type: char(5)), VALUE._col4 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1751,25 +1761,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_multi_ints_varchar - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: varchar(50)), c2 (type: varchar(50)), c3 (type: varchar(50)), c4 (type: varchar(50)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: varchar(50)), _col2 (type: varchar(50)), _col3 (type: varchar(50)), _col4 (type: varchar(50)), _col5 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: varchar(50)), VALUE._col1 (type: varchar(50)), VALUE._col2 (type: varchar(50)), VALUE._col3 (type: varchar(50)), VALUE._col4 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 330 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1912,25 +1922,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_multi_ints_varchar_trunc - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: varchar(5)), c2 (type: varchar(5)), c3 (type: varchar(5)), c4 (type: varchar(5)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: varchar(5)), _col2 (type: varchar(5)), _col3 (type: varchar(5)), _col4 (type: varchar(5)), _col5 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: varchar(5)), VALUE._col1 (type: varchar(5)), VALUE._col2 (type: varchar(5)), VALUE._col3 (type: varchar(5)), VALUE._col4 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 304 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 150 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2071,25 +2081,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_floating_string - Statistics: Num rows: 10 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: string), c2 (type: string), c3 (type: string), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 380 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2230,25 +2240,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_floating_char - Statistics: Num rows: 10 Data size: 385 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: char(50)), c2 (type: char(50)), c3 (type: char(50)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 385 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 385 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: char(50)), _col2 (type: char(50)), _col3 (type: char(50)), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: char(50)), VALUE._col1 (type: char(50)), VALUE._col2 (type: char(50)), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 385 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 385 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2389,25 +2399,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_floating_char_trunc - Statistics: Num rows: 10 Data size: 357 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: char(7)), c2 (type: char(7)), c3 (type: char(7)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 357 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 357 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: char(7)), _col2 (type: char(7)), _col3 (type: char(7)), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: char(7)), VALUE._col1 (type: char(7)), VALUE._col2 (type: char(7)), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 357 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 357 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 215 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2548,25 +2558,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_floating_varchar - Statistics: Num rows: 10 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 203 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: varchar(50)), c2 (type: varchar(50)), c3 (type: varchar(50)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 203 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 203 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: varchar(50)), _col2 (type: varchar(50)), _col3 (type: varchar(50)), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: varchar(50)), VALUE._col1 (type: varchar(50)), VALUE._col2 (type: varchar(50)), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 203 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 203 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2707,25 +2717,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_numeric_group_string_group_floating_varchar_trunc - Statistics: Num rows: 10 Data size: 362 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: varchar(7)), c2 (type: varchar(7)), c3 (type: varchar(7)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 362 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 362 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: varchar(7)), _col2 (type: varchar(7)), _col3 (type: varchar(7)), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: varchar(7)), VALUE._col1 (type: varchar(7)), VALUE._col2 (type: varchar(7)), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 362 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 362 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 220 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -2876,25 +2886,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_string_group_string_group_string - Statistics: Num rows: 10 Data size: 548 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 236 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: char(50)), c2 (type: char(9)), c3 (type: varchar(50)), c4 (type: char(9)), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 548 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 236 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 548 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 236 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: char(50)), _col2 (type: char(9)), _col3 (type: varchar(50)), _col4 (type: char(9)), _col5 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: char(50)), VALUE._col1 (type: char(9)), VALUE._col2 (type: varchar(50)), VALUE._col3 (type: char(9)), VALUE._col4 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 548 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 236 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 548 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 236 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -3035,25 +3045,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_string_group_string_group_char - Statistics: Num rows: 10 Data size: 463 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 178 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: varchar(50)), c2 (type: varchar(9)), c3 (type: string), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 463 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 178 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 463 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 178 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: varchar(50)), _col2 (type: varchar(9)), _col3 (type: string), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: varchar(50)), VALUE._col1 (type: varchar(9)), VALUE._col2 (type: string), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 463 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 178 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 463 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 178 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -3194,25 +3204,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_string_group_string_group_varchar - Statistics: Num rows: 10 Data size: 452 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 187 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: char(50)), c2 (type: char(9)), c3 (type: string), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 452 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 187 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 452 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 187 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: char(50)), _col2 (type: char(9)), _col3 (type: string), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: char(50)), VALUE._col1 (type: char(9)), VALUE._col2 (type: string), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 452 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 187 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 452 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 187 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -3367,25 +3377,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_lower_to_higher_numeric_group_tinyint - Statistics: Num rows: 10 Data size: 446 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 118 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: smallint), c2 (type: int), c3 (type: bigint), c4 (type: decimal(38,18)), c5 (type: float), c6 (type: double), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 10 Data size: 446 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 118 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 446 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 118 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: smallint), _col2 (type: int), _col3 (type: bigint), _col4 (type: decimal(38,18)), _col5 (type: float), _col6 (type: double), _col7 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: smallint), VALUE._col1 (type: int), VALUE._col2 (type: bigint), VALUE._col3 (type: decimal(38,18)), VALUE._col4 (type: float), VALUE._col5 (type: double), VALUE._col6 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 10 Data size: 446 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 118 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 446 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 118 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -3530,25 +3540,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_lower_to_higher_numeric_group_smallint - Statistics: Num rows: 10 Data size: 445 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 130 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: int), c2 (type: bigint), c3 (type: decimal(38,18)), c4 (type: float), c5 (type: double), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 10 Data size: 445 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 130 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 445 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 130 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: int), _col2 (type: bigint), _col3 (type: decimal(38,18)), _col4 (type: float), _col5 (type: double), _col6 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: int), VALUE._col1 (type: bigint), VALUE._col2 (type: decimal(38,18)), VALUE._col3 (type: float), VALUE._col4 (type: double), VALUE._col5 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 10 Data size: 445 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 130 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 445 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 130 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -3691,25 +3701,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_lower_to_higher_numeric_group_int - Statistics: Num rows: 10 Data size: 429 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 132 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: bigint), c2 (type: decimal(38,18)), c3 (type: float), c4 (type: double), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 429 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 132 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 429 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 132 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: bigint), _col2 (type: decimal(38,18)), _col3 (type: float), _col4 (type: double), _col5 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: bigint), VALUE._col1 (type: decimal(38,18)), VALUE._col2 (type: float), VALUE._col3 (type: double), VALUE._col4 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 10 Data size: 429 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 132 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 429 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 132 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -3850,25 +3860,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_lower_to_higher_numeric_group_bigint - Statistics: Num rows: 10 Data size: 411 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 127 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: decimal(38,18)), c2 (type: float), c3 (type: double), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 411 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 127 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 411 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 127 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: decimal(38,18)), _col2 (type: float), _col3 (type: double), _col4 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: decimal(38,18)), VALUE._col1 (type: float), VALUE._col2 (type: double), VALUE._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 411 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 127 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 411 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 127 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -4007,25 +4017,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_lower_to_higher_numeric_group_decimal - Statistics: Num rows: 10 Data size: 335 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 238 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: float), c2 (type: double), b (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 10 Data size: 335 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 238 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 335 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 238 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: float), _col2 (type: double), _col3 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: float), VALUE._col1 (type: double), VALUE._col2 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 10 Data size: 335 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 238 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 335 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 238 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -4162,25 +4172,25 @@ STAGE PLANS: Map Operator Tree: TableScan alias: table_change_lower_to_higher_numeric_group_float - Statistics: Num rows: 10 Data size: 148 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: insert_num (type: int), c1 (type: double), b (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 10 Data size: 148 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 10 Data size: 148 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: double), _col2 (type: string) Execution mode: vectorized Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: double), VALUE._col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 10 Data size: 148 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 148 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 4 Data size: 78 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/show_create_table_alter.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/show_create_table_alter.q.out b/ql/src/test/results/clientpositive/show_create_table_alter.q.out index 32819ea..d09f30b 100644 --- a/ql/src/test/results/clientpositive/show_create_table_alter.q.out +++ b/ql/src/test/results/clientpositive/show_create_table_alter.q.out @@ -35,6 +35,11 @@ OUTPUTFORMAT LOCATION #### A masked pattern was here #### TBLPROPERTIES ( + 'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}', + 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', + 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: -- Add a comment to the table, change the EXTERNAL property, and test SHOW CREATE TABLE on the change. ALTER TABLE tmp_showcrt1 SET TBLPROPERTIES ('comment'='temporary table', 'EXTERNAL'='FALSE') @@ -73,6 +78,8 @@ TBLPROPERTIES ( 'EXTERNAL'='FALSE', #### A masked pattern was here #### 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: -- Alter the table comment, change the EXTERNAL property back and test SHOW CREATE TABLE on the change. @@ -111,6 +118,8 @@ LOCATION TBLPROPERTIES ( #### A masked pattern was here #### 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: -- Change the 'SORTBUCKETCOLSPREFIX' property and test SHOW CREATE TABLE. The output should not change. @@ -149,6 +158,8 @@ LOCATION TBLPROPERTIES ( #### A masked pattern was here #### 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: -- Alter the storage handler of the table, and test SHOW CREATE TABLE. @@ -187,6 +198,8 @@ LOCATION TBLPROPERTIES ( #### A masked pattern was here #### 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: DROP TABLE tmp_showcrt1 http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/show_create_table_db_table.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/show_create_table_db_table.q.out b/ql/src/test/results/clientpositive/show_create_table_db_table.q.out index 495f4b5..daf63e9 100644 --- a/ql/src/test/results/clientpositive/show_create_table_db_table.q.out +++ b/ql/src/test/results/clientpositive/show_create_table_db_table.q.out @@ -46,6 +46,11 @@ OUTPUTFORMAT LOCATION #### A masked pattern was here #### TBLPROPERTIES ( + 'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}', + 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', + 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: DROP TABLE tmp_feng.tmp_showcrt PREHOOK: type: DROPTABLE http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/show_create_table_serde.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/show_create_table_serde.q.out b/ql/src/test/results/clientpositive/show_create_table_serde.q.out index 2350d98..a7bcb44 100644 --- a/ql/src/test/results/clientpositive/show_create_table_serde.q.out +++ b/ql/src/test/results/clientpositive/show_create_table_serde.q.out @@ -42,6 +42,8 @@ LOCATION TBLPROPERTIES ( #### A masked pattern was here #### 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: DROP TABLE tmp_showcrt1 @@ -90,6 +92,11 @@ OUTPUTFORMAT LOCATION #### A masked pattern was here #### TBLPROPERTIES ( + 'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}', + 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', + 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: DROP TABLE tmp_showcrt1 PREHOOK: type: DROPTABLE @@ -139,6 +146,11 @@ OUTPUTFORMAT LOCATION #### A masked pattern was here #### TBLPROPERTIES ( + 'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}', + 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', + 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: DROP TABLE tmp_showcrt1 PREHOOK: type: DROPTABLE @@ -183,6 +195,11 @@ WITH SERDEPROPERTIES ( LOCATION #### A masked pattern was here #### TBLPROPERTIES ( + 'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}', + 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', + 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: DROP TABLE tmp_showcrt1 PREHOOK: type: DROPTABLE http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/show_tblproperties.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/show_tblproperties.q.out b/ql/src/test/results/clientpositive/show_tblproperties.q.out index 63bbe6d..e1c6670 100644 --- a/ql/src/test/results/clientpositive/show_tblproperties.q.out +++ b/ql/src/test/results/clientpositive/show_tblproperties.q.out @@ -39,6 +39,8 @@ POSTHOOK: type: SHOW_TBLPROPERTIES bar bar value #### A masked pattern was here #### numFiles 0 +numRows 0 +rawDataSize 0 tmp true totalSize 0 #### A masked pattern was here #### @@ -54,6 +56,8 @@ POSTHOOK: type: SHOW_TBLPROPERTIES bar bar value #### A masked pattern was here #### numFiles 0 +numRows 0 +rawDataSize 0 tmp true totalSize 0 #### A masked pattern was here #### @@ -107,6 +111,8 @@ POSTHOOK: type: SHOW_TBLPROPERTIES bar bar value #### A masked pattern was here #### numFiles 0 +numRows 0 +rawDataSize 0 tmp true totalSize 0 #### A masked pattern was here #### @@ -124,6 +130,8 @@ POSTHOOK: type: SHOW_TBLPROPERTIES bar bar value1 #### A masked pattern was here #### numFiles 0 +numRows 0 +rawDataSize 0 tmp true1 totalSize 0 #### A masked pattern was here #### @@ -147,6 +155,8 @@ POSTHOOK: type: SHOW_TBLPROPERTIES bar bar value1 #### A masked pattern was here #### numFiles 0 +numRows 0 +rawDataSize 0 tmp true1 totalSize 0 #### A masked pattern was here #### http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out index dfb571d..e2be217 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_1.q.out @@ -174,8 +174,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -219,8 +221,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -374,8 +378,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -419,8 +425,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -564,8 +572,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -665,8 +675,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -710,8 +722,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out index 76ea0a8..eff3671 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_12.q.out @@ -194,8 +194,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -266,8 +268,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_medium numFiles 3 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -331,8 +335,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_medium numFiles 3 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -439,8 +445,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -484,8 +492,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out index c942d11..e09df8c 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_3.q.out @@ -154,8 +154,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -308,8 +310,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -452,8 +456,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -497,8 +503,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -599,8 +607,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out index ba2547a..a59c8a4 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_4.q.out @@ -170,8 +170,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -324,8 +326,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -468,8 +472,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -513,8 +519,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -615,8 +623,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out index 99fca25..04e5f40 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_5.q.out @@ -140,6 +140,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -159,6 +161,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -285,6 +289,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -304,6 +310,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -420,6 +428,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows 0 + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -439,6 +449,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows 0 + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -518,6 +530,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -537,6 +551,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out index 00c601f..6f8307b 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_7.q.out @@ -187,8 +187,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -232,8 +234,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -389,8 +393,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -434,8 +440,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -581,8 +589,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -626,8 +636,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -728,8 +740,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -773,8 +787,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out index 5564ceb..51f71c0 100644 --- a/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out +++ b/ql/src/test/results/clientpositive/spark/auto_sortmerge_join_8.q.out @@ -187,8 +187,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -232,8 +234,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -389,8 +393,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -434,8 +440,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -583,8 +591,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -628,8 +638,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_small numFiles 2 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -730,8 +742,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -775,8 +789,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.bucket_big numFiles 4 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe