http://git-wip-us.apache.org/repos/asf/hive/blob/6f5c1135/ql/src/test/results/clientpositive/bucket4.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/bucket4.q.out b/ql/src/test/results/clientpositive/bucket4.q.out deleted file mode 100644 index d06c7ed..0000000 --- a/ql/src/test/results/clientpositive/bucket4.q.out +++ /dev/null @@ -1,467 +0,0 @@ -PREHOOK: query: CREATE TABLE bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucket4_1 -POSTHOOK: query: CREATE TABLE bucket4_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucket4_1 -PREHOOK: query: explain extended -insert overwrite table bucket4_1 -select * from src -PREHOOK: type: QUERY -POSTHOOK: query: explain extended -insert overwrite table bucket4_1 -select * from src -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: UDFToInteger(_col0) (type: int) - null sort order: a - sort order: + - Map-reduce partition columns: UDFToInteger(_col0) (type: int) - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: string), _col1 (type: string) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [src] - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: UDFToInteger(VALUE._col0) (type: int), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 2 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.bucket4_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket4_1 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.bucket4_1 - TotalFiles: 2 - GatherStats: true - MultiFileSpray: true - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.bucket4_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket4_1 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.bucket4_1 - - Stage: Stage-2 - Stats-Aggr Operator -#### A masked pattern was here #### - -PREHOOK: query: insert overwrite table bucket4_1 -select * from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@bucket4_1 -POSTHOOK: query: insert overwrite table bucket4_1 -select * from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@bucket4_1 -POSTHOOK: Lineage: bucket4_1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: bucket4_1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain -select * from bucket4_1 tablesample (bucket 1 out of 2) s -PREHOOK: type: QUERY -POSTHOOK: query: explain -select * from bucket4_1 tablesample (bucket 1 out of 2) s -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: s - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (((hash(key) & 2147483647) % 2) = 0) (type: boolean) - Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from bucket4_1 tablesample (bucket 1 out of 2) s -PREHOOK: type: QUERY -PREHOOK: Input: default@bucket4_1 -#### A masked pattern was here #### -POSTHOOK: query: select * from bucket4_1 tablesample (bucket 1 out of 2) s -POSTHOOK: type: QUERY -POSTHOOK: Input: default@bucket4_1 -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -2 val_2 -4 val_4 -8 val_8 -10 val_10 -12 val_12 -12 val_12 -18 val_18 -18 val_18 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -28 val_28 -30 val_30 -34 val_34 -42 val_42 -42 val_42 -44 val_44 -54 val_54 -58 val_58 -58 val_58 -64 val_64 -66 val_66 -70 val_70 -70 val_70 -70 val_70 -72 val_72 -72 val_72 -74 val_74 -76 val_76 -76 val_76 -78 val_78 -80 val_80 -82 val_82 -84 val_84 -84 val_84 -86 val_86 -90 val_90 -90 val_90 -90 val_90 -92 val_92 -96 val_96 -98 val_98 -98 val_98 -100 val_100 -100 val_100 -104 val_104 -104 val_104 -114 val_114 -116 val_116 -118 val_118 -118 val_118 -120 val_120 -120 val_120 -126 val_126 -128 val_128 -128 val_128 -128 val_128 -134 val_134 -134 val_134 -136 val_136 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -146 val_146 -146 val_146 -150 val_150 -152 val_152 -152 val_152 -156 val_156 -158 val_158 -160 val_160 -162 val_162 -164 val_164 -164 val_164 -166 val_166 -168 val_168 -170 val_170 -172 val_172 -172 val_172 -174 val_174 -174 val_174 -176 val_176 -176 val_176 -178 val_178 -180 val_180 -186 val_186 -190 val_190 -192 val_192 -194 val_194 -196 val_196 -200 val_200 -200 val_200 -202 val_202 -208 val_208 -208 val_208 -208 val_208 -214 val_214 -216 val_216 -216 val_216 -218 val_218 -222 val_222 -224 val_224 -224 val_224 -226 val_226 -228 val_228 -230 val_230 -230 val_230 -230 val_230 -230 val_230 -230 val_230 -238 val_238 -238 val_238 -242 val_242 -242 val_242 -244 val_244 -248 val_248 -252 val_252 -256 val_256 -256 val_256 -258 val_258 -260 val_260 -262 val_262 -266 val_266 -272 val_272 -272 val_272 -274 val_274 -278 val_278 -278 val_278 -280 val_280 -280 val_280 -282 val_282 -282 val_282 -284 val_284 -286 val_286 -288 val_288 -288 val_288 -292 val_292 -296 val_296 -298 val_298 -298 val_298 -298 val_298 -302 val_302 -306 val_306 -308 val_308 -310 val_310 -316 val_316 -316 val_316 -316 val_316 -318 val_318 -318 val_318 -318 val_318 -322 val_322 -322 val_322 -332 val_332 -336 val_336 -338 val_338 -342 val_342 -342 val_342 -344 val_344 -344 val_344 -348 val_348 -348 val_348 -348 val_348 -348 val_348 -348 val_348 -356 val_356 -360 val_360 -362 val_362 -364 val_364 -366 val_366 -368 val_368 -374 val_374 -378 val_378 -382 val_382 -382 val_382 -384 val_384 -384 val_384 -384 val_384 -386 val_386 -392 val_392 -394 val_394 -396 val_396 -396 val_396 -396 val_396 -400 val_400 -402 val_402 -404 val_404 -404 val_404 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -414 val_414 -414 val_414 -418 val_418 -424 val_424 -424 val_424 -430 val_430 -430 val_430 -430 val_430 -432 val_432 -436 val_436 -438 val_438 -438 val_438 -438 val_438 -444 val_444 -446 val_446 -448 val_448 -452 val_452 -454 val_454 -454 val_454 -454 val_454 -458 val_458 -458 val_458 -460 val_460 -462 val_462 -462 val_462 -466 val_466 -466 val_466 -466 val_466 -468 val_468 -468 val_468 -468 val_468 -468 val_468 -470 val_470 -472 val_472 -478 val_478 -478 val_478 -480 val_480 -480 val_480 -480 val_480 -482 val_482 -484 val_484 -490 val_490 -492 val_492 -492 val_492 -494 val_494 -496 val_496 -498 val_498 -498 val_498 -498 val_498
http://git-wip-us.apache.org/repos/asf/hive/blob/6f5c1135/ql/src/test/results/clientpositive/bucket5.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/bucket5.q.out b/ql/src/test/results/clientpositive/bucket5.q.out deleted file mode 100644 index 30994a7..0000000 --- a/ql/src/test/results/clientpositive/bucket5.q.out +++ /dev/null @@ -1,610 +0,0 @@ -PREHOOK: query: -- Tests that when a multi insert inserts into a bucketed table and a table which is not bucketed --- the bucketed table is not merged and the table which is not bucketed is - -CREATE TABLE bucketed_table(key INT, value STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucketed_table -POSTHOOK: query: -- Tests that when a multi insert inserts into a bucketed table and a table which is not bucketed --- the bucketed table is not merged and the table which is not bucketed is - -CREATE TABLE bucketed_table(key INT, value STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucketed_table -PREHOOK: query: CREATE TABLE unbucketed_table(key INT, value STRING) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@unbucketed_table -POSTHOOK: query: CREATE TABLE unbucketed_table(key INT, value STRING) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@unbucketed_table -PREHOOK: query: EXPLAIN EXTENDED -FROM src -INSERT OVERWRITE TABLE bucketed_table SELECT key, value -INSERT OVERWRITE TABLE unbucketed_table SELECT key, value cluster by key -PREHOOK: type: QUERY -POSTHOOK: query: EXPLAIN EXTENDED -FROM src -INSERT OVERWRITE TABLE bucketed_table SELECT key, value -INSERT OVERWRITE TABLE unbucketed_table SELECT key, value cluster by key -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0 - Stage-4 depends on stages: Stage-2 - Stage-10 depends on stages: Stage-4 , consists of Stage-7, Stage-6, Stage-8 - Stage-7 - Stage-1 depends on stages: Stage-7, Stage-6, Stage-9 - Stage-5 depends on stages: Stage-1 - Stage-6 - Stage-8 - Stage-9 depends on stages: Stage-8 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: UDFToInteger(_col0) (type: int) - null sort order: a - sort order: + - Map-reduce partition columns: UDFToInteger(_col0) (type: int) - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: string), _col1 (type: string) - auto parallelism: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - columns _col0,_col1 - columns.types string,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [src] - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: UDFToInteger(VALUE._col0) (type: int), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.bucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.bucketed_table - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.bucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.bucketed_table - - Stage: Stage-3 - Stats-Aggr Operator -#### A masked pattern was here #### - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: a - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col1 (type: string) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - columns _col0,_col1 - columns.types string,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - columns _col0,_col1 - columns.types string,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: UDFToInteger(KEY.reducesinkkey0) (type: int), VALUE._col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 2 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 29 Data size: 5812 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.unbucketed_table - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - - Stage: Stage-10 - Conditional Operator - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-1 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.unbucketed_table - - Stage: Stage-5 - Stats-Aggr Operator -#### A masked pattern was here #### - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.unbucketed_table - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10005 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.unbucketed_table - name: default.unbucketed_table - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-8 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.unbucketed_table - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10005 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.unbucketed_table - name: default.unbucketed_table - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-9 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM src -INSERT OVERWRITE TABLE bucketed_table SELECT key, value -INSERT OVERWRITE TABLE unbucketed_table SELECT key, value cluster by key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@bucketed_table -PREHOOK: Output: default@unbucketed_table -POSTHOOK: query: FROM src -INSERT OVERWRITE TABLE bucketed_table SELECT key, value -INSERT OVERWRITE TABLE unbucketed_table SELECT key, value cluster by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@bucketed_table -POSTHOOK: Output: default@unbucketed_table -POSTHOOK: Lineage: bucketed_table.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: bucketed_table.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: unbucketed_table.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: unbucketed_table.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: DESC FORMATTED bucketed_table -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@bucketed_table -POSTHOOK: query: DESC FORMATTED bucketed_table -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@bucketed_table -# col_name data_type comment - -key int -value string - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - SORTBUCKETCOLSPREFIX TRUE - numFiles 2 - numRows 500 - rawDataSize 5312 - totalSize 5812 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: 2 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: SELECT * FROM bucketed_table TABLESAMPLE (BUCKET 1 OUT OF 2) s LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@bucketed_table -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM bucketed_table TABLESAMPLE (BUCKET 1 OUT OF 2) s LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@bucketed_table -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -2 val_2 -4 val_4 -8 val_8 -10 val_10 -12 val_12 -12 val_12 -18 val_18 -PREHOOK: query: SELECT * FROM bucketed_table TABLESAMPLE (BUCKET 2 OUT OF 2) s LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@bucketed_table -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM bucketed_table TABLESAMPLE (BUCKET 2 OUT OF 2) s LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@bucketed_table -#### A masked pattern was here #### -5 val_5 -5 val_5 -5 val_5 -9 val_9 -11 val_11 -15 val_15 -15 val_15 -17 val_17 -19 val_19 -27 val_27 -PREHOOK: query: -- Should be 2 (not merged) -SELECT COUNT(DISTINCT INPUT__FILE__NAME) FROM bucketed_table -PREHOOK: type: QUERY -PREHOOK: Input: default@bucketed_table -#### A masked pattern was here #### -POSTHOOK: query: -- Should be 2 (not merged) -SELECT COUNT(DISTINCT INPUT__FILE__NAME) FROM bucketed_table -POSTHOOK: type: QUERY -POSTHOOK: Input: default@bucketed_table -#### A masked pattern was here #### -2 -PREHOOK: query: -- Should be 1 (merged) -SELECT COUNT(DISTINCT INPUT__FILE__NAME) FROM unbucketed_table -PREHOOK: type: QUERY -PREHOOK: Input: default@unbucketed_table -#### A masked pattern was here #### -POSTHOOK: query: -- Should be 1 (merged) -SELECT COUNT(DISTINCT INPUT__FILE__NAME) FROM unbucketed_table -POSTHOOK: type: QUERY -POSTHOOK: Input: default@unbucketed_table -#### A masked pattern was here #### -1 http://git-wip-us.apache.org/repos/asf/hive/blob/6f5c1135/ql/src/test/results/clientpositive/bucket6.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/bucket6.q.out b/ql/src/test/results/clientpositive/bucket6.q.out deleted file mode 100644 index c7f696f..0000000 --- a/ql/src/test/results/clientpositive/bucket6.q.out +++ /dev/null @@ -1,189 +0,0 @@ -PREHOOK: query: CREATE TABLE src_bucket(key STRING, value STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@src_bucket -POSTHOOK: query: CREATE TABLE src_bucket(key STRING, value STRING) CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@src_bucket -PREHOOK: query: explain -insert into table src_bucket select key,value from srcpart -PREHOOK: type: QUERY -POSTHOOK: query: explain -insert into table src_bucket select key,value from srcpart -POSTHOOK: type: QUERY -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - Statistics: Num rows: 116 Data size: 23248 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 116 Data size: 23248 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 116 Data size: 23248 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 116 Data size: 23248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 116 Data size: 23248 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src_bucket - - Stage: Stage-0 - Move Operator - tables: - replace: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src_bucket - - Stage: Stage-2 - Stats-Aggr Operator - -PREHOOK: query: insert into table src_bucket select key,value from srcpart -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -PREHOOK: Output: default@src_bucket -POSTHOOK: query: insert into table src_bucket select key,value from srcpart -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -POSTHOOK: Output: default@src_bucket -POSTHOOK: Lineage: src_bucket.key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: src_bucket.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from src_bucket limit 100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src_bucket -#### A masked pattern was here #### -POSTHOOK: query: select * from src_bucket limit 100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_bucket -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -103 val_103 -103 val_103 -103 val_103 -103 val_103 -103 val_103 -103 val_103 -103 val_103 -103 val_103 -105 val_105 -105 val_105 -105 val_105 -105 val_105 -11 val_11 -11 val_11 -11 val_11 -11 val_11 -114 val_114 -114 val_114 -114 val_114 -114 val_114 -116 val_116 -116 val_116 -116 val_116 -116 val_116 -118 val_118 -118 val_118 -118 val_118 -118 val_118 -118 val_118 -118 val_118 -118 val_118 -118 val_118 -125 val_125 -125 val_125 -125 val_125 -125 val_125 -125 val_125 -125 val_125 -125 val_125 -125 val_125 -129 val_129 -129 val_129 -129 val_129 -129 val_129 -129 val_129 -129 val_129 -129 val_129 -129 val_129 -134 val_134 -134 val_134 -134 val_134 -134 val_134 -134 val_134 -134 val_134 -134 val_134 -134 val_134 -136 val_136 -136 val_136 -136 val_136 -136 val_136 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -143 val_143 -143 val_143 -143 val_143 -143 val_143 -145 val_145 -145 val_145 -145 val_145 -145 val_145 -149 val_149 -149 val_149 -149 val_149 -149 val_149