http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out index d40a693..de1305f 100644 --- a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out +++ b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.java1.7.out @@ -439,6 +439,8 @@ Table: list_bucketing_dynamic_part #### A masked pattern was here #### Partition Parameters: numFiles 3 + numRows 984 + rawDataSize 9488 totalSize 10586 #### A masked pattern was here #### @@ -555,8 +557,10 @@ STAGE PLANS: #### A masked pattern was here #### name default.list_bucketing_dynamic_part numFiles 3 + numRows 984 partition_columns ds/hr partition_columns.types string:string + rawDataSize 9488 serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -585,16 +589,16 @@ STAGE PLANS: Processor Tree: TableScan alias: list_bucketing_dynamic_part - Statistics: Num rows: 16 Data size: 136 Basic stats: PARTIAL Column stats: NONE + Statistics: Num rows: 1000 Data size: 9624 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 2406 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: '484' (type: string), 'val_484' (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 34 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 250 Data size: 2406 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: select * from list_bucketing_dynamic_part where key = '484' and value = 'val_484'
http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out b/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out index 8f5492f..65d896b 100644 --- a/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out +++ b/ql/src/test/results/clientpositive/mapjoin_memcheck.q.out @@ -44,14 +44,14 @@ STAGE PLANS: $hdt$_0:src1 TableScan alias: src1 - Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) - Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string), value (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator keys: 0 _col0 (type: string) @@ -62,14 +62,14 @@ STAGE PLANS: Map Operator Tree: TableScan alias: src1 - Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: key is not null (type: boolean) - Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string), value (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 70 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 80 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -77,10 +77,10 @@ STAGE PLANS: 0 _col0 (type: string) 1 _col0 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 11 Data size: 77 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 11 Data size: 77 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/metadataonly1.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/metadataonly1.q.out b/ql/src/test/results/clientpositive/metadataonly1.q.out index 0281407..4b4c57c 100644 --- a/ql/src/test/results/clientpositive/metadataonly1.q.out +++ b/ql/src/test/results/clientpositive/metadataonly1.q.out @@ -132,17 +132,22 @@ STAGE PLANS: partition values: ds 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -256,17 +261,22 @@ STAGE PLANS: partition values: ds 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -380,17 +390,22 @@ STAGE PLANS: partition values: ds 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -515,17 +530,22 @@ STAGE PLANS: partition values: ds 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -554,17 +574,22 @@ STAGE PLANS: partition values: ds 2 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -678,17 +703,22 @@ STAGE PLANS: partition values: ds 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -718,17 +748,22 @@ STAGE PLANS: partition values: ds 2 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -948,17 +983,22 @@ STAGE PLANS: ds 1 hr 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -988,17 +1028,22 @@ STAGE PLANS: ds 1 hr 2 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -1028,17 +1073,22 @@ STAGE PLANS: ds 1 hr 3 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -1163,17 +1213,22 @@ STAGE PLANS: ds 1 hr 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1204,17 +1259,22 @@ STAGE PLANS: ds 1 hr 2 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1245,17 +1305,22 @@ STAGE PLANS: ds 1 hr 3 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1375,17 +1440,22 @@ STAGE PLANS: partition values: ds 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -1414,17 +1484,22 @@ STAGE PLANS: partition values: ds 2 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test1 + numFiles 0 + numRows 0 partition_columns ds partition_columns.types string + rawDataSize 0 serialization.ddl struct test1 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -1598,17 +1673,22 @@ STAGE PLANS: ds 01:10:10 hr 01 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -1638,17 +1718,22 @@ STAGE PLANS: ds 01:10:20 hr 02 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -1678,17 +1763,22 @@ STAGE PLANS: ds 1 hr 1 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -1718,17 +1808,22 @@ STAGE PLANS: ds 1 hr 2 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -1758,17 +1853,22 @@ STAGE PLANS: ds 1 hr 3 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns a,b columns.comments columns.types int:double #### A masked pattern was here #### name default.test2 + numFiles 0 + numRows 0 partition_columns ds/hr partition_columns.types string:string + rawDataSize 0 serialization.ddl struct test2 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/nullformat.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/nullformat.q.out b/ql/src/test/results/clientpositive/nullformat.q.out index af91470..2a80359 100644 --- a/ql/src/test/results/clientpositive/nullformat.q.out +++ b/ql/src/test/results/clientpositive/nullformat.q.out @@ -93,6 +93,11 @@ OUTPUTFORMAT LOCATION #### A masked pattern was here #### TBLPROPERTIES ( + 'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}', + 'numFiles'='0', + 'numRows'='0', + 'rawDataSize'='0', + 'totalSize'='0', #### A masked pattern was here #### PREHOOK: query: -- load null data from another table and verify that the null is stored in the expected format INSERT OVERWRITE TABLE null_tab1 SELECT a,b FROM base_tab http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_create.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/orc_create.q.out b/ql/src/test/results/clientpositive/orc_create.q.out index 3013fda..20c3fce 100644 --- a/ql/src/test/results/clientpositive/orc_create.q.out +++ b/ql/src/test/results/clientpositive/orc_create.q.out @@ -70,6 +70,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information @@ -319,6 +324,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_llap.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/orc_llap.q.out b/ql/src/test/results/clientpositive/orc_llap.q.out index 6fc73b7..bae69bb 100644 --- a/ql/src/test/results/clientpositive/orc_llap.q.out +++ b/ql/src/test/results/clientpositive/orc_llap.q.out @@ -719,17 +719,17 @@ STAGE PLANS: TableScan alias: orc_llap filterExpr: ((cint > 10) and cbigint is not null) (type: boolean) - Statistics: Num rows: 98779 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((cint > 10) and cbigint is not null) (type: boolean) - Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cint (type: int), csmallint (type: smallint), cbigint (type: bigint) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 32926 Data size: 526817 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -792,17 +792,17 @@ STAGE PLANS: TableScan alias: orc_llap filterExpr: ((cint > 10) and cbigint is not null) (type: boolean) - Statistics: Num rows: 4938 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((cint > 10) and cbigint is not null) (type: boolean) - Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ctinyint (type: tinyint), csmallint (type: smallint), cint (type: int), cbigint (type: bigint), cfloat (type: float), cdouble (type: double), cstring1 (type: string), cstring2 (type: string), ctimestamp1 (type: timestamp), ctimestamp2 (type: timestamp), cboolean1 (type: boolean), cboolean2 (type: boolean) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 - Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1646 Data size: 526823 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 81920 Data size: 19386626 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -874,17 +874,17 @@ STAGE PLANS: TableScan alias: orc_llap filterExpr: ((cint > 5) and (cint < 10)) (type: boolean) - Statistics: Num rows: 15196 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((cint > 5) and (cint < 10)) (type: boolean) - Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cstring2 (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1688 Data size: 175561 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 27306 Data size: 6462051 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -944,22 +944,22 @@ STAGE PLANS: Map Operator Tree: TableScan alias: orc_llap - Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cstring1 (type: string), cstring2 (type: string) outputColumnNames: cstring1, cstring2 - Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() keys: cstring1 (type: string), cstring2 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 7902 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE value expressions: _col2 (type: bigint) Execution mode: vectorized LLAP IO: all inputs @@ -969,10 +969,10 @@ STAGE PLANS: keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3951 Data size: 790234 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 122880 Data size: 29079940 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 3951 Data size: 790234 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 122880 Data size: 29079940 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1039,14 +1039,14 @@ STAGE PLANS: TableScan alias: o1 filterExpr: (csmallint is not null and cbigint is not null) (type: boolean) - Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (csmallint is not null and cbigint is not null) (type: boolean) - Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: csmallint (type: smallint), cstring1 (type: string) outputColumnNames: _col0, _col2 - Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator keys: 0 _col0 (type: smallint) @@ -1058,14 +1058,14 @@ STAGE PLANS: TableScan alias: o1 filterExpr: (csmallint is not null and cbigint is not null) (type: boolean) - Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (csmallint is not null and cbigint is not null) (type: boolean) - Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: csmallint (type: smallint), cstring2 (type: string) outputColumnNames: _col0, _col2 - Statistics: Num rows: 14111 Data size: 1580469 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 245760 Data size: 58159880 Basic stats: COMPLETE Column stats: NONE Map Join Operator condition map: Inner Join 0 to 1 @@ -1073,14 +1073,14 @@ STAGE PLANS: 0 _col0 (type: smallint) 1 _col0 (type: smallint) outputColumnNames: _col2, _col5 - Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col2 (type: string), _col5 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 15522 Data size: 1738515 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 270336 Data size: 63975869 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out b/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out index 7b361b7..38321e9 100644 --- a/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out +++ b/ql/src/test/results/clientpositive/orc_predicate_pushdown.q.out @@ -135,11 +135,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: orc_pred - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: hash(t) (type: int) outputColumnNames: _col0 - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col0) mode: hash @@ -183,11 +183,11 @@ STAGE PLANS: Map Operator Tree: TableScan alias: orc_pred - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: hash(t) (type: int) outputColumnNames: _col0 - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col0) mode: hash @@ -311,14 +311,14 @@ STAGE PLANS: Map Operator Tree: TableScan alias: orc_pred - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean) - Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: hash(t) (type: int) outputColumnNames: _col0 - Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col0) mode: hash @@ -369,14 +369,14 @@ STAGE PLANS: TableScan alias: orc_pred filterExpr: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean) - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6037 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((t < 0) and (UDFToInteger(t) > -2)) (type: boolean) - Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: hash(t) (type: int) outputColumnNames: _col0 - Statistics: Num rows: 116 Data size: 34409 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 670 Data size: 2680 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col0) mode: hash @@ -458,17 +458,17 @@ STAGE PLANS: Map Operator Tree: TableScan alias: orc_pred - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean) - Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: -1 (type: tinyint), s (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -501,17 +501,17 @@ STAGE PLANS: TableScan alias: orc_pred filterExpr: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean) - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((t = -1) and s is not null and (s like 'bob%')) (type: boolean) - Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: -1 (type: tinyint), s (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 262 Data size: 77718 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 58 Data size: 6037 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -589,26 +589,26 @@ STAGE PLANS: Map Operator Tree: TableScan alias: orc_pred - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean) - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: t (type: tinyint), s (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: tinyint), _col1 (type: string) sort order: ++ - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: tinyint), KEY.reducesinkkey1 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -645,26 +645,26 @@ STAGE PLANS: TableScan alias: orc_pred filterExpr: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean) - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 232 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (s is not null and (s like 'bob%') and (not (t) IN (-1, -2, -3)) and t BETWEEN 25 AND 30) (type: boolean) - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: t (type: tinyint), s (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: tinyint), _col1 (type: string) sort order: ++ - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: tinyint), KEY.reducesinkkey1 (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 131 Data size: 38859 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 29 Data size: 3018 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -766,31 +766,31 @@ STAGE PLANS: Map Operator Tree: TableScan alias: orc_pred - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean) - Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col3 (type: string) sort order: - - Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE TopN Hash Memory Usage: 0.1 value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double) Reduce Operator Tree: Select Operator expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 3 - Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -835,31 +835,31 @@ STAGE PLANS: TableScan alias: orc_pred filterExpr: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean) - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((d >= 10.0) and (d < 12.0) and (s like '%son') and (t > 0) and si BETWEEN 300 AND 400 and (not (s like '%car%'))) (type: boolean) - Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col3 (type: string) sort order: - - Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE TopN Hash Memory Usage: 0.1 value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double) Reduce Operator Tree: Select Operator expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 5 Data size: 1483 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 3 - Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 3 Data size: 888 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -970,28 +970,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: orc_pred - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean) - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col3 (type: string) sort order: - - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE TopN Hash Memory Usage: 0.1 value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double) Reduce Operator Tree: Select Operator expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false table: @@ -1006,20 +1006,20 @@ STAGE PLANS: Reduce Output Operator key expressions: _col3 (type: string) sort order: - - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE TopN Hash Memory Usage: 0.1 value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double) Reduce Operator Tree: Select Operator expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -1069,28 +1069,28 @@ STAGE PLANS: TableScan alias: orc_pred filterExpr: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean) - Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 208 Data size: 24150 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: ((t > 10) and (t <> 101) and (d >= 10.0) and (d < 12.0) and (s like '%son') and (not (s like '%car%')) and (t > 0) and si BETWEEN 300 AND 400) (type: boolean) - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: t (type: tinyint), si (type: smallint), d (type: double), s (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col3 (type: string) sort order: - - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE TopN Hash Memory Usage: 0.1 value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double) Reduce Operator Tree: Select Operator expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false table: @@ -1105,20 +1105,20 @@ STAGE PLANS: Reduce Output Operator key expressions: _col3 (type: string) sort order: - - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE TopN Hash Memory Usage: 0.1 value expressions: _col0 (type: tinyint), _col1 (type: smallint), _col2 (type: double) Reduce Operator Tree: Select Operator expressions: VALUE._col0 (type: tinyint), VALUE._col1 (type: smallint), VALUE._col2 (type: double), KEY.reducesinkkey0 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 3 - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 296 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 116 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_array_null_element.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/parquet_array_null_element.q.out b/ql/src/test/results/clientpositive/parquet_array_null_element.q.out index 387f01e..75d2d27 100644 --- a/ql/src/test/results/clientpositive/parquet_array_null_element.q.out +++ b/ql/src/test/results/clientpositive/parquet_array_null_element.q.out @@ -70,6 +70,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_create.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/parquet_create.q.out b/ql/src/test/results/clientpositive/parquet_create.q.out index c6d33ff..cc4b735 100644 --- a/ql/src/test/results/clientpositive/parquet_create.q.out +++ b/ql/src/test/results/clientpositive/parquet_create.q.out @@ -73,6 +73,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out b/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out index d6affd6..e96aa80 100644 --- a/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out +++ b/ql/src/test/results/clientpositive/parquet_mixed_partition_formats.q.out @@ -126,6 +126,8 @@ Table: parquet_mixed_partition_formats #### A masked pattern was here #### Partition Parameters: numFiles 1 + numRows 0 + rawDataSize 0 totalSize 2521 #### A masked pattern was here #### @@ -247,6 +249,8 @@ Table: parquet_mixed_partition_formats #### A masked pattern was here #### Partition Parameters: numFiles 1 + numRows 0 + rawDataSize 0 totalSize 2521 #### A masked pattern was here #### http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/parquet_serde.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/parquet_serde.q.out b/ql/src/test/results/clientpositive/parquet_serde.q.out index c1e594a..6d5f0f8 100644 --- a/ql/src/test/results/clientpositive/parquet_serde.q.out +++ b/ql/src/test/results/clientpositive/parquet_serde.q.out @@ -72,6 +72,8 @@ Table: parquet_mixed_fileformat #### A masked pattern was here #### Partition Parameters: numFiles 1 + numRows 0 + rawDataSize 0 totalSize 36 #### A masked pattern was here #### @@ -173,6 +175,8 @@ Table: parquet_mixed_fileformat #### A masked pattern was here #### Partition Parameters: numFiles 1 + numRows 0 + rawDataSize 0 totalSize 36 #### A masked pattern was here #### http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out index 82f4750..876d0db 100644 --- a/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out +++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props.q.out @@ -38,8 +38,13 @@ Database: default Table: mytbl #### A masked pattern was here #### Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} a myval b yourval + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out index e4b8003..95c2d6e 100644 --- a/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out +++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props_empty.q.out @@ -34,6 +34,11 @@ Database: default Table: mytbl #### A masked pattern was here #### Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out b/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out index 106448a..0be588a 100644 --- a/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out +++ b/ql/src/test/results/clientpositive/part_inherit_tbl_props_with_star.q.out @@ -38,9 +38,14 @@ Database: default Table: mytbl #### A masked pattern was here #### Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} a myval b yourval c noval + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/partition_coltype_literals.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out index bc159eb..b7a09d2 100644 --- a/ql/src/test/results/clientpositive/partition_coltype_literals.q.out +++ b/ql/src/test/results/clientpositive/partition_coltype_literals.q.out @@ -45,6 +45,11 @@ Database: default Table: partcoltypenum #### A masked pattern was here #### Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 0 + numRows 0 + rawDataSize 0 + totalSize 0 #### A masked pattern was here #### # Storage Information @@ -92,8 +97,11 @@ Database: default Table: partcoltypenum #### A masked pattern was here #### Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} #### A masked pattern was here #### numFiles 0 + numRows 0 + rawDataSize 0 totalSize 0 #### A masked pattern was here #### @@ -142,8 +150,11 @@ Database: default Table: partcoltypenum #### A masked pattern was here #### Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} #### A masked pattern was here #### numFiles 0 + numRows 0 + rawDataSize 0 totalSize 0 #### A masked pattern was here #### @@ -193,8 +204,11 @@ Database: default Table: partcoltypenum #### A masked pattern was here #### Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} #### A masked pattern was here #### numFiles 0 + numRows 0 + rawDataSize 0 totalSize 0 #### A masked pattern was here #### http://git-wip-us.apache.org/repos/asf/hive/blob/244ce09c/ql/src/test/results/clientpositive/pcr.q.out ---------------------------------------------------------------------- diff --git a/ql/src/test/results/clientpositive/pcr.q.out b/ql/src/test/results/clientpositive/pcr.q.out index b53226e..7222617 100644 --- a/ql/src/test/results/clientpositive/pcr.q.out +++ b/ql/src/test/results/clientpositive/pcr.q.out @@ -3419,15 +3419,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t2 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 @@ -3449,15 +3454,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t3 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 @@ -3533,15 +3543,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t2 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 @@ -3564,15 +3579,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t2 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 @@ -3588,30 +3608,40 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t2 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t2 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 @@ -3633,15 +3663,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t2 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 @@ -3657,30 +3692,40 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t2 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t2 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 @@ -3712,15 +3757,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t3 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 @@ -3743,15 +3793,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t3 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 @@ -3767,30 +3822,40 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t3 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t3 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 @@ -3812,15 +3877,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t3 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 @@ -3836,30 +3906,40 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t3 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 columns key,value columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t3 + numFiles 0 + numRows 0 + rawDataSize 0 serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3