Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_1.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_1.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_1.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_1.q.out Thu Oct 30 12:31:47 2014 @@ -23,7 +23,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_1 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_1 select cast('2011-01-01 01:01:01' as timestamp) from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -107,7 +106,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_1 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_1 select '2011-01-01 01:01:01' from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -191,7 +189,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_1 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_1 select '2011-01-01 01:01:01.1' from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -275,7 +272,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_1 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_1 select '2011-01-01 01:01:01.0001' from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -359,7 +355,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_1 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_1 select '2011-01-01 01:01:01.000100000' from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -443,7 +438,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_1 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_1 select '2011-01-01 01:01:01.001000011' from src tablesample (1 rows) POSTHOOK: type: QUERY
Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_2.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_2.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_2.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_2.q.out Thu Oct 30 12:31:47 2014 @@ -23,7 +23,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_2 select cast('2011-01-01 01:01:01' as timestamp) from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -107,7 +106,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_2 select '2011-01-01 01:01:01' from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -191,7 +189,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_2 select '2011-01-01 01:01:01.1' from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -275,7 +272,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_2 select '2011-01-01 01:01:01.0001' from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -359,7 +355,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_2 select '2011-01-01 01:01:01.000100000' from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -443,7 +438,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_2 select '2011-01-01 01:01:01.001000011' from src tablesample (1 rows) POSTHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_3.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_3.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_3.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_3.q.out Thu Oct 30 12:31:47 2014 @@ -23,7 +23,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_3 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_3 select cast(cast('1.3041352164485E9' as double) as timestamp) from src tablesample (1 rows) POSTHOOK: type: QUERY Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_lazy.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_lazy.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_lazy.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_lazy.q.out Thu Oct 30 12:31:47 2014 @@ -14,7 +14,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_lazy -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table timestamp_lazy select cast('2011-01-01 01:01:01' as timestamp), key, value from src tablesample (5 rows) POSTHOOK: type: QUERY POSTHOOK: Input: default@src Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_udf.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_udf.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_udf.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/timestamp_udf.q.out Thu Oct 30 12:31:47 2014 @@ -31,8 +31,6 @@ PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@timestamp_udf PREHOOK: Output: default@timestamp_udf_string -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: from (select * from src tablesample (1 rows)) s insert overwrite table timestamp_udf select '2011-05-06 07:08:09.1234567' Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/transform1.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/transform1.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/transform1.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/transform1.q.out Thu Oct 30 12:31:47 2014 @@ -72,7 +72,6 @@ select array(1,2,3) from src tablesample PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@transform1_t2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table transform1_t2 select array(1,2,3) from src tablesample (1 rows) POSTHOOK: type: QUERY @@ -98,21 +97,21 @@ STAGE PLANS: Map Operator Tree: TableScan alias: transform1_t2 - Statistics: Num rows: -1 Data size: 6 Basic stats: PARTIAL Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 5 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: '012' (type: string) outputColumnNames: _col0 - Statistics: Num rows: -1 Data size: 6 Basic stats: PARTIAL Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE Transform Operator command: cat output info: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Statistics: Num rows: -1 Data size: 6 Basic stats: PARTIAL Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: -1 Data size: 6 Basic stats: PARTIAL Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union10.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union10.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union10.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union10.q.out Thu Oct 30 12:31:47 2014 @@ -175,7 +175,6 @@ PREHOOK: query: insert overwrite table t PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@tmptable -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table tmptable select unionsrc.key, unionsrc.value FROM (select 'tst1' as key, count(1) as value from src s1 UNION ALL Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union18.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union18.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union18.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union18.q.out Thu Oct 30 12:31:47 2014 @@ -161,8 +161,6 @@ PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@dest1 PREHOOK: Output: default@dest2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1 UNION ALL select s2.key as key, s2.value as value from src s2) unionsrc Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union19.q.out Thu Oct 30 12:31:47 2014 @@ -188,8 +188,6 @@ PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@dest1 PREHOOK: Output: default@dest2 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1 UNION ALL select s2.key as key, s2.value as value from src s2) unionsrc Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union25.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union25.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union25.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union25.q.out Thu Oct 30 12:31:47 2014 @@ -12,7 +12,6 @@ PREHOOK: type: QUERY PREHOOK: Input: default@srcpart PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 PREHOOK: Output: default@tmp_srcpart@ds=2008-04-08/hr=11 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table tmp_srcpart partition (ds='2008-04-08', hr='11') select key, value from srcpart where ds='2008-04-08' and hr='11' POSTHOOK: type: QUERY @@ -159,14 +158,14 @@ STAGE PLANS: keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 279 Data size: 5562 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col2 (type: bigint), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 279 Data size: 5562 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 279 Data size: 5562 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union28.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union28.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union28.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union28.q.out Thu Oct 30 12:31:47 2014 @@ -185,7 +185,6 @@ select * from ( PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@union_subq_union -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table union_subq_union select * from ( select key, value from src Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union29.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union29.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union29.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union29.q.out Thu Oct 30 12:31:47 2014 @@ -135,7 +135,6 @@ select * from ( PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@union_subq_union -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table union_subq_union select * from ( select key, value from src Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union3.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union3.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union3.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union3.q.out Thu Oct 30 12:31:47 2014 @@ -238,7 +238,6 @@ FROM ( PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@union_out -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table union_out SELECT * FROM ( Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union30.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union30.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union30.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union30.q.out Thu Oct 30 12:31:47 2014 @@ -232,7 +232,6 @@ select key, value from src PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@union_subq_union -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table union_subq_union select * from ( Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union33.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union33.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union33.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union33.q.out Thu Oct 30 12:31:47 2014 @@ -151,7 +151,6 @@ UNION ALL PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@test_src -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: INSERT OVERWRITE TABLE test_src SELECT key, value FROM ( SELECT key, value FROM src @@ -311,7 +310,6 @@ UNION ALL PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@test_src -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: INSERT OVERWRITE TABLE test_src SELECT key, value FROM ( SELECT key, COUNT(*) AS value FROM src Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union4.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union4.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union4.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union4.q.out Thu Oct 30 12:31:47 2014 @@ -135,7 +135,6 @@ select unionsrc.key, unionsrc.value FROM PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@tmptable -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table tmptable select unionsrc.key, unionsrc.value FROM (select 'tst1' as key, count(1) as value from src s1 UNION ALL Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/union6.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/union6.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/union6.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/union6.q.out Thu Oct 30 12:31:47 2014 @@ -117,7 +117,6 @@ PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Input: default@src1 PREHOOK: Output: default@tmptable -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table tmptable select unionsrc.key, unionsrc.value FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1 UNION ALL Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_between_in.q.out Thu Oct 30 12:31:47 2014 @@ -3,7 +3,6 @@ PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@alltypesorc PREHOOK: Output: database:default PREHOOK: Output: default@decimal_date_test -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: CREATE TABLE decimal_date_test STORED AS ORC AS SELECT cdouble, CAST (((cdouble*22.1)/37) AS DECIMAL(20,10)) AS cdecimal1, CAST (((cdouble*9.3)/13) AS DECIMAL(23,14)) AS cdecimal2, CAST(CAST((CAST(cint AS BIGINT) *ctinyint) AS TIMESTAMP) AS DATE) AS cdate FROM alltypesorc ORDER BY cdate POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@alltypesorc @@ -28,28 +27,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_date_test - Statistics: Num rows: 1952 Data size: 109320 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2467616 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (cdate) IN (1969-10-26, 1969-07-14) (type: boolean) - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cdate (type: date) outputColumnNames: _col0 - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: date) sort order: + - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reducer 2 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: date) outputColumnNames: _col0 - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -81,12 +80,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_date_test - Statistics: Num rows: 1952 Data size: 109320 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2467616 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (not (cdate) IN (1969-10-26, 1969-07-14, 1970-01-21)) (type: boolean) - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Select Operator - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -142,28 +141,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_date_test - Statistics: Num rows: 976 Data size: 109320 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2467616 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (cdecimal1) IN (2365.8945945946, 881.0135135135, -3367.6517567568) (type: boolean) - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cdecimal1 (type: decimal(20,10)) outputColumnNames: _col0 - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: decimal(20,10)) sort order: + - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reducer 2 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: decimal(20,10)) outputColumnNames: _col0 - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -195,12 +194,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_date_test - Statistics: Num rows: 976 Data size: 109320 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2467616 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (not (cdecimal1) IN (2365.8945945946, 881.0135135135, -3367.6517567568)) (type: boolean) - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Select Operator - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash @@ -256,28 +255,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_date_test - Statistics: Num rows: 1952 Data size: 109320 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2467616 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: cdate BETWEEN 1969-12-30 AND 1970-01-02 (type: boolean) - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cdate (type: date) outputColumnNames: _col0 - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: date) sort order: + - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reducer 2 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: date) outputColumnNames: _col0 - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -309,28 +308,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_date_test - Statistics: Num rows: 1952 Data size: 109320 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2467616 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: cdate NOT BETWEEN 1968-05-01 AND 1971-09-01 (type: boolean) - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cdate (type: date) outputColumnNames: _col0 - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: date) sort order: + - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reducer 2 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: date) outputColumnNames: _col0 - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 976 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -362,28 +361,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_date_test - Statistics: Num rows: 976 Data size: 109320 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2467616 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: cdecimal1 BETWEEN -20 AND 45.9918918919 (type: boolean) - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cdecimal1 (type: decimal(20,10)) outputColumnNames: _col0 - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: decimal(20,10)) sort order: + - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reducer 2 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: decimal(20,10)) outputColumnNames: _col0 - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -415,12 +414,12 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_date_test - Statistics: Num rows: 976 Data size: 109320 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2467616 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: cdecimal1 NOT BETWEEN -2000 AND 4390.1351351351 (type: boolean) - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Select Operator - Statistics: Num rows: 488 Data size: 54660 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1233808 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() mode: hash Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_cast_constant.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_cast_constant.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_cast_constant.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_cast_constant.q.out Thu Oct 30 12:31:47 2014 @@ -82,7 +82,6 @@ PREHOOK: query: INSERT INTO TABLE over1k PREHOOK: type: QUERY PREHOOK: Input: default@over1k PREHOOK: Output: default@over1korc -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: INSERT INTO TABLE over1korc SELECT * FROM over1k POSTHOOK: type: QUERY POSTHOOK: Input: default@over1k @@ -128,22 +127,22 @@ STAGE PLANS: Map Operator Tree: TableScan alias: over1korc - Statistics: Num rows: 5898 Data size: 23594 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: i (type: int) outputColumnNames: i - Statistics: Num rows: 5898 Data size: 23594 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: avg(50), avg(UDFToDouble(50)), avg(CAST( 50 AS decimal(10,0))) keys: i (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 5898 Data size: 23594 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 5898 Data size: 23594 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: struct<count:bigint,sum:double,input:int>), _col2 (type: struct<count:bigint,sum:double,input:double>), _col3 (type: struct<count:bigint,sum:decimal(12,0),input:decimal(10,0)>) Execution mode: vectorized Reducer 2 @@ -153,28 +152,28 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2949 Data size: 11797 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: double), _col2 (type: double), _col3 (type: decimal(14,4)) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2949 Data size: 11797 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 2949 Data size: 11797 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: decimal(14,4)) Reducer 3 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: double), VALUE._col1 (type: double), VALUE._col2 (type: decimal(14,4)) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2949 Data size: 11797 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 524 Data size: 155436 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 10 - Statistics: Num rows: 10 Data size: 40 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 40 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10 Data size: 2960 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_char_4.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_char_4.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_char_4.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_char_4.q.out Thu Oct 30 12:31:47 2014 @@ -92,7 +92,6 @@ PREHOOK: query: INSERT INTO TABLE vector PREHOOK: type: QUERY PREHOOK: Input: default@vectortab2k PREHOOK: Output: default@vectortab2korc -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: INSERT INTO TABLE vectortab2korc SELECT * FROM vectortab2k POSTHOOK: type: QUERY POSTHOOK: Input: default@vectortab2k @@ -143,14 +142,14 @@ STAGE PLANS: Map Operator Tree: TableScan alias: vectortab2korc - Statistics: Num rows: 731 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: CAST( t AS CHAR(10) (type: char(10)), CAST( si AS CHAR(10) (type: char(10)), CAST( i AS CHAR(20) (type: char(20)), CAST( b AS CHAR(30) (type: char(30)), CAST( f AS CHAR(20) (type: char(20)), CAST( d AS CHAR(20) (type: char(20)), CAST( s AS CHAR(50) (type: char(50)) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 731 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 731 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_count_distinct.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_count_distinct.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_count_distinct.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_count_distinct.q.out Thu Oct 30 12:31:47 2014 @@ -197,7 +197,6 @@ select ws_sold_date_sk, ws_sold_time_sk, PREHOOK: type: QUERY PREHOOK: Input: default@web_sales_txt PREHOOK: Output: default@web_sales -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table web_sales partition (ws_web_site_sk) select ws_sold_date_sk, ws_sold_time_sk, ws_ship_date_sk, ws_item_sk, @@ -1255,21 +1254,21 @@ STAGE PLANS: Map Operator Tree: TableScan alias: web_sales - Statistics: Num rows: 58981 Data size: 235968 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 3504000 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: ws_order_number (type: int) outputColumnNames: ws_order_number - Statistics: Num rows: 58981 Data size: 235968 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 3504000 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT ws_order_number) keys: ws_order_number (type: int) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 58981 Data size: 235968 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 3504000 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + - Statistics: Num rows: 58981 Data size: 235968 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 3504000 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reducer 2 Reduce Operator Tree: Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_data_types.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_data_types.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== Files hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_data_types.q.out (original) and hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_data_types.q.out Thu Oct 30 12:31:47 2014 differ Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_aggregate.q.out Thu Oct 30 12:31:47 2014 @@ -7,7 +7,6 @@ PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@alltypesorc PREHOOK: Output: database:default PREHOOK: Output: default@decimal_vgby -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: CREATE TABLE decimal_vgby STORED AS ORC AS SELECT cdouble, CAST (((cdouble*22.1)/37) AS DECIMAL(20,10)) AS cdecimal1, CAST (((cdouble*9.3)/13) AS DECIMAL(23,14)) AS cdecimal2, @@ -48,22 +47,22 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_vgby - Statistics: Num rows: 559 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cint (type: int), cdecimal1 (type: decimal(20,10)), cdecimal2 (type: decimal(23,14)) outputColumnNames: cint, cdecimal1, cdecimal2 - Statistics: Num rows: 559 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(cdecimal1), max(cdecimal1), min(cdecimal1), sum(cdecimal1), count(cdecimal2), max(cdecimal2), min(cdecimal2), sum(cdecimal2), count() keys: cint (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 - Statistics: Num rows: 559 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 559 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: bigint), _col2 (type: decimal(20,10)), _col3 (type: decimal(20,10)), _col4 (type: decimal(30,10)), _col5 (type: bigint), _col6 (type: decimal(23,14)), _col7 (type: decimal(23,14)), _col8 (type: decimal(33,14)), _col9 (type: bigint) Execution mode: vectorized Reducer 2 @@ -73,17 +72,17 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 - Statistics: Num rows: 279 Data size: 63714 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1082530 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (_col9 > 1) (type: boolean) - Statistics: Num rows: 93 Data size: 21238 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2048 Data size: 360843 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: decimal(20,10)), _col3 (type: decimal(20,10)), _col4 (type: decimal(30,10)), _col5 (type: bigint), _col6 (type: decimal(23,14)), _col7 (type: decimal(23,14)), _col8 (type: decimal(33,14)) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 93 Data size: 21238 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2048 Data size: 360843 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 93 Data size: 21238 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2048 Data size: 360843 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -153,22 +152,22 @@ STAGE PLANS: Map Operator Tree: TableScan alias: decimal_vgby - Statistics: Num rows: 559 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cint (type: int), cdecimal1 (type: decimal(20,10)), cdecimal2 (type: decimal(23,14)) outputColumnNames: cint, cdecimal1, cdecimal2 - Statistics: Num rows: 559 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(cdecimal1), max(cdecimal1), min(cdecimal1), sum(cdecimal1), avg(cdecimal1), stddev_pop(cdecimal1), stddev_samp(cdecimal1), count(cdecimal2), max(cdecimal2), min(cdecimal2), sum(cdecimal2), avg(cdecimal2), stddev_pop(cdecimal2), stddev_samp(cdecimal2), count() keys: cint (type: int) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15 - Statistics: Num rows: 559 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 559 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: bigint), _col2 (type: decimal(20,10)), _col3 (type: decimal(20,10)), _col4 (type: decimal(30,10)), _col5 (type: struct<count:bigint,sum:decimal(30,10),input:decimal(20,10)>), _col6 (type: struct<count:bigint,sum:double,variance:double>), _col7 (type: struct<count:bigint,sum:double,variance:double>), _col8 (type: bigint), _col9 (type: decimal(23,14)), _col10 (type: decimal(23,14)), _col11 (type: decimal(33,14)), _col12 (type: struct<count:bigint,sum:decimal(33,14),input:decimal(23,14)>), _col13 (type: struct<count:bigint,sum:double,variance:double>), _col14 (type: struct<count:bigint,sum:double,variance:double>), _col15 (type: bigint) Execution mode: vectorized Reducer 2 @@ -178,17 +177,17 @@ STAGE PLANS: keys: KEY._col0 (type: int) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15 - Statistics: Num rows: 279 Data size: 63714 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 6144 Data size: 1082530 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (_col15 > 1) (type: boolean) - Statistics: Num rows: 93 Data size: 21238 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2048 Data size: 360843 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: int), _col1 (type: bigint), _col2 (type: decimal(20,10)), _col3 (type: decimal(20,10)), _col4 (type: decimal(30,10)), _col5 (type: decimal(24,14)), _col6 (type: double), _col7 (type: double), _col8 (type: bigint), _col9 (type: decimal(23,14)), _col10 (type: decimal(23,14)), _col11 (type: decimal(33,14)), _col12 (type: decimal(27,18)), _col13 (type: double), _col14 (type: double) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14 - Statistics: Num rows: 93 Data size: 21238 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2048 Data size: 360843 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 93 Data size: 21238 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2048 Data size: 360843 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_decimal_mapjoin.q.out Thu Oct 30 12:31:47 2014 @@ -9,7 +9,6 @@ PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@alltypesorc PREHOOK: Output: database:default PREHOOK: Output: default@decimal_mapjoin -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: -- SORT_QUERY_RESULTS CREATE TABLE decimal_mapjoin STORED AS ORC AS @@ -46,28 +45,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: r - Statistics: Num rows: 1100 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (cint is not null and (cint = 6981)) (type: boolean) - Statistics: Num rows: 275 Data size: 31914 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3072 Data size: 541265 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: 6981 (type: int) sort order: + - Statistics: Num rows: 275 Data size: 31914 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3072 Data size: 541265 Basic stats: COMPLETE Column stats: NONE value expressions: cdecimal2 (type: decimal(23,14)) Execution mode: vectorized Map 3 Map Operator Tree: TableScan alias: l - Statistics: Num rows: 1100 Data size: 127658 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 2165060 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (cint is not null and (cint = 6981)) (type: boolean) - Statistics: Num rows: 275 Data size: 31914 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3072 Data size: 541265 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: 6981 (type: int) sort order: + - Statistics: Num rows: 275 Data size: 31914 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3072 Data size: 541265 Basic stats: COMPLETE Column stats: NONE value expressions: cdecimal1 (type: decimal(20,10)) Execution mode: vectorized Reducer 2 @@ -79,14 +78,14 @@ STAGE PLANS: 0 {VALUE._col1} 1 {VALUE._col2} outputColumnNames: _col1, _col9 - Statistics: Num rows: 302 Data size: 35105 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3379 Data size: 595391 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: 6981 (type: int), 6981 (type: int), _col1 (type: decimal(20,10)), _col9 (type: decimal(23,14)) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 302 Data size: 35105 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3379 Data size: 595391 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 302 Data size: 35105 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3379 Data size: 595391 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_orderby_5.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_orderby_5.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_orderby_5.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_orderby_5.q.out Thu Oct 30 12:31:47 2014 @@ -84,7 +84,6 @@ PREHOOK: query: INSERT INTO TABLE vector PREHOOK: type: QUERY PREHOOK: Input: default@vectortab2k PREHOOK: Output: default@vectortab2korc -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: INSERT INTO TABLE vectortab2korc SELECT * FROM vectortab2k POSTHOOK: type: QUERY POSTHOOK: Input: default@vectortab2k @@ -124,22 +123,22 @@ STAGE PLANS: Map Operator Tree: TableScan alias: vectortab2korc - Statistics: Num rows: 8043 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: bo (type: boolean), b (type: bigint) outputColumnNames: bo, b - Statistics: Num rows: 8043 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(b) keys: bo (type: boolean) mode: hash outputColumnNames: _col0, _col1 - Statistics: Num rows: 8043 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: boolean) sort order: + Map-reduce partition columns: _col0 (type: boolean) - Statistics: Num rows: 8043 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: bigint) Execution mode: vectorized Reducer 2 @@ -149,15 +148,15 @@ STAGE PLANS: keys: KEY._col0 (type: boolean) mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 4021 Data size: 48252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: boolean), _col1 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 4021 Data size: 48252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: boolean) sort order: - - Statistics: Num rows: 4021 Data size: 48252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: bigint) Execution mode: vectorized Reducer 3 @@ -165,10 +164,10 @@ STAGE PLANS: Select Operator expressions: KEY.reducesinkkey0 (type: boolean), VALUE._col0 (type: bigint) outputColumnNames: _col0, _col1 - Statistics: Num rows: 4021 Data size: 48252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 4021 Data size: 48252 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1000 Data size: 459356 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_string_concat.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_string_concat.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_string_concat.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_string_concat.q.out Thu Oct 30 12:31:47 2014 @@ -82,7 +82,6 @@ PREHOOK: query: INSERT INTO TABLE over1k PREHOOK: type: QUERY PREHOOK: Input: default@over1k PREHOOK: Output: default@over1korc -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: INSERT INTO TABLE over1korc SELECT * FROM over1k POSTHOOK: type: QUERY POSTHOOK: Input: default@over1k @@ -118,14 +117,14 @@ STAGE PLANS: Processor Tree: TableScan alias: over1korc - Statistics: Num rows: 235 Data size: 23594 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: s (type: string), concat(concat(' ', s), ' ') (type: string), concat(concat('|', rtrim(concat(concat(' ', s), ' '))), '|') (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 235 Data size: 23594 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1049 Data size: 311170 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 20 - Statistics: Num rows: 20 Data size: 2000 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 20 Data size: 5920 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: SELECT s AS `string`, Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_varchar_4.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_varchar_4.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_varchar_4.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vector_varchar_4.q.out Thu Oct 30 12:31:47 2014 @@ -92,7 +92,6 @@ PREHOOK: query: INSERT INTO TABLE vector PREHOOK: type: QUERY PREHOOK: Input: default@vectortab2k PREHOOK: Output: default@vectortab2korc -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: INSERT INTO TABLE vectortab2korc SELECT * FROM vectortab2k POSTHOOK: type: QUERY POSTHOOK: Input: default@vectortab2k @@ -143,14 +142,14 @@ STAGE PLANS: Map Operator Tree: TableScan alias: vectortab2korc - Statistics: Num rows: 731 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: CAST( t AS varchar(10)) (type: varchar(10)), CAST( si AS varchar(10)) (type: varchar(10)), CAST( i AS varchar(20)) (type: varchar(20)), CAST( b AS varchar(30)) (type: varchar(30)), CAST( f AS varchar(20)) (type: varchar(20)), CAST( d AS varchar(20)) (type: varchar(20)), CAST( s AS varchar(50)) (type: varchar(50)) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 731 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 731 Data size: 96518 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2000 Data size: 918712 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_decimal_date.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_decimal_date.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_decimal_date.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_decimal_date.q.out Thu Oct 30 12:31:47 2014 @@ -3,7 +3,6 @@ PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@alltypesorc PREHOOK: Output: database:default PREHOOK: Output: default@date_decimal_test -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: CREATE TABLE date_decimal_test STORED AS ORC AS SELECT cint, cdouble, CAST (CAST (cint AS TIMESTAMP) AS DATE) AS cdate, CAST (((cdouble*22.1)/37) AS DECIMAL(20,10)) AS cdecimal FROM alltypesorc POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@alltypesorc @@ -23,17 +22,17 @@ STAGE PLANS: Processor Tree: TableScan alias: date_decimal_test - Statistics: Num rows: 482 Data size: 86777 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 12288 Data size: 1651260 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (cint is not null and cdouble is not null) (type: boolean) - Statistics: Num rows: 121 Data size: 21784 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3072 Data size: 412815 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: cdate (type: date), cdecimal (type: decimal(20,10)) outputColumnNames: _col0, _col1 - Statistics: Num rows: 121 Data size: 21784 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3072 Data size: 412815 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 10 - Statistics: Num rows: 10 Data size: 1800 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10 Data size: 1340 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: SELECT cdate, cdecimal from date_decimal_test where cint IS NOT NULL AND cdouble IS NOT NULL LIMIT 10 Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_part.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_part.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_part.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_part.q.out Thu Oct 30 12:31:47 2014 @@ -10,7 +10,6 @@ PREHOOK: query: insert overwrite table a PREHOOK: type: QUERY PREHOOK: Input: default@alltypesorc PREHOOK: Output: default@alltypesorc_part@ds=2011 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table alltypesorc_part partition (ds='2011') select * from alltypesorc limit 100 POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypesorc @@ -31,7 +30,6 @@ PREHOOK: query: insert overwrite table a PREHOOK: type: QUERY PREHOOK: Input: default@alltypesorc PREHOOK: Output: default@alltypesorc_part@ds=2012 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table alltypesorc_part partition (ds='2012') select * from alltypesorc limit 100 POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypesorc Modified: hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_part_project.q.out URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_part_project.q.out?rev=1635477&r1=1635476&r2=1635477&view=diff ============================================================================== --- hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_part_project.q.out (original) +++ hive/branches/spark/ql/src/test/results/clientpositive/spark/vectorization_part_project.q.out Thu Oct 30 12:31:47 2014 @@ -10,7 +10,6 @@ PREHOOK: query: insert overwrite table a PREHOOK: type: QUERY PREHOOK: Input: default@alltypesorc PREHOOK: Output: default@alltypesorc_part@ds=2011 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table alltypesorc_part partition (ds='2011') select * from alltypesorc limit 100 POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypesorc @@ -31,7 +30,6 @@ PREHOOK: query: insert overwrite table a PREHOOK: type: QUERY PREHOOK: Input: default@alltypesorc PREHOOK: Output: default@alltypesorc_part@ds=2012 -[Error 30017]: Skipping stats aggregation by error org.apache.hadoop.hive.ql.metadata.HiveException: [Error 30015]: Stats aggregator of type counter cannot be connected to POSTHOOK: query: insert overwrite table alltypesorc_part partition (ds='2012') select * from alltypesorc limit 100 POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypesorc @@ -67,28 +65,28 @@ STAGE PLANS: Map Operator Tree: TableScan alias: alltypesorc_part - Statistics: Num rows: 706 Data size: 5648 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: (cdouble + 2) (type: double) outputColumnNames: _col0 - Statistics: Num rows: 706 Data size: 5648 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: double) sort order: + - Statistics: Num rows: 706 Data size: 5648 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE Execution mode: vectorized Reducer 2 Reduce Operator Tree: Select Operator expressions: KEY.reducesinkkey0 (type: double) outputColumnNames: _col0 - Statistics: Num rows: 706 Data size: 5648 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 200 Data size: 41576 Basic stats: COMPLETE Column stats: NONE Limit Number of rows: 10 - Statistics: Num rows: 10 Data size: 80 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10 Data size: 2070 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 80 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 10 Data size: 2070 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat