hive git commit: HIVE-19130: NPE is thrown when REPL LOAD applied drop partition event (Sankar Hariappan, reviewed by Mahesh Kumar Behera, Thejas M Nair)
Repository: hive Updated Branches: refs/heads/branch-3 50f144d1f -> 150e35167 HIVE-19130: NPE is thrown when REPL LOAD applied drop partition event (Sankar Hariappan, reviewed by Mahesh Kumar Behera, Thejas M Nair) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/150e3516 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/150e3516 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/150e3516 Branch: refs/heads/branch-3 Commit: 150e35167ab49135fd808a987bfa7fb582a531ba Parents: 50f144d Author: Sankar Hariappan Authored: Fri Apr 13 11:28:49 2018 +0530 Committer: Sankar Hariappan Committed: Fri Apr 13 11:28:49 2018 +0530 -- ...TestReplicationScenariosAcrossInstances.java | 55 .../hadoop/hive/ql/parse/WarehouseInstance.java | 5 ++ .../org/apache/hadoop/hive/ql/exec/DDLTask.java | 8 ++- .../apache/hadoop/hive/ql/parse/EximUtil.java | 10 .../hive/ql/parse/ImportSemanticAnalyzer.java | 31 +-- .../parse/repl/load/message/TableHandler.java | 1 + 6 files changed, 106 insertions(+), 4 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/150e3516/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java -- diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java index 689ca76..70e1aa7 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java @@ -620,4 +620,59 @@ public class TestReplicationScenariosAcrossInstances { .run("show functions like '" + replicatedDbName + "*'") .verifyResult(null); } + + @Test + public void testIncrementalReplWithEventsBatchHavingDropCreateTable() throws Throwable { +// Bootstrap dump with empty db +WarehouseInstance.Tuple bootstrapTuple = primary.dump(primaryDbName, null); + +// Bootstrap load in replica +replica.load(replicatedDbName, bootstrapTuple.dumpLocation) +.status(replicatedDbName) +.verifyResult(bootstrapTuple.lastReplicationId); + +// First incremental dump +WarehouseInstance.Tuple firstIncremental = primary.run("use " + primaryDbName) +.run("create table table1 (i int)") +.run("create table table2 (id int) partitioned by (country string)") +.run("insert into table1 values (1)") +.run("insert into table2 partition(country='india') values(1)") +.dump(primaryDbName, bootstrapTuple.lastReplicationId); + +// Second incremental dump +WarehouseInstance.Tuple secondIncremental = primary.run("use " + primaryDbName) +.run("drop table table1") +.run("drop table table2") +.run("create table table2 (id int) partitioned by (country string)") +.run("alter table table2 add partition(country='india')") +.run("alter table table2 drop partition(country='india')") +.run("insert into table2 partition(country='us') values(2)") +.run("create table table1 (i int)") +.run("insert into table1 values (2)") +.dump(primaryDbName, firstIncremental.lastReplicationId); + +// First incremental load +replica.load(replicatedDbName, firstIncremental.dumpLocation) +.status(replicatedDbName) +.verifyResult(firstIncremental.lastReplicationId) +.run("use " + replicatedDbName) +.run("show tables") +.verifyResults(new String[] {"table1", "table2"}) +.run("select * from table1") +.verifyResults(new String[] {"1"}) +.run("select id from table2 order by id") +.verifyResults(new String[] {"1"}); + +// Second incremental load +replica.load(replicatedDbName, secondIncremental.dumpLocation) +.status(replicatedDbName) +.verifyResult(secondIncremental.lastReplicationId) +.run("use " + replicatedDbName) +.run("show tables") +.verifyResults(new String[] {"table1", "table2"}) +.run("select * from table1") +.verifyResults(new String[] {"2"}) +.run("select id from table2 order by id") +.verifyResults(new String[] {"2"}); + } } http://git-wip-us.apache.org/repos/asf/hive/blob/150e3516/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java -
hive git commit: HIVE-19170 : Fix TestMiniDruidKafkaCliDriver (Nishant Bangarwa via Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/master 6a929c58f -> 6f9090c1d HIVE-19170 : Fix TestMiniDruidKafkaCliDriver (Nishant Bangarwa via Ashutosh Chauhan) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/6f9090c1 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/6f9090c1 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/6f9090c1 Branch: refs/heads/master Commit: 6f9090c1de9bd6d2d98312a74b73802044cfa23e Parents: 6a929c5 Author: Nishant Bangarwa Authored: Thu Apr 12 22:41:54 2018 -0700 Committer: Ashutosh Chauhan Committed: Thu Apr 12 22:41:54 2018 -0700 -- itests/src/test/resources/testconfiguration.properties | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/6f9090c1/itests/src/test/resources/testconfiguration.properties -- diff --git a/itests/src/test/resources/testconfiguration.properties b/itests/src/test/resources/testconfiguration.properties index c5d4e9f..2db98c9 100644 --- a/itests/src/test/resources/testconfiguration.properties +++ b/itests/src/test/resources/testconfiguration.properties @@ -1698,5 +1698,6 @@ druid.query.files=druidmini_test1.q,\ druidmini_expressions.q,\ druidmini_extractTime.q,\ druidmini_test_alter.q,\ - druidkafkamini_basic.q,\ druidmini_floorTime.q + +druid.kafka.query.files=druidkafkamini_basic.q
hive git commit: HIVE-19172 : NPE due to null EnvironmentContext in DDLTask (Nishant Bangarwa via Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/master b7c64b15c -> 6a929c58f HIVE-19172 : NPE due to null EnvironmentContext in DDLTask (Nishant Bangarwa via Ashutosh Chauhan) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/6a929c58 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/6a929c58 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/6a929c58 Branch: refs/heads/master Commit: 6a929c58f943c41685f6e06a519381e696ad91ad Parents: b7c64b1 Author: Nishant Bangarwa Authored: Thu Apr 12 22:39:37 2018 -0700 Committer: Ashutosh Chauhan Committed: Thu Apr 12 22:39:37 2018 -0700 -- ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java | 3 +++ 1 file changed, 3 insertions(+) -- http://git-wip-us.apache.org/repos/asf/hive/blob/6a929c58/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java index 61a0432..b9b1830 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java @@ -3827,6 +3827,9 @@ public class DDLTask extends Task implements Serializable { try { EnvironmentContext environmentContext = alterTbl.getEnvironmentContext(); + if (environmentContext == null) { +environmentContext = new EnvironmentContext(); + } environmentContext.putToProperties(HiveMetaHook.ALTER_TABLE_OPERATION_TYPE, alterTbl.getOp().name()); if (allPartitions == null) { db.alterTable(alterTbl.getOldName(), tbl, alterTbl.getIsCascade(), environmentContext);
hive git commit: HIVE-17429: Hive JDBC doesn't return rows when querying Impala (Zach Amsden, reviewed by Aihua Xu)
Repository: hive Updated Branches: refs/heads/branch-2.3 f78128a71 -> 1ca13b3ce HIVE-17429: Hive JDBC doesn't return rows when querying Impala (Zach Amsden, reviewed by Aihua Xu) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/1ca13b3c Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/1ca13b3c Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/1ca13b3c Branch: refs/heads/branch-2.3 Commit: 1ca13b3cec97c4114090687079b6efbffeb498c3 Parents: f78128a Author: Aihua Xu Authored: Thu Sep 7 17:17:14 2017 -0700 Committer: Aihua Xu Committed: Thu Apr 12 15:33:16 2018 -0700 -- jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/1ca13b3c/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java -- diff --git a/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java b/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java index c385e2c..4c98517 100644 --- a/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java +++ b/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java @@ -254,7 +254,7 @@ public class HiveStatement implements java.sql.Statement { TGetOperationStatusResp status = waitForOperationToComplete(); // The query should be completed by now -if (!status.isHasResultSet()) { +if (!status.isHasResultSet() && !stmtHandle.isHasResultSet()) { return false; } resultSet = new HiveQueryResultSet.Builder(this).setClient(client).setSessionHandle(sessHandle)
hive git commit: HIVE-17429: Hive JDBC doesn't return rows when querying Impala (Zach Amsden, reviewed by Aihua Xu)
Repository: hive Updated Branches: refs/heads/branch-2.2 9e10b88c3 -> 1846c5a73 HIVE-17429: Hive JDBC doesn't return rows when querying Impala (Zach Amsden, reviewed by Aihua Xu) Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/1846c5a7 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/1846c5a7 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/1846c5a7 Branch: refs/heads/branch-2.2 Commit: 1846c5a737ee32b49dbeb6dd58ca7968f68fa31a Parents: 9e10b88 Author: Aihua Xu Authored: Thu Sep 7 17:17:14 2017 -0700 Committer: Aihua Xu Committed: Thu Apr 12 15:26:30 2018 -0700 -- jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/1846c5a7/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java -- diff --git a/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java b/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java index ed0ed08..d54c964 100644 --- a/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java +++ b/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java @@ -254,7 +254,7 @@ public class HiveStatement implements java.sql.Statement { TGetOperationStatusResp status = waitForOperationToComplete(); // The query should be completed by now -if (!status.isHasResultSet()) { +if (!status.isHasResultSet() && !stmtHandle.isHasResultSet()) { return false; } resultSet = new HiveQueryResultSet.Builder(this).setClient(client).setSessionHandle(sessHandle)
[10/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query4.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query4.q.out b/ql/src/test/results/clientpositive/perf/spark/query4.q.out index 5074d49..75c4fbb 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query4.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query4.q.out @@ -225,20 +225,20 @@ STAGE PLANS: Reducer 10 <- Map 13 (PARTITION-LEVEL SORT, 398), Map 9 (PARTITION-LEVEL SORT, 398) Reducer 11 <- Map 14 (PARTITION-LEVEL SORT, 975), Reducer 10 (PARTITION-LEVEL SORT, 975) Reducer 12 <- Reducer 11 (GROUP, 481) -Reducer 16 <- Map 15 (PARTITION-LEVEL SORT, 306), Map 19 (PARTITION-LEVEL SORT, 306) -Reducer 17 <- Map 20 (PARTITION-LEVEL SORT, 873), Reducer 16 (PARTITION-LEVEL SORT, 873) -Reducer 18 <- Reducer 17 (GROUP, 369) +Reducer 16 <- Map 15 (PARTITION-LEVEL SORT, 398), Map 19 (PARTITION-LEVEL SORT, 398) +Reducer 17 <- Map 20 (PARTITION-LEVEL SORT, 975), Reducer 16 (PARTITION-LEVEL SORT, 975) +Reducer 18 <- Reducer 17 (GROUP, 481) Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 306), Map 7 (PARTITION-LEVEL SORT, 306) -Reducer 22 <- Map 21 (PARTITION-LEVEL SORT, 154), Map 25 (PARTITION-LEVEL SORT, 154) -Reducer 23 <- Map 26 (PARTITION-LEVEL SORT, 706), Reducer 22 (PARTITION-LEVEL SORT, 706) -Reducer 24 <- Reducer 23 (GROUP, 186) +Reducer 22 <- Map 21 (PARTITION-LEVEL SORT, 306), Map 25 (PARTITION-LEVEL SORT, 306) +Reducer 23 <- Map 26 (PARTITION-LEVEL SORT, 873), Reducer 22 (PARTITION-LEVEL SORT, 873) +Reducer 24 <- Reducer 23 (GROUP, 369) Reducer 28 <- Map 27 (PARTITION-LEVEL SORT, 154), Map 31 (PARTITION-LEVEL SORT, 154) Reducer 29 <- Map 32 (PARTITION-LEVEL SORT, 706), Reducer 28 (PARTITION-LEVEL SORT, 706) Reducer 3 <- Map 8 (PARTITION-LEVEL SORT, 873), Reducer 2 (PARTITION-LEVEL SORT, 873) Reducer 30 <- Reducer 29 (GROUP, 186) -Reducer 34 <- Map 33 (PARTITION-LEVEL SORT, 398), Map 37 (PARTITION-LEVEL SORT, 398) -Reducer 35 <- Map 38 (PARTITION-LEVEL SORT, 975), Reducer 34 (PARTITION-LEVEL SORT, 975) -Reducer 36 <- Reducer 35 (GROUP, 481) +Reducer 34 <- Map 33 (PARTITION-LEVEL SORT, 154), Map 37 (PARTITION-LEVEL SORT, 154) +Reducer 35 <- Map 38 (PARTITION-LEVEL SORT, 706), Reducer 34 (PARTITION-LEVEL SORT, 706) +Reducer 36 <- Reducer 35 (GROUP, 186) Reducer 4 <- Reducer 3 (GROUP, 369) Reducer 5 <- Reducer 12 (PARTITION-LEVEL SORT, 690), Reducer 18 (PARTITION-LEVEL SORT, 690), Reducer 24 (PARTITION-LEVEL SORT, 690), Reducer 30 (PARTITION-LEVEL SORT, 690), Reducer 36 (PARTITION-LEVEL SORT, 690), Reducer 4 (PARTITION-LEVEL SORT, 690) Reducer 6 <- Reducer 5 (SORT, 1) @@ -300,20 +300,20 @@ STAGE PLANS: Map 15 Map Operator Tree: TableScan - alias: catalog_sales - Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE + alias: store_sales + Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: (cs_bill_customer_sk is not null and cs_sold_date_sk is not null) (type: boolean) -Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE +predicate: (ss_customer_sk is not null and ss_sold_date_sk is not null) (type: boolean) +Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: cs_sold_date_sk (type: int), cs_bill_customer_sk (type: int), cs_ext_discount_amt (type: decimal(7,2)), cs_ext_sales_price (type: decimal(7,2)), cs_ext_wholesale_cost (type: decimal(7,2)), cs_ext_list_price (type: decimal(7,2)) + expressions: ss_sold_date_sk (type: int), ss_customer_sk (type: int), ss_ext_discount_amt (type: decimal(7,2)), ss_ext_sales_price (type: decimal(7,2)), ss_ext_wholesale_cost (type: decimal(7,2)), ss_ext_list_price (type: decimal(7,2)) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (typ
[15/15] hive git commit: HIVE-19175 : TestMiniLlapLocalCliDriver.testCliDriver update_access_time_non_current_db failing (Vineet Garg via Peter Vary)
HIVE-19175 : TestMiniLlapLocalCliDriver.testCliDriver update_access_time_non_current_db failing (Vineet Garg via Peter Vary) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/50f144d1 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/50f144d1 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/50f144d1 Branch: refs/heads/branch-3 Commit: 50f144d1f62b6c74d7d2b5f6a20232ab77f32be0 Parents: 5a95a82 Author: Vineet Garg Authored: Thu Apr 12 09:34:31 2018 -0700 Committer: Vineet Garg Committed: Thu Apr 12 10:58:27 2018 -0700 -- .../test/resources/testconfiguration.properties | 2 +- .../update_access_time_non_current_db.q.out | 47 .../update_access_time_non_current_db.q.out | 47 3 files changed, 48 insertions(+), 48 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/50f144d1/itests/src/test/resources/testconfiguration.properties -- diff --git a/itests/src/test/resources/testconfiguration.properties b/itests/src/test/resources/testconfiguration.properties index 48d62a8..c5d4e9f 100644 --- a/itests/src/test/resources/testconfiguration.properties +++ b/itests/src/test/resources/testconfiguration.properties @@ -285,7 +285,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\ union8.q,\ union9.q,\ union_stats.q,\ - update_access_time_non_current_db.q, \ update_after_multiple_inserts.q,\ update_all_non_partitioned.q,\ update_all_partitioned.q,\ @@ -739,6 +738,7 @@ minillaplocal.query.files=\ union_fast_stats.q,\ union_remove_26.q,\ union_top_level.q,\ + update_access_time_non_current_db.q, \ vector_annotate_stats_select.q,\ vector_auto_smb_mapjoin_14.q,\ vector_char_varchar_1.q,\ http://git-wip-us.apache.org/repos/asf/hive/blob/50f144d1/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out b/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out new file mode 100644 index 000..7fda3ac --- /dev/null +++ b/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out @@ -0,0 +1,47 @@ +PREHOOK: query: create database temp1 +PREHOOK: type: CREATEDATABASE +PREHOOK: Output: database:temp1 +POSTHOOK: query: create database temp1 +POSTHOOK: type: CREATEDATABASE +POSTHOOK: Output: database:temp1 +PREHOOK: query: use temp1 +PREHOOK: type: SWITCHDATABASE +PREHOOK: Input: database:temp1 +POSTHOOK: query: use temp1 +POSTHOOK: type: SWITCHDATABASE +POSTHOOK: Input: database:temp1 +PREHOOK: query: create table test1(id int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:temp1 +PREHOOK: Output: temp1@test1 +POSTHOOK: query: create table test1(id int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:temp1 +POSTHOOK: Output: temp1@test1 +PREHOOK: query: create database temp2 +PREHOOK: type: CREATEDATABASE +PREHOOK: Output: database:temp2 +POSTHOOK: query: create database temp2 +POSTHOOK: type: CREATEDATABASE +POSTHOOK: Output: database:temp2 +PREHOOK: query: use temp2 +PREHOOK: type: SWITCHDATABASE +PREHOOK: Input: database:temp2 +POSTHOOK: query: use temp2 +POSTHOOK: type: SWITCHDATABASE +POSTHOOK: Input: database:temp2 +PREHOOK: query: create table test2(id int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:temp2 +PREHOOK: Output: temp2@test2 +POSTHOOK: query: create table test2(id int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:temp2 +POSTHOOK: Output: temp2@test2 +POSTHOOK: query: use temp1 +POSTHOOK: type: SWITCHDATABASE +POSTHOOK: Input: database:temp1 +POSTHOOK: query: desc temp2.test2 +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: temp2@test2 +id int http://git-wip-us.apache.org/repos/asf/hive/blob/50f144d1/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out -- diff --git a/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out b/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out deleted file mode 100644 index 7fda3ac..000 --- a/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out +++ /dev/null @@ -1,47 +0,0 @@ -PREHOOK: query: create database temp1 -PREHOOK: type: CREATEDATABASE -PREHOOK: Output: database:temp1 -POSTHOOK: query: create database temp1 -POSTHOOK: type: CREATEDATABASE -POSTHOOK: Output: database:temp1 -PREHOOK: query: use temp1 -PREHOOK: type: SWITCHDATABASE -PREHOOK: Input: database:temp1 -PO
[14/15] hive git commit: HIVE-19156 : TestMiniLlapLocalCliDriver.vectorized_dynamic_semijoin_reduction.q is broken (Jason Dere via Deepak Jaiswal)
HIVE-19156 : TestMiniLlapLocalCliDriver.vectorized_dynamic_semijoin_reduction.q is broken (Jason Dere via Deepak Jaiswal) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/5a95a826 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/5a95a826 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/5a95a826 Branch: refs/heads/branch-3 Commit: 5a95a826a1ce6db56a5a721ef99e3f76a9360abf Parents: 12a1784 Author: Jason Dere Authored: Thu Apr 12 08:47:25 2018 -0700 Committer: Vineet Garg Committed: Thu Apr 12 10:58:15 2018 -0700 -- .../ql/exec/tez/DynamicValueRegistryTez.java | 19 +-- 1 file changed, 17 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/5a95a826/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java index ec1e84b..2d99f50 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java @@ -61,6 +61,11 @@ public class DynamicValueRegistryTez implements DynamicValueRegistry { } } + static class NullValue { + } + + static final NullValue NULL_VALUE = new NullValue(); + protected Map values = new ConcurrentHashMap<>(); public DynamicValueRegistryTez() { @@ -71,11 +76,21 @@ public class DynamicValueRegistryTez implements DynamicValueRegistry { if (!values.containsKey(key)) { throw new NoDynamicValuesException("Value does not exist in registry: " + key); } -return values.get(key); +Object val = values.get(key); + +if (val == NULL_VALUE) { + return null; +} +return val; } protected void setValue(String key, Object value) { -values.put(key, value); +if (value == null) { + // ConcurrentHashMap does not allow null - use a substitute value. + values.put(key, NULL_VALUE); +} else { + values.put(key, value); +} } @Override
[08/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query50.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query50.q.out b/ql/src/test/results/clientpositive/perf/spark/query50.q.out index b978201..fde0c4e 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query50.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query50.q.out @@ -124,7 +124,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 6 +Map 10 Map Operator Tree: TableScan alias: store @@ -138,7 +138,7 @@ STAGE PLANS: Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col3 (type: int) + 0 _col10 (type: int) 1 _col0 (type: int) Local Work: Map Reduce Local Work @@ -146,63 +146,16 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 476), Reducer 8 (PARTITION-LEVEL SORT, 476) -Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 481), Reducer 2 (PARTITION-LEVEL SORT, 481) -Reducer 4 <- Reducer 3 (GROUP, 529) -Reducer 5 <- Reducer 4 (SORT, 1) -Reducer 8 <- Map 7 (PARTITION-LEVEL SORT, 36), Map 9 (PARTITION-LEVEL SORT, 36) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 36), Map 7 (PARTITION-LEVEL SORT, 36) +Reducer 3 <- Map 8 (PARTITION-LEVEL SORT, 436), Reducer 2 (PARTITION-LEVEL SORT, 436) +Reducer 4 <- Map 9 (PARTITION-LEVEL SORT, 438), Reducer 3 (PARTITION-LEVEL SORT, 438) +Reducer 5 <- Reducer 4 (GROUP, 529) +Reducer 6 <- Reducer 5 (SORT, 1) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: store_sales - Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: (ss_customer_sk is not null and ss_item_sk is not null and ss_sold_date_sk is not null and ss_store_sk is not null and ss_ticket_number is not null) (type: boolean) -Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: ss_sold_date_sk (type: int), ss_item_sk (type: int), ss_customer_sk (type: int), ss_store_sk (type: int), ss_ticket_number (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE - Map Join Operator -condition map: - Inner Join 0 to 1 -keys: - 0 _col3 (type: int) - 1 _col0 (type: int) -outputColumnNames: _col0, _col1, _col2, _col4, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15 -input vertices: - 1 Map 6 -Statistics: Num rows: 633595212 Data size: 55895953508 Basic stats: COMPLETE Column stats: NONE -Reduce Output Operator - key expressions: _col1 (type: int), _col2 (type: int), _col4 (type: int) - sort order: +++ - Map-reduce partition columns: _col1 (type: int), _col2 (type: int), _col4 (type: int) - Statistics: Num rows: 633595212 Data size: 55895953508 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col6 (type: string), _col7 (type: int), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: string) -Local Work: - Map Reduce Local Work -Map 10 -Map Operator Tree: -TableScan - alias: d1 - Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: d_date_sk is not null (type: boolean) -Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: d_date_sk (type: int) - outputColumnNames: _col0 - Statistics: Num rows:
[03/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query81.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query81.q.out b/ql/src/test/results/clientpositive/perf/spark/query81.q.out index 23998d6..5b2d5b3 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query81.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query81.q.out @@ -66,37 +66,72 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 25), Map 14 (PARTITION-LEVEL SORT, 25) -Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 344), Reducer 11 (PARTITION-LEVEL SORT, 344) -Reducer 13 <- Reducer 12 (GROUP PARTITION-LEVEL SORT, 349) -Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 697), Map 18 (PARTITION-LEVEL SORT, 697) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 25), Map 8 (PARTITION-LEVEL SORT, 25) -Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 344), Reducer 2 (PARTITION-LEVEL SORT, 344) -Reducer 4 <- Reducer 3 (GROUP, 349) -Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 262), Reducer 4 (PARTITION-LEVEL SORT, 262) -Reducer 6 <- Reducer 17 (PARTITION-LEVEL SORT, 656), Reducer 5 (PARTITION-LEVEL SORT, 656) -Reducer 7 <- Reducer 6 (SORT, 1) +Reducer 10 <- Reducer 16 (PARTITION-LEVEL SORT, 262), Reducer 9 (PARTITION-LEVEL SORT, 262) +Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 25), Map 17 (PARTITION-LEVEL SORT, 25) +Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 344), Reducer 14 (PARTITION-LEVEL SORT, 344) +Reducer 16 <- Reducer 15 (GROUP PARTITION-LEVEL SORT, 349) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 697), Map 5 (PARTITION-LEVEL SORT, 697) +Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 656), Reducer 2 (PARTITION-LEVEL SORT, 656) +Reducer 4 <- Reducer 3 (SORT, 1) +Reducer 7 <- Map 11 (PARTITION-LEVEL SORT, 25), Map 6 (PARTITION-LEVEL SORT, 25) +Reducer 8 <- Map 12 (PARTITION-LEVEL SORT, 344), Reducer 7 (PARTITION-LEVEL SORT, 344) +Reducer 9 <- Reducer 8 (GROUP, 349) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: catalog_returns - Statistics: Num rows: 28798881 Data size: 3057234680 Basic stats: COMPLETE Column stats: NONE + alias: customer + Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: (cr_returned_date_sk is not null and cr_returning_addr_sk is not null and cr_returning_customer_sk is not null) (type: boolean) -Statistics: Num rows: 28798881 Data size: 3057234680 Basic stats: COMPLETE Column stats: NONE +predicate: (c_current_addr_sk is not null and c_customer_sk is not null) (type: boolean) +Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: cr_returned_date_sk (type: int), cr_returning_customer_sk (type: int), cr_returning_addr_sk (type: int), cr_return_amt_inc_tax (type: decimal(7,2)) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 28798881 Data size: 3057234680 Basic stats: COMPLETE Column stats: NONE + expressions: c_customer_sk (type: int), c_customer_id (type: string), c_current_addr_sk (type: int), c_salutation (type: string), c_first_name (type: string), c_last_name (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator +key expressions: _col2 (type: int) +sort order: + +Map-reduce partition columns: _col2 (type: int) +Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE +value expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) +Map 11 +Map Operator Tree: +TableScan + alias: date_dim + Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE + Filter Operator +predicate: ((d_year = 1998) and d_date_sk is not null) (type: boolean) +Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE +Select
[11/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query29.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query29.q.out b/ql/src/test/results/clientpositive/perf/spark/query29.q.out index 456fce7..530a625 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query29.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query29.q.out @@ -124,7 +124,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 14 +Map 13 Map Operator Tree: TableScan alias: d2 @@ -146,51 +146,51 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 10 <- Map 12 (PARTITION-LEVEL SORT, 306), Map 9 (PARTITION-LEVEL SORT, 306) -Reducer 11 <- Map 13 (PARTITION-LEVEL SORT, 374), Reducer 10 (PARTITION-LEVEL SORT, 374) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 398), Map 7 (PARTITION-LEVEL SORT, 398) -Reducer 3 <- Map 8 (PARTITION-LEVEL SORT, 442), Reducer 2 (PARTITION-LEVEL SORT, 442) -Reducer 4 <- Reducer 11 (PARTITION-LEVEL SORT, 850), Reducer 3 (PARTITION-LEVEL SORT, 850) -Reducer 5 <- Reducer 4 (GROUP, 582) -Reducer 6 <- Reducer 5 (SORT, 1) +Reducer 10 <- Map 14 (PARTITION-LEVEL SORT, 486), Reducer 9 (PARTITION-LEVEL SORT, 486) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 306), Map 6 (PARTITION-LEVEL SORT, 306) +Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 917), Reducer 2 (PARTITION-LEVEL SORT, 917) +Reducer 4 <- Reducer 3 (GROUP, 640) +Reducer 5 <- Reducer 4 (SORT, 1) +Reducer 8 <- Map 11 (PARTITION-LEVEL SORT, 398), Map 7 (PARTITION-LEVEL SORT, 398) +Reducer 9 <- Map 12 (PARTITION-LEVEL SORT, 476), Reducer 8 (PARTITION-LEVEL SORT, 476) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: store_sales - Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE + alias: catalog_sales + Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: (ss_customer_sk is not null and ss_item_sk is not null and ss_sold_date_sk is not null and ss_store_sk is not null and ss_ticket_number is not null) (type: boolean) -Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE +predicate: (cs_bill_customer_sk is not null and cs_item_sk is not null and cs_sold_date_sk is not null) (type: boolean) +Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: ss_sold_date_sk (type: int), ss_item_sk (type: int), ss_customer_sk (type: int), ss_store_sk (type: int), ss_ticket_number (type: int), ss_quantity (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE + expressions: cs_sold_date_sk (type: int), cs_bill_customer_sk (type: int), cs_item_sk (type: int), cs_quantity (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int) sort order: + Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int) -Map 12 +Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE +value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int) +Map 11 Map Operator Tree: TableScan - alias: d3 + alias: d1 Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ((d_year) IN (1999, 2000, 2001) and d_date_sk is not null) (type: boolean) -Statistics: Num rows: 36525 Data size: 40871475 Basic stats: COMPLETE Column stats: NONE +predicate: ((d
[12/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query19.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query19.q.out b/ql/src/test/results/clientpositive/perf/spark/query19.q.out index 453bc56..281445c 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query19.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query19.q.out @@ -70,7 +70,7 @@ STAGE PLANS: Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col3 (type: int) + 0 _col7 (type: int) 1 _col0 (type: int) Local Work: Map Reduce Local Work @@ -78,35 +78,17 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 398), Map 8 (PARTITION-LEVEL SORT, 398) -Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 440), Reducer 2 (PARTITION-LEVEL SORT, 440) -Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 1009), Reducer 3 (PARTITION-LEVEL SORT, 1009) -Reducer 5 <- Map 11 (PARTITION-LEVEL SORT, 846), Reducer 4 (PARTITION-LEVEL SORT, 846) -Reducer 6 <- Reducer 5 (GROUP, 640) -Reducer 7 <- Reducer 6 (SORT, 1) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 855), Map 6 (PARTITION-LEVEL SORT, 855) +Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 1009), Reducer 9 (PARTITION-LEVEL SORT, 1009) +Reducer 4 <- Reducer 3 (GROUP, 582) +Reducer 5 <- Reducer 4 (SORT, 1) +Reducer 8 <- Map 10 (PARTITION-LEVEL SORT, 398), Map 7 (PARTITION-LEVEL SORT, 398) +Reducer 9 <- Map 11 (PARTITION-LEVEL SORT, 440), Reducer 8 (PARTITION-LEVEL SORT, 440) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: store_sales - Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: (ss_customer_sk is not null and ss_item_sk is not null and ss_sold_date_sk is not null and ss_store_sk is not null) (type: boolean) -Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: ss_sold_date_sk (type: int), ss_item_sk (type: int), ss_customer_sk (type: int), ss_store_sk (type: int), ss_ext_sales_price (type: decimal(7,2)) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col0 (type: int) -sort order: + -Map-reduce partition columns: _col0 (type: int) -Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)) -Map 10 -Map Operator Tree: -TableScan alias: customer Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE Filter Operator @@ -117,30 +99,12 @@ STAGE PLANS: outputColumnNames: _col0, _col1 Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col0 (type: int) +key expressions: _col1 (type: int) sort order: + -Map-reduce partition columns: _col0 (type: int) +Map-reduce partition columns: _col1 (type: int) Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE -value expressions: _col1 (type: int) -Map 11 -Map Operator Tree: -TableScan - alias: customer_address - Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: ca_address_sk is not null (type: boolean) -Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: ca_address_sk (type: int), ca_zip (type:
[13/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/12a1784d Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/12a1784d Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/12a1784d Branch: refs/heads/branch-3 Commit: 12a1784de44f185a963c34462a8f863db53920ce Parents: 91d0a0a Author: Zoltan Haindrich Authored: Wed Apr 11 20:05:01 2018 -0700 Committer: Vineet Garg Committed: Thu Apr 12 10:57:23 2018 -0700 -- data/conf/perf-reg/spark/hive-site.xml | 6 + data/conf/perf-reg/tez/hive-site.xml| 5 + .../hive/cli/control/CorePerfCliDriver.java | 6 +- .../hadoop/hive/ql/MetaStoreDumpUtility.java| 231 + .../org/apache/hadoop/hive/ql/QTestUtil.java| 221 + .../hive/ql/parse/BaseSemanticAnalyzer.java | 2 +- .../hadoop/hive/ql/parse/SemanticAnalyzer.java | 2 +- .../clientpositive/perf/spark/query11.q.out | 48 +- .../clientpositive/perf/spark/query15.q.out | 164 +++ .../clientpositive/perf/spark/query16.q.out | 70 +-- .../clientpositive/perf/spark/query18.q.out | 216 .../clientpositive/perf/spark/query19.q.out | 218 - .../clientpositive/perf/spark/query21.q.out | 114 +++-- .../clientpositive/perf/spark/query24.q.out | 282 +-- .../clientpositive/perf/spark/query25.q.out | 118 ++--- .../clientpositive/perf/spark/query29.q.out | 280 +-- .../clientpositive/perf/spark/query30.q.out | 262 +- .../clientpositive/perf/spark/query32.q.out | 132 ++--- .../clientpositive/perf/spark/query34.q.out | 34 +- .../clientpositive/perf/spark/query35.q.out | 74 ++- .../clientpositive/perf/spark/query37.q.out | 16 +- .../clientpositive/perf/spark/query4.q.out | 214 .../clientpositive/perf/spark/query40.q.out | 116 +++-- .../clientpositive/perf/spark/query44.q.out | 246 +- .../clientpositive/perf/spark/query45.q.out | 204 .../clientpositive/perf/spark/query46.q.out | 104 ++-- .../clientpositive/perf/spark/query47.q.out | 92 ++-- .../clientpositive/perf/spark/query48.q.out | 94 ++-- .../clientpositive/perf/spark/query5.q.out | 38 +- .../clientpositive/perf/spark/query50.q.out | 196 .../clientpositive/perf/spark/query53.q.out | 105 ++-- .../clientpositive/perf/spark/query54.q.out | 277 ++- .../clientpositive/perf/spark/query57.q.out | 92 ++-- .../clientpositive/perf/spark/query58.q.out | 490 +-- .../clientpositive/perf/spark/query6.q.out | 350 +++-- .../clientpositive/perf/spark/query61.q.out | 40 +- .../clientpositive/perf/spark/query63.q.out | 105 ++-- .../clientpositive/perf/spark/query65.q.out | 100 ++-- .../clientpositive/perf/spark/query66.q.out | 20 +- .../clientpositive/perf/spark/query67.q.out | 137 +++--- .../clientpositive/perf/spark/query68.q.out | 104 ++-- .../clientpositive/perf/spark/query72.q.out | 461 + .../clientpositive/perf/spark/query73.q.out | 34 +- .../clientpositive/perf/spark/query75.q.out | 248 +- .../clientpositive/perf/spark/query76.q.out | 130 +++-- .../clientpositive/perf/spark/query77.q.out | 66 +-- .../clientpositive/perf/spark/query78.q.out | 273 +-- .../clientpositive/perf/spark/query79.q.out | 10 +- .../clientpositive/perf/spark/query8.q.out | 174 +++ .../clientpositive/perf/spark/query80.q.out | 417 .../clientpositive/perf/spark/query81.q.out | 270 +- .../clientpositive/perf/spark/query82.q.out | 16 +- .../clientpositive/perf/spark/query83.q.out | 354 +++--- .../clientpositive/perf/spark/query85.q.out | 309 ++-- .../clientpositive/perf/spark/query88.q.out | 272 +- .../clientpositive/perf/spark/query89.q.out | 26 +- .../clientpositive/perf/spark/query90.q.out | 68 +-- .../clientpositive/perf/spark/query91.q.out | 14 +- .../clientpositive/perf/spark/query92.q.out | 106 ++-- .../clientpositive/perf/spark/query94.q.out | 70 +-- .../clientpositive/perf/spark/query95.q.out | 100 ++-- .../clientpositive/perf/spark/query97.q.out | 54 +- .../clientpositive/perf/spark/query99.q.out | 42 +- 63 files changed, 4603 insertions(+), 4536 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/data/conf/perf-reg/spark/hive-site.xml -- diff --git a/data/conf/perf-reg/spark/hive-site.xml b/data/conf/perf-reg/spark/hive-site.xml index
[06/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query65.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query65.q.out b/ql/src/test/results/clientpositive/perf/spark/query65.q.out index 575fc5c..3b3baef 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query65.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query65.q.out @@ -57,8 +57,7 @@ POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage Stage-3 depends on stages: Stage-2 - Stage-4 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-4 + Stage-1 depends on stages: Stage-3 Stage-0 depends on stages: Stage-1 STAGE PLANS: @@ -66,29 +65,6 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 11 -Map Operator Tree: -TableScan - alias: store - Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE - Filter Operator -predicate: s_store_sk is not null (type: boolean) -Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE -Select Operator - expressions: s_store_sk (type: int), s_store_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE - Spark HashTable Sink Operator -keys: - 0 _col0 (type: int) - 1 _col0 (type: int) -Local Work: - Map Reduce Local Work - - Stage: Stage-3 -Spark - A masked pattern was here - Vertices: Map 6 Map Operator Tree: TableScan @@ -108,7 +84,7 @@ STAGE PLANS: Local Work: Map Reduce Local Work - Stage: Stage-4 + Stage: Stage-3 Spark A masked pattern was here Vertices: @@ -135,8 +111,8 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (GROUP, 437) -Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 328), Reducer 8 (PARTITION-LEVEL SORT, 328) -Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 86), Reducer 3 (PARTITION-LEVEL SORT, 86) +Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 328), Reducer 2 (PARTITION-LEVEL SORT, 328), Reducer 8 (PARTITION-LEVEL SORT, 328) +Reducer 4 <- Map 11 (PARTITION-LEVEL SORT, 166), Reducer 3 (PARTITION-LEVEL SORT, 166) Reducer 5 <- Reducer 4 (SORT, 1) Reducer 8 <- Map 7 (GROUP PARTITION-LEVEL SORT, 437) A masked pattern was here @@ -180,6 +156,24 @@ STAGE PLANS: Map 10 Map Operator Tree: TableScan + alias: store + Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE + Filter Operator +predicate: s_store_sk is not null (type: boolean) +Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE +Select Operator + expressions: s_store_sk (type: int), s_store_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator +key expressions: _col0 (type: int) +sort order: + +Map-reduce partition columns: _col0 (type: int) +Statistics: Num rows: 1704 Data size: 3256276 Basic stats: COMPLETE Column stats: NONE +value expressions: _col1 (type: string) +Map 11 +Map Operator Tree: +TableScan alias: item Statistics: Num rows: 462000 Data size: 663560457 Basic stats: COMPLETE Column stats: NONE Filter Operator @@ -250,23 +244,23 @@ STAGE PLANS: Join Operator condition map: Inner Join 0 to 1 + Inner Join 0 to 2 keys: 0 _col0 (type: int) 1 _col0 (type: int) -outputColumnNames: _col0, _col1, _col2, _col4 -Statistics: Num rows: 348477374 Data size: 30742775095 Basic stats: COMPLETE Column stats: NONE + 2 _col0 (type: int) +outputColumnNames: _col1, _col2, _col4, _col6 +Statistics: Num rows: 696954748 Data size: 61485550191 Basic stats: COMPLETE Column stats: NONE Filter Operator predica
[05/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query75.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query75.q.out b/ql/src/test/results/clientpositive/perf/spark/query75.q.out index f4fca34..4b76570 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query75.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query75.q.out @@ -144,25 +144,25 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 400), Map 16 (PARTITION-LEVEL SORT, 400) -Reducer 14 <- Map 10 (PARTITION-LEVEL SORT, 438), Reducer 13 (PARTITION-LEVEL SORT, 438) +Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 398), Map 16 (PARTITION-LEVEL SORT, 398) +Reducer 14 <- Map 10 (PARTITION-LEVEL SORT, 440), Reducer 13 (PARTITION-LEVEL SORT, 440) Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 516), Reducer 14 (PARTITION-LEVEL SORT, 516) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 308), Map 16 (PARTITION-LEVEL SORT, 308) -Reducer 20 <- Map 16 (PARTITION-LEVEL SORT, 156), Map 19 (PARTITION-LEVEL SORT, 156) -Reducer 21 <- Map 10 (PARTITION-LEVEL SORT, 169), Reducer 20 (PARTITION-LEVEL SORT, 169) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 306), Map 16 (PARTITION-LEVEL SORT, 306) +Reducer 20 <- Map 16 (PARTITION-LEVEL SORT, 154), Map 19 (PARTITION-LEVEL SORT, 154) +Reducer 21 <- Map 10 (PARTITION-LEVEL SORT, 171), Reducer 20 (PARTITION-LEVEL SORT, 171) Reducer 22 <- Map 25 (PARTITION-LEVEL SORT, 196), Reducer 21 (PARTITION-LEVEL SORT, 196) -Reducer 27 <- Map 26 (PARTITION-LEVEL SORT, 308), Map 32 (PARTITION-LEVEL SORT, 308) -Reducer 28 <- Map 33 (PARTITION-LEVEL SORT, 336), Reducer 27 (PARTITION-LEVEL SORT, 336) +Reducer 27 <- Map 26 (PARTITION-LEVEL SORT, 306), Map 32 (PARTITION-LEVEL SORT, 306) +Reducer 28 <- Map 33 (PARTITION-LEVEL SORT, 338), Reducer 27 (PARTITION-LEVEL SORT, 338) Reducer 29 <- Map 34 (PARTITION-LEVEL SORT, 393), Reducer 28 (PARTITION-LEVEL SORT, 393) -Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 336), Reducer 2 (PARTITION-LEVEL SORT, 336) +Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 338), Reducer 2 (PARTITION-LEVEL SORT, 338) Reducer 30 <- Reducer 29 (GROUP, 934), Reducer 38 (GROUP, 934) Reducer 31 <- Reducer 30 (GROUP PARTITION-LEVEL SORT, 671), Reducer 45 (GROUP PARTITION-LEVEL SORT, 671) -Reducer 36 <- Map 32 (PARTITION-LEVEL SORT, 400), Map 35 (PARTITION-LEVEL SORT, 400) -Reducer 37 <- Map 33 (PARTITION-LEVEL SORT, 438), Reducer 36 (PARTITION-LEVEL SORT, 438) +Reducer 36 <- Map 32 (PARTITION-LEVEL SORT, 398), Map 35 (PARTITION-LEVEL SORT, 398) +Reducer 37 <- Map 33 (PARTITION-LEVEL SORT, 440), Reducer 36 (PARTITION-LEVEL SORT, 440) Reducer 38 <- Map 41 (PARTITION-LEVEL SORT, 516), Reducer 37 (PARTITION-LEVEL SORT, 516) Reducer 4 <- Map 11 (PARTITION-LEVEL SORT, 393), Reducer 3 (PARTITION-LEVEL SORT, 393) -Reducer 43 <- Map 32 (PARTITION-LEVEL SORT, 156), Map 42 (PARTITION-LEVEL SORT, 156) -Reducer 44 <- Map 33 (PARTITION-LEVEL SORT, 169), Reducer 43 (PARTITION-LEVEL SORT, 169) +Reducer 43 <- Map 32 (PARTITION-LEVEL SORT, 154), Map 42 (PARTITION-LEVEL SORT, 154) +Reducer 44 <- Map 33 (PARTITION-LEVEL SORT, 171), Reducer 43 (PARTITION-LEVEL SORT, 171) Reducer 45 <- Map 48 (PARTITION-LEVEL SORT, 196), Reducer 44 (PARTITION-LEVEL SORT, 196) Reducer 5 <- Reducer 15 (GROUP, 934), Reducer 4 (GROUP, 934) Reducer 6 <- Reducer 22 (GROUP PARTITION-LEVEL SORT, 671), Reducer 5 (GROUP PARTITION-LEVEL SORT, 671) @@ -183,28 +183,29 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col1 (type: int) +key expressions: _col0 (type: int) sort order: + -Map-reduce partition columns: _col1 (type: int) +Map-reduce partition columns: _col0 (type: int) Statistics: Num rows: 287989836 Data size: 38999608952 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)) +value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: decimal(7,2)) Map 10 Map Operator Tree: TableScan - alias: date_dim - Statistics: Num rows: 73049 Data size: 81741831 Basic stats:
[07/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query58.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query58.q.out b/ql/src/test/results/clientpositive/perf/spark/query58.q.out index eb5dffe..f06cbef 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query58.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query58.q.out @@ -1,6 +1,6 @@ -Warning: Map Join MAPJOIN[180][bigTable=?] in task 'Stage-1:MAPRED' is a cross product -Warning: Map Join MAPJOIN[181][bigTable=?] in task 'Stage-1:MAPRED' is a cross product -Warning: Map Join MAPJOIN[182][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[183][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[184][bigTable=?] in task 'Stage-1:MAPRED' is a cross product +Warning: Map Join MAPJOIN[185][bigTable=?] in task 'Stage-1:MAPRED' is a cross product PREHOOK: query: explain with ss_items as (select i_item_id item_id @@ -140,10 +140,10 @@ STAGE PLANS: Stage: Stage-2 Spark Edges: -Reducer 12 <- Map 11 (GROUP, 1) +Reducer 11 <- Map 10 (GROUP, 1) A masked pattern was here Vertices: -Map 11 +Map 10 Map Operator Tree: TableScan alias: date_dim @@ -162,7 +162,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 12 +Reducer 11 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -184,10 +184,10 @@ STAGE PLANS: Stage: Stage-3 Spark Edges: -Reducer 25 <- Map 24 (GROUP, 1) +Reducer 24 <- Map 23 (GROUP, 1) A masked pattern was here Vertices: -Map 24 +Map 23 Map Operator Tree: TableScan alias: date_dim @@ -206,7 +206,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 25 +Reducer 24 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -228,10 +228,10 @@ STAGE PLANS: Stage: Stage-4 Spark Edges: -Reducer 38 <- Map 37 (GROUP, 1) +Reducer 37 <- Map 36 (GROUP, 1) A masked pattern was here Vertices: -Map 37 +Map 36 Map Operator Tree: TableScan alias: date_dim @@ -250,7 +250,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint) -Reducer 38 +Reducer 37 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -272,47 +272,76 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 10 <- Reducer 9 (GROUP, 2) -Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 2), Reducer 23 (PARTITION-LEVEL SORT, 2) -Reducer 18 <- Map 27 (PARTITION-LEVEL SORT, 398), Reducer 17 (PARTITION-LEVEL SORT, 398) -Reducer 19 <- Map 28 (PARTITION-LEVEL SORT, 442), Reducer 18 (PARTITION-LEVEL SORT, 442) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2), Reducer 10 (PARTITION-LEVEL SORT, 2) -Reducer 20 <- Reducer 19 (GROUP, 481) -Reducer 22 <- Map 21 (PARTITION-LEVEL SORT, 2), Map 26 (PARTITION-LEVEL SORT, 2) -Reducer 23 <- Reducer 22 (GROUP, 2) -Reducer 3 <- Map 14 (PARTITION-LEVEL SORT, 306), Reducer 2 (PARTITION-LEVEL SORT, 306) -Reducer 30 <- Map 29 (PARTITION-LEVEL SORT, 2), Reducer 36 (PARTITION-LEVEL SORT, 2) -Reducer 31 <- Map 40 (PARTITION-LEVEL SORT, 154), Reducer 30 (PARTITION-LEVEL SORT, 154) -Reducer 32 <- Map 41 (PARTITION-LEVEL SORT, 174), Reducer 31 (PARTITION-LEVEL SORT, 174) -Reducer 33 <- Reducer 32 (GROUP, 186) -Reducer 35 <- Map 34 (PARTITION-LEVEL SORT, 2), Map 39 (PARTITION-LEVEL SORT, 2) -Reducer 36 <- Reducer 35 (GROUP, 2) -Reducer 4 <- Map 15 (PARTITION-LEVEL SORT, 341), Reducer 3 (PARTITION-LEVEL SORT, 341) -Reducer 5 <- Reducer 4 (GROUP, 369) -Reducer 6 <- Reducer 20 (PARTITION-LEVEL SORT, 518), Reducer 33 (PARTITION-LEVEL SORT, 518), Reducer 5 (PARTITION-LEVEL SORT, 518) -Reducer 7 <- Reducer 6 (SORT, 1) -Reducer 9 <- Map 13 (PARTITION-LEVEL SORT, 2), Map 8 (PARTITION-LEVEL SORT, 2) +Reducer 13 <- Map 12 (P
[02/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query85.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query85.q.out b/ql/src/test/results/clientpositive/perf/spark/query85.q.out index c82dcf7..d60751c 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query85.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query85.q.out @@ -175,7 +175,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 15 +Map 13 Map Operator Tree: TableScan alias: reason @@ -212,21 +212,21 @@ STAGE PLANS: Statistics: Num rows: 4602 Data size: 2696178 Basic stats: COMPLETE Column stats: NONE Spark HashTable Sink Operator keys: - 0 _col0 (type: int) - 1 _col2 (type: int) + 0 _col10 (type: int) + 1 _col0 (type: int) Local Work: Map Reduce Local Work Stage: Stage-1 Spark Edges: -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 20), Map 14 (PARTITION-LEVEL SORT, 20) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 90), Map 8 (PARTITION-LEVEL SORT, 90) -Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 17), Reducer 2 (PARTITION-LEVEL SORT, 17) -Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 19), Reducer 3 (PARTITION-LEVEL SORT, 19) -Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 35), Reducer 4 (PARTITION-LEVEL SORT, 35) -Reducer 6 <- Reducer 5 (GROUP, 2) -Reducer 7 <- Reducer 6 (SORT, 1) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 28), Map 9 (PARTITION-LEVEL SORT, 28) +Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 98), Reducer 2 (PARTITION-LEVEL SORT, 98) +Reducer 4 <- Map 12 (PARTITION-LEVEL SORT, 5), Reducer 3 (PARTITION-LEVEL SORT, 5) +Reducer 5 <- Map 14 (PARTITION-LEVEL SORT, 11), Reducer 4 (PARTITION-LEVEL SORT, 11) +Reducer 6 <- Map 15 (PARTITION-LEVEL SORT, 7), Reducer 5 (PARTITION-LEVEL SORT, 7) +Reducer 7 <- Reducer 6 (GROUP, 7) +Reducer 8 <- Reducer 7 (SORT, 1) A masked pattern was here Vertices: Map 1 @@ -242,61 +242,32 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 Statistics: Num rows: 14398467 Data size: 1325194184 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col2 (type: int) -sort order: + -Map-reduce partition columns: _col2 (type: int) +key expressions: _col0 (type: int), _col5 (type: int) +sort order: ++ +Map-reduce partition columns: _col0 (type: int), _col5 (type: int) Statistics: Num rows: 14398467 Data size: 1325194184 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: decimal(7,2)), _col7 (type: decimal(7,2)) +value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col6 (type: decimal(7,2)), _col7 (type: decimal(7,2)) Map 10 Map Operator Tree: TableScan - alias: cd2 - Statistics: Num rows: 1861800 Data size: 717186159 Basic stats: COMPLETE Column stats: NONE + alias: customer_address + Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: (((cd_education_status = '4 yr Degree') or (cd_education_status = 'Primary') or (cd_education_status = 'Advanced Degree')) and ((cd_marital_status = 'M') or (cd_marital_status = 'D') or (cd_marital_status = 'U')) and cd_demo_sk is not null and cd_education_status is not null and cd_marital_status is not null) (type: boolean) -Statistics: Num rows: 1861800 Data size: 717186159 Basic stats: COMPLETE Column stats: NONE +predicate: ((ca_country = 'United States') and (ca_state) IN ('KY', 'GA', 'NM', 'MT', 'OR', 'IN', 'WI', 'MO', 'WV') and ca_address_sk is not null) (type: boolean) +Statistics: Num rows: 1000 Data size: 10148798821 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: cd_demo_sk (type: int), cd_marital_status (type: string), cd_education_status (type: string) - outputColumnNames: _col0, _col
[01/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
Repository: hive Updated Branches: refs/heads/branch-3 91d0a0a2f -> 50f144d1f http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query92.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query92.q.out b/ql/src/test/results/clientpositive/perf/spark/query92.q.out index 70c4c5a..e7b8632 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query92.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query92.q.out @@ -67,7 +67,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 7 +Map 5 Map Operator Tree: TableScan alias: date_dim @@ -90,7 +90,7 @@ STAGE PLANS: Spark A masked pattern was here Vertices: -Map 10 +Map 9 Map Operator Tree: TableScan alias: date_dim @@ -112,11 +112,11 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 156), Map 6 (PARTITION-LEVEL SORT, 156) -Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 270), Reducer 9 (PARTITION-LEVEL SORT, 270) -Reducer 4 <- Reducer 3 (GROUP, 1) -Reducer 5 <- Reducer 4 (SORT, 1) -Reducer 9 <- Map 8 (GROUP, 169) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 261), Reducer 8 (PARTITION-LEVEL SORT, 261) +Reducer 3 <- Reducer 2 (GROUP, 1) +Reducer 4 <- Reducer 3 (SORT, 1) +Reducer 7 <- Map 6 (GROUP, 169) +Reducer 8 <- Map 10 (PARTITION-LEVEL SORT, 87), Reducer 7 (PARTITION-LEVEL SORT, 87) A masked pattern was here Vertices: Map 1 @@ -131,13 +131,25 @@ STAGE PLANS: expressions: ws_sold_date_sk (type: int), ws_item_sk (type: int), ws_ext_discount_amt (type: decimal(7,2)) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 144002668 Data size: 19580198212 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator -key expressions: _col1 (type: int) -sort order: + -Map-reduce partition columns: _col1 (type: int) -Statistics: Num rows: 144002668 Data size: 19580198212 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: int), _col2 (type: decimal(7,2)) -Map 6 + Map Join Operator +condition map: + Inner Join 0 to 1 +keys: + 0 _col0 (type: int) + 1 _col0 (type: int) +outputColumnNames: _col1, _col2 +input vertices: + 1 Map 5 +Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE +Reduce Output Operator + key expressions: _col1 (type: int) + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: decimal(7,2)) +Local Work: + Map Reduce Local Work +Map 10 Map Operator Tree: TableScan alias: item @@ -154,7 +166,7 @@ STAGE PLANS: sort order: + Map-reduce partition columns: _col0 (type: int) Statistics: Num rows: 231000 Data size: 331780228 Basic stats: COMPLETE Column stats: NONE -Map 8 +Map 6 Map Operator Tree: TableScan alias: web_sales @@ -174,7 +186,7 @@ STAGE PLANS: 1 _col0 (type: int) outputColumnNames: _col1, _col2 input vertices: - 1 Map 10 + 1 Map 9 Statistics: Num rows: 158402938 Data size: 21538218500 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(_col2), count(_col2) @@ -191,50 +203,22 @@ STAGE PLANS: Local Work: Map Reduce Local Work Reducer 2 -Local Work: - Map Reduce Local Work Reduce Operator Tree: Join Operator condition map: Inner Join 0 to 1 keys: 0 _col1 (type: int) - 1 _col0 (type: int) -outp
[04/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query78.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query78.q.out b/ql/src/test/results/clientpositive/perf/spark/query78.q.out index c2e8577..1467c5f 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query78.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query78.q.out @@ -120,41 +120,57 @@ STAGE PLANS: Stage: Stage-1 Spark Edges: -Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 164), Map 14 (PARTITION-LEVEL SORT, 164) -Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 85), Reducer 11 (PARTITION-LEVEL SORT, 85) -Reducer 13 <- Reducer 12 (GROUP, 93) -Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 85), Reducer 20 (PARTITION-LEVEL SORT, 85) +Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 85), Reducer 14 (PARTITION-LEVEL SORT, 85) +Reducer 12 <- Reducer 11 (GROUP, 93) +Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 164), Map 15 (PARTITION-LEVEL SORT, 164) +Reducer 17 <- Map 10 (PARTITION-LEVEL SORT, 85), Reducer 20 (PARTITION-LEVEL SORT, 85) Reducer 18 <- Reducer 17 (GROUP, 93) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 432), Map 8 (PARTITION-LEVEL SORT, 432) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 219), Reducer 8 (PARTITION-LEVEL SORT, 219) Reducer 20 <- Map 19 (PARTITION-LEVEL SORT, 177), Map 21 (PARTITION-LEVEL SORT, 177) -Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 219), Reducer 2 (PARTITION-LEVEL SORT, 219) -Reducer 4 <- Reducer 3 (GROUP, 241) -Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 167), Reducer 4 (PARTITION-LEVEL SORT, 167) -Reducer 6 <- Reducer 18 (PARTITION-LEVEL SORT, 91), Reducer 5 (PARTITION-LEVEL SORT, 91) -Reducer 7 <- Reducer 6 (SORT, 1) +Reducer 3 <- Reducer 2 (GROUP, 241) +Reducer 4 <- Reducer 12 (PARTITION-LEVEL SORT, 167), Reducer 3 (PARTITION-LEVEL SORT, 167) +Reducer 5 <- Reducer 18 (PARTITION-LEVEL SORT, 91), Reducer 4 (PARTITION-LEVEL SORT, 91) +Reducer 6 <- Reducer 5 (SORT, 1) +Reducer 8 <- Map 7 (PARTITION-LEVEL SORT, 432), Map 9 (PARTITION-LEVEL SORT, 432) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: store_sales - Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE + alias: date_dim + Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ss_sold_date_sk is not null (type: boolean) -Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE +predicate: ((d_year = 2000) and d_date_sk is not null) (type: boolean) +Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: ss_sold_date_sk (type: int), ss_item_sk (type: int), ss_customer_sk (type: int), ss_ticket_number (type: int), ss_quantity (type: int), ss_wholesale_cost (type: decimal(7,2)), ss_sales_price (type: decimal(7,2)) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE + expressions: d_date_sk (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator -key expressions: _col1 (type: int), _col3 (type: int) -sort order: ++ -Map-reduce partition columns: _col1 (type: int), _col3 (type: int) -Statistics: Num rows: 575995635 Data size: 50814502088 Basic stats: COMPLETE Column stats: NONE -value expressions: _col0 (type: int), _col2 (type: int), _col4 (type: int), _col5 (type: decimal(7,2)), _col6 (type: decimal(7,2)) +key expressions: _col0 (type: int) +sort order: + +Map-reduce partition columns: _col0 (type: int) +Statistics: Num rows: 36524 Data size: 40870356 Basic stats: COMPLETE Column stats: NONE Map 10 Map Operator Tree: TableScan + alias: date_dim + Statistics: Num rows: 73049 Data size: 81741831 Basic stats: COMPLETE Column stats: NONE + Filter Operator +
[09/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query45.q.out -- diff --git a/ql/src/test/results/clientpositive/perf/spark/query45.q.out b/ql/src/test/results/clientpositive/perf/spark/query45.q.out index 7e1cc88..b674400 100644 --- a/ql/src/test/results/clientpositive/perf/spark/query45.q.out +++ b/ql/src/test/results/clientpositive/perf/spark/query45.q.out @@ -46,10 +46,10 @@ STAGE PLANS: Stage: Stage-2 Spark Edges: -Reducer 6 <- Map 5 (GROUP, 1) +Reducer 16 <- Map 15 (GROUP, 1) A masked pattern was here Vertices: -Map 5 +Map 15 Map Operator Tree: TableScan alias: item @@ -70,7 +70,7 @@ STAGE PLANS: sort order: Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE value expressions: _col0 (type: bigint), _col1 (type: bigint) -Reducer 6 +Reducer 16 Local Work: Map Reduce Local Work Reduce Operator Tree: @@ -88,45 +88,33 @@ STAGE PLANS: Spark Edges: Reducer 11 <- Map 10 (GROUP, 3) -Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 154), Map 15 (PARTITION-LEVEL SORT, 154) -Reducer 14 <- Map 16 (PARTITION-LEVEL SORT, 706), Reducer 13 (PARTITION-LEVEL SORT, 706) -Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 527), Reducer 9 (PARTITION-LEVEL SORT, 527) -Reducer 3 <- Reducer 2 (GROUP, 224) -Reducer 4 <- Reducer 3 (SORT, 1) +Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 154), Map 14 (PARTITION-LEVEL SORT, 154) +Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 855), Map 6 (PARTITION-LEVEL SORT, 855) +Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 777), Reducer 9 (PARTITION-LEVEL SORT, 777) +Reducer 4 <- Reducer 3 (GROUP, 230) +Reducer 5 <- Reducer 4 (SORT, 1) Reducer 8 <- Map 7 (PARTITION-LEVEL SORT, 7), Reducer 11 (PARTITION-LEVEL SORT, 7) -Reducer 9 <- Reducer 14 (PARTITION-LEVEL SORT, 191), Reducer 8 (PARTITION-LEVEL SORT, 191) +Reducer 9 <- Reducer 13 (PARTITION-LEVEL SORT, 174), Reducer 8 (PARTITION-LEVEL SORT, 174) A masked pattern was here Vertices: Map 1 Map Operator Tree: TableScan - alias: customer_address - Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE + alias: customer + Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE Filter Operator -predicate: ca_address_sk is not null (type: boolean) -Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE +predicate: (c_current_addr_sk is not null and c_customer_sk is not null) (type: boolean) +Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: ca_address_sk (type: int), ca_county (type: string), ca_zip (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4000 Data size: 40595195284 Basic stats: COMPLETE Column stats: NONE - Map Join Operator -condition map: - Inner Join 0 to 1 -keys: - 0 - 1 -outputColumnNames: _col0, _col1, _col2, _col3, _col4 -input vertices: - 1 Reducer 6 -Statistics: Num rows: 4000 Data size: 41275195284 Basic stats: COMPLETE Column stats: NONE -Reduce Output Operator - key expressions: _col0 (type: int) - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 4000 Data size: 41275195284 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: bigint), _col4 (type: bigint) -Local Work: - Map Reduce Local Work + expressions: c_customer_sk (type: int), c_current_addr_sk (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 8000 Data size: 68801615852 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator +key expressions: _col1 (
hive git commit: HIVE-19175 : TestMiniLlapLocalCliDriver.testCliDriver update_access_time_non_current_db failing (Vineet Garg via Peter Vary)
Repository: hive Updated Branches: refs/heads/master a2dd09f86 -> b7c64b15c HIVE-19175 : TestMiniLlapLocalCliDriver.testCliDriver update_access_time_non_current_db failing (Vineet Garg via Peter Vary) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/b7c64b15 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/b7c64b15 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/b7c64b15 Branch: refs/heads/master Commit: b7c64b15c28b78c889e44f58f5e61374696ab9a5 Parents: a2dd09f Author: Vineet Garg Authored: Thu Apr 12 09:34:31 2018 -0700 Committer: Ashutosh Chauhan Committed: Thu Apr 12 09:34:31 2018 -0700 -- .../test/resources/testconfiguration.properties | 2 +- .../update_access_time_non_current_db.q.out | 47 .../update_access_time_non_current_db.q.out | 47 3 files changed, 48 insertions(+), 48 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/b7c64b15/itests/src/test/resources/testconfiguration.properties -- diff --git a/itests/src/test/resources/testconfiguration.properties b/itests/src/test/resources/testconfiguration.properties index 48d62a8..c5d4e9f 100644 --- a/itests/src/test/resources/testconfiguration.properties +++ b/itests/src/test/resources/testconfiguration.properties @@ -285,7 +285,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\ union8.q,\ union9.q,\ union_stats.q,\ - update_access_time_non_current_db.q, \ update_after_multiple_inserts.q,\ update_all_non_partitioned.q,\ update_all_partitioned.q,\ @@ -739,6 +738,7 @@ minillaplocal.query.files=\ union_fast_stats.q,\ union_remove_26.q,\ union_top_level.q,\ + update_access_time_non_current_db.q, \ vector_annotate_stats_select.q,\ vector_auto_smb_mapjoin_14.q,\ vector_char_varchar_1.q,\ http://git-wip-us.apache.org/repos/asf/hive/blob/b7c64b15/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out -- diff --git a/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out b/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out new file mode 100644 index 000..7fda3ac --- /dev/null +++ b/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out @@ -0,0 +1,47 @@ +PREHOOK: query: create database temp1 +PREHOOK: type: CREATEDATABASE +PREHOOK: Output: database:temp1 +POSTHOOK: query: create database temp1 +POSTHOOK: type: CREATEDATABASE +POSTHOOK: Output: database:temp1 +PREHOOK: query: use temp1 +PREHOOK: type: SWITCHDATABASE +PREHOOK: Input: database:temp1 +POSTHOOK: query: use temp1 +POSTHOOK: type: SWITCHDATABASE +POSTHOOK: Input: database:temp1 +PREHOOK: query: create table test1(id int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:temp1 +PREHOOK: Output: temp1@test1 +POSTHOOK: query: create table test1(id int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:temp1 +POSTHOOK: Output: temp1@test1 +PREHOOK: query: create database temp2 +PREHOOK: type: CREATEDATABASE +PREHOOK: Output: database:temp2 +POSTHOOK: query: create database temp2 +POSTHOOK: type: CREATEDATABASE +POSTHOOK: Output: database:temp2 +PREHOOK: query: use temp2 +PREHOOK: type: SWITCHDATABASE +PREHOOK: Input: database:temp2 +POSTHOOK: query: use temp2 +POSTHOOK: type: SWITCHDATABASE +POSTHOOK: Input: database:temp2 +PREHOOK: query: create table test2(id int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:temp2 +PREHOOK: Output: temp2@test2 +POSTHOOK: query: create table test2(id int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:temp2 +POSTHOOK: Output: temp2@test2 +POSTHOOK: query: use temp1 +POSTHOOK: type: SWITCHDATABASE +POSTHOOK: Input: database:temp1 +POSTHOOK: query: desc temp2.test2 +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: temp2@test2 +id int http://git-wip-us.apache.org/repos/asf/hive/blob/b7c64b15/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out -- diff --git a/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out b/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out deleted file mode 100644 index 7fda3ac..000 --- a/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out +++ /dev/null @@ -1,47 +0,0 @@ -PREHOOK: query: create database temp1 -PREHOOK: type: CREATEDATABASE -PREHOOK: Output: database:temp1 -POSTHOOK: query: create database temp1 -POSTHOOK: type: CREATEDATABASE -POSTHOOK: Output: database:temp1 -PREHOOK:
hive git commit: HIVE-19156 : TestMiniLlapLocalCliDriver.vectorized_dynamic_semijoin_reduction.q is broken (Jason Dere via Deepak Jaiswal)
Repository: hive Updated Branches: refs/heads/master 2e027cff7 -> a2dd09f86 HIVE-19156 : TestMiniLlapLocalCliDriver.vectorized_dynamic_semijoin_reduction.q is broken (Jason Dere via Deepak Jaiswal) Signed-off-by: Ashutosh Chauhan Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/a2dd09f8 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/a2dd09f8 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/a2dd09f8 Branch: refs/heads/master Commit: a2dd09f86386ca3f51eca90948d6ec1963ece982 Parents: 2e027cf Author: Jason Dere Authored: Thu Apr 12 08:47:25 2018 -0700 Committer: Ashutosh Chauhan Committed: Thu Apr 12 08:47:25 2018 -0700 -- .../ql/exec/tez/DynamicValueRegistryTez.java | 19 +-- 1 file changed, 17 insertions(+), 2 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hive/blob/a2dd09f8/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java -- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java index ec1e84b..2d99f50 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java @@ -61,6 +61,11 @@ public class DynamicValueRegistryTez implements DynamicValueRegistry { } } + static class NullValue { + } + + static final NullValue NULL_VALUE = new NullValue(); + protected Map values = new ConcurrentHashMap<>(); public DynamicValueRegistryTez() { @@ -71,11 +76,21 @@ public class DynamicValueRegistryTez implements DynamicValueRegistry { if (!values.containsKey(key)) { throw new NoDynamicValuesException("Value does not exist in registry: " + key); } -return values.get(key); +Object val = values.get(key); + +if (val == NULL_VALUE) { + return null; +} +return val; } protected void setValue(String key, Object value) { -values.put(key, value); +if (value == null) { + // ConcurrentHashMap does not allow null - use a substitute value. + values.put(key, NULL_VALUE); +} else { + values.put(key, value); +} } @Override