hive git commit: HIVE-19130: NPE is thrown when REPL LOAD applied drop partition event (Sankar Hariappan, reviewed by Mahesh Kumar Behera, Thejas M Nair)

2018-04-12 Thread sankarh
Repository: hive
Updated Branches:
  refs/heads/branch-3 50f144d1f -> 150e35167


HIVE-19130: NPE is thrown when REPL LOAD applied drop partition event (Sankar 
Hariappan, reviewed by Mahesh Kumar Behera, Thejas M Nair)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/150e3516
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/150e3516
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/150e3516

Branch: refs/heads/branch-3
Commit: 150e35167ab49135fd808a987bfa7fb582a531ba
Parents: 50f144d
Author: Sankar Hariappan 
Authored: Fri Apr 13 11:28:49 2018 +0530
Committer: Sankar Hariappan 
Committed: Fri Apr 13 11:28:49 2018 +0530

--
 ...TestReplicationScenariosAcrossInstances.java | 55 
 .../hadoop/hive/ql/parse/WarehouseInstance.java |  5 ++
 .../org/apache/hadoop/hive/ql/exec/DDLTask.java |  8 ++-
 .../apache/hadoop/hive/ql/parse/EximUtil.java   | 10 
 .../hive/ql/parse/ImportSemanticAnalyzer.java   | 31 +--
 .../parse/repl/load/message/TableHandler.java   |  1 +
 6 files changed, 106 insertions(+), 4 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/150e3516/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java
--
diff --git 
a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java
 
b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java
index 689ca76..70e1aa7 100644
--- 
a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java
+++ 
b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/TestReplicationScenariosAcrossInstances.java
@@ -620,4 +620,59 @@ public class TestReplicationScenariosAcrossInstances {
 .run("show functions like '" + replicatedDbName + "*'")
 .verifyResult(null);
   }
+
+  @Test
+  public void testIncrementalReplWithEventsBatchHavingDropCreateTable() throws 
Throwable {
+// Bootstrap dump with empty db
+WarehouseInstance.Tuple bootstrapTuple = primary.dump(primaryDbName, null);
+
+// Bootstrap load in replica
+replica.load(replicatedDbName, bootstrapTuple.dumpLocation)
+.status(replicatedDbName)
+.verifyResult(bootstrapTuple.lastReplicationId);
+
+// First incremental dump
+WarehouseInstance.Tuple firstIncremental = primary.run("use " + 
primaryDbName)
+.run("create table table1 (i int)")
+.run("create table table2 (id int) partitioned by (country 
string)")
+.run("insert into table1 values (1)")
+.run("insert into table2 partition(country='india') values(1)")
+.dump(primaryDbName, bootstrapTuple.lastReplicationId);
+
+// Second incremental dump
+WarehouseInstance.Tuple secondIncremental = primary.run("use " + 
primaryDbName)
+.run("drop table table1")
+.run("drop table table2")
+.run("create table table2 (id int) partitioned by (country 
string)")
+.run("alter table table2 add partition(country='india')")
+.run("alter table table2 drop partition(country='india')")
+.run("insert into table2 partition(country='us') values(2)")
+.run("create table table1 (i int)")
+.run("insert into table1 values (2)")
+.dump(primaryDbName, firstIncremental.lastReplicationId);
+
+// First incremental load
+replica.load(replicatedDbName, firstIncremental.dumpLocation)
+.status(replicatedDbName)
+.verifyResult(firstIncremental.lastReplicationId)
+.run("use " + replicatedDbName)
+.run("show tables")
+.verifyResults(new String[] {"table1", "table2"})
+.run("select * from table1")
+.verifyResults(new String[] {"1"})
+.run("select id from table2 order by id")
+.verifyResults(new String[] {"1"});
+
+// Second incremental load
+replica.load(replicatedDbName, secondIncremental.dumpLocation)
+.status(replicatedDbName)
+.verifyResult(secondIncremental.lastReplicationId)
+.run("use " + replicatedDbName)
+.run("show tables")
+.verifyResults(new String[] {"table1", "table2"})
+.run("select * from table1")
+.verifyResults(new String[] {"2"})
+.run("select id from table2 order by id")
+.verifyResults(new String[] {"2"});
+  }
 }

http://git-wip-us.apache.org/repos/asf/hive/blob/150e3516/itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/parse/WarehouseInstance.java
-

hive git commit: HIVE-19170 : Fix TestMiniDruidKafkaCliDriver (Nishant Bangarwa via Ashutosh Chauhan)

2018-04-12 Thread hashutosh
Repository: hive
Updated Branches:
  refs/heads/master 6a929c58f -> 6f9090c1d


HIVE-19170 : Fix TestMiniDruidKafkaCliDriver (Nishant Bangarwa via Ashutosh 
Chauhan)

Signed-off-by: Ashutosh Chauhan 


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/6f9090c1
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/6f9090c1
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/6f9090c1

Branch: refs/heads/master
Commit: 6f9090c1de9bd6d2d98312a74b73802044cfa23e
Parents: 6a929c5
Author: Nishant Bangarwa 
Authored: Thu Apr 12 22:41:54 2018 -0700
Committer: Ashutosh Chauhan 
Committed: Thu Apr 12 22:41:54 2018 -0700

--
 itests/src/test/resources/testconfiguration.properties | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/6f9090c1/itests/src/test/resources/testconfiguration.properties
--
diff --git a/itests/src/test/resources/testconfiguration.properties 
b/itests/src/test/resources/testconfiguration.properties
index c5d4e9f..2db98c9 100644
--- a/itests/src/test/resources/testconfiguration.properties
+++ b/itests/src/test/resources/testconfiguration.properties
@@ -1698,5 +1698,6 @@ druid.query.files=druidmini_test1.q,\
   druidmini_expressions.q,\
   druidmini_extractTime.q,\
   druidmini_test_alter.q,\
-  druidkafkamini_basic.q,\
   druidmini_floorTime.q
+
+druid.kafka.query.files=druidkafkamini_basic.q



hive git commit: HIVE-19172 : NPE due to null EnvironmentContext in DDLTask (Nishant Bangarwa via Ashutosh Chauhan)

2018-04-12 Thread hashutosh
Repository: hive
Updated Branches:
  refs/heads/master b7c64b15c -> 6a929c58f


HIVE-19172 : NPE due to null EnvironmentContext in DDLTask (Nishant Bangarwa 
via Ashutosh Chauhan)

Signed-off-by: Ashutosh Chauhan 


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/6a929c58
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/6a929c58
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/6a929c58

Branch: refs/heads/master
Commit: 6a929c58f943c41685f6e06a519381e696ad91ad
Parents: b7c64b1
Author: Nishant Bangarwa 
Authored: Thu Apr 12 22:39:37 2018 -0700
Committer: Ashutosh Chauhan 
Committed: Thu Apr 12 22:39:37 2018 -0700

--
 ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java | 3 +++
 1 file changed, 3 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/6a929c58/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
--
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
index 61a0432..b9b1830 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java
@@ -3827,6 +3827,9 @@ public class DDLTask extends Task implements 
Serializable {
 
 try {
   EnvironmentContext environmentContext = alterTbl.getEnvironmentContext();
+  if (environmentContext == null) {
+environmentContext = new EnvironmentContext();
+  }
   
environmentContext.putToProperties(HiveMetaHook.ALTER_TABLE_OPERATION_TYPE, 
alterTbl.getOp().name());
   if (allPartitions == null) {
 db.alterTable(alterTbl.getOldName(), tbl, alterTbl.getIsCascade(), 
environmentContext);



hive git commit: HIVE-17429: Hive JDBC doesn't return rows when querying Impala (Zach Amsden, reviewed by Aihua Xu)

2018-04-12 Thread aihuaxu
Repository: hive
Updated Branches:
  refs/heads/branch-2.3 f78128a71 -> 1ca13b3ce


HIVE-17429: Hive JDBC doesn't return rows when querying Impala (Zach Amsden, 
reviewed by Aihua Xu)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/1ca13b3c
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/1ca13b3c
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/1ca13b3c

Branch: refs/heads/branch-2.3
Commit: 1ca13b3cec97c4114090687079b6efbffeb498c3
Parents: f78128a
Author: Aihua Xu 
Authored: Thu Sep 7 17:17:14 2017 -0700
Committer: Aihua Xu 
Committed: Thu Apr 12 15:33:16 2018 -0700

--
 jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/1ca13b3c/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
--
diff --git a/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java 
b/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
index c385e2c..4c98517 100644
--- a/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
+++ b/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
@@ -254,7 +254,7 @@ public class HiveStatement implements java.sql.Statement {
 TGetOperationStatusResp status = waitForOperationToComplete();
 
 // The query should be completed by now
-if (!status.isHasResultSet()) {
+if (!status.isHasResultSet() && !stmtHandle.isHasResultSet()) {
   return false;
 }
 resultSet =  new 
HiveQueryResultSet.Builder(this).setClient(client).setSessionHandle(sessHandle)



hive git commit: HIVE-17429: Hive JDBC doesn't return rows when querying Impala (Zach Amsden, reviewed by Aihua Xu)

2018-04-12 Thread aihuaxu
Repository: hive
Updated Branches:
  refs/heads/branch-2.2 9e10b88c3 -> 1846c5a73


HIVE-17429: Hive JDBC doesn't return rows when querying Impala (Zach Amsden, 
reviewed by Aihua Xu)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/1846c5a7
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/1846c5a7
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/1846c5a7

Branch: refs/heads/branch-2.2
Commit: 1846c5a737ee32b49dbeb6dd58ca7968f68fa31a
Parents: 9e10b88
Author: Aihua Xu 
Authored: Thu Sep 7 17:17:14 2017 -0700
Committer: Aihua Xu 
Committed: Thu Apr 12 15:26:30 2018 -0700

--
 jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/1846c5a7/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
--
diff --git a/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java 
b/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
index ed0ed08..d54c964 100644
--- a/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
+++ b/jdbc/src/java/org/apache/hive/jdbc/HiveStatement.java
@@ -254,7 +254,7 @@ public class HiveStatement implements java.sql.Statement {
 TGetOperationStatusResp status = waitForOperationToComplete();
 
 // The query should be completed by now
-if (!status.isHasResultSet()) {
+if (!status.isHasResultSet() && !stmtHandle.isHasResultSet()) {
   return false;
 }
 resultSet =  new 
HiveQueryResultSet.Builder(this).setClient(client).setSessionHandle(sessHandle)



[10/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query4.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query4.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query4.q.out
index 5074d49..75c4fbb 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query4.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query4.q.out
@@ -225,20 +225,20 @@ STAGE PLANS:
 Reducer 10 <- Map 13 (PARTITION-LEVEL SORT, 398), Map 9 
(PARTITION-LEVEL SORT, 398)
 Reducer 11 <- Map 14 (PARTITION-LEVEL SORT, 975), Reducer 10 
(PARTITION-LEVEL SORT, 975)
 Reducer 12 <- Reducer 11 (GROUP, 481)
-Reducer 16 <- Map 15 (PARTITION-LEVEL SORT, 306), Map 19 
(PARTITION-LEVEL SORT, 306)
-Reducer 17 <- Map 20 (PARTITION-LEVEL SORT, 873), Reducer 16 
(PARTITION-LEVEL SORT, 873)
-Reducer 18 <- Reducer 17 (GROUP, 369)
+Reducer 16 <- Map 15 (PARTITION-LEVEL SORT, 398), Map 19 
(PARTITION-LEVEL SORT, 398)
+Reducer 17 <- Map 20 (PARTITION-LEVEL SORT, 975), Reducer 16 
(PARTITION-LEVEL SORT, 975)
+Reducer 18 <- Reducer 17 (GROUP, 481)
 Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 306), Map 7 (PARTITION-LEVEL 
SORT, 306)
-Reducer 22 <- Map 21 (PARTITION-LEVEL SORT, 154), Map 25 
(PARTITION-LEVEL SORT, 154)
-Reducer 23 <- Map 26 (PARTITION-LEVEL SORT, 706), Reducer 22 
(PARTITION-LEVEL SORT, 706)
-Reducer 24 <- Reducer 23 (GROUP, 186)
+Reducer 22 <- Map 21 (PARTITION-LEVEL SORT, 306), Map 25 
(PARTITION-LEVEL SORT, 306)
+Reducer 23 <- Map 26 (PARTITION-LEVEL SORT, 873), Reducer 22 
(PARTITION-LEVEL SORT, 873)
+Reducer 24 <- Reducer 23 (GROUP, 369)
 Reducer 28 <- Map 27 (PARTITION-LEVEL SORT, 154), Map 31 
(PARTITION-LEVEL SORT, 154)
 Reducer 29 <- Map 32 (PARTITION-LEVEL SORT, 706), Reducer 28 
(PARTITION-LEVEL SORT, 706)
 Reducer 3 <- Map 8 (PARTITION-LEVEL SORT, 873), Reducer 2 
(PARTITION-LEVEL SORT, 873)
 Reducer 30 <- Reducer 29 (GROUP, 186)
-Reducer 34 <- Map 33 (PARTITION-LEVEL SORT, 398), Map 37 
(PARTITION-LEVEL SORT, 398)
-Reducer 35 <- Map 38 (PARTITION-LEVEL SORT, 975), Reducer 34 
(PARTITION-LEVEL SORT, 975)
-Reducer 36 <- Reducer 35 (GROUP, 481)
+Reducer 34 <- Map 33 (PARTITION-LEVEL SORT, 154), Map 37 
(PARTITION-LEVEL SORT, 154)
+Reducer 35 <- Map 38 (PARTITION-LEVEL SORT, 706), Reducer 34 
(PARTITION-LEVEL SORT, 706)
+Reducer 36 <- Reducer 35 (GROUP, 186)
 Reducer 4 <- Reducer 3 (GROUP, 369)
 Reducer 5 <- Reducer 12 (PARTITION-LEVEL SORT, 690), Reducer 18 
(PARTITION-LEVEL SORT, 690), Reducer 24 (PARTITION-LEVEL SORT, 690), Reducer 30 
(PARTITION-LEVEL SORT, 690), Reducer 36 (PARTITION-LEVEL SORT, 690), Reducer 4 
(PARTITION-LEVEL SORT, 690)
 Reducer 6 <- Reducer 5 (SORT, 1)
@@ -300,20 +300,20 @@ STAGE PLANS:
 Map 15 
 Map Operator Tree:
 TableScan
-  alias: catalog_sales
-  Statistics: Num rows: 287989836 Data size: 38999608952 Basic 
stats: COMPLETE Column stats: NONE
+  alias: store_sales
+  Statistics: Num rows: 575995635 Data size: 50814502088 Basic 
stats: COMPLETE Column stats: NONE
   Filter Operator
-predicate: (cs_bill_customer_sk is not null and 
cs_sold_date_sk is not null) (type: boolean)
-Statistics: Num rows: 287989836 Data size: 38999608952 
Basic stats: COMPLETE Column stats: NONE
+predicate: (ss_customer_sk is not null and ss_sold_date_sk 
is not null) (type: boolean)
+Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
 Select Operator
-  expressions: cs_sold_date_sk (type: int), 
cs_bill_customer_sk (type: int), cs_ext_discount_amt (type: decimal(7,2)), 
cs_ext_sales_price (type: decimal(7,2)), cs_ext_wholesale_cost (type: 
decimal(7,2)), cs_ext_list_price (type: decimal(7,2))
+  expressions: ss_sold_date_sk (type: int), ss_customer_sk 
(type: int), ss_ext_discount_amt (type: decimal(7,2)), ss_ext_sales_price 
(type: decimal(7,2)), ss_ext_wholesale_cost (type: decimal(7,2)), 
ss_ext_list_price (type: decimal(7,2))
   outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5
-  Statistics: Num rows: 287989836 Data size: 38999608952 
Basic stats: COMPLETE Column stats: NONE
+  Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
   Reduce Output Operator
 key expressions: _col0 (type: int)
 sort order: +
 Map-reduce partition columns: _col0 (typ

[15/15] hive git commit: HIVE-19175 : TestMiniLlapLocalCliDriver.testCliDriver update_access_time_non_current_db failing (Vineet Garg via Peter Vary)

2018-04-12 Thread vgarg
HIVE-19175 : TestMiniLlapLocalCliDriver.testCliDriver 
update_access_time_non_current_db failing (Vineet Garg via Peter Vary)

Signed-off-by: Ashutosh Chauhan 


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/50f144d1
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/50f144d1
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/50f144d1

Branch: refs/heads/branch-3
Commit: 50f144d1f62b6c74d7d2b5f6a20232ab77f32be0
Parents: 5a95a82
Author: Vineet Garg 
Authored: Thu Apr 12 09:34:31 2018 -0700
Committer: Vineet Garg 
Committed: Thu Apr 12 10:58:27 2018 -0700

--
 .../test/resources/testconfiguration.properties |  2 +-
 .../update_access_time_non_current_db.q.out | 47 
 .../update_access_time_non_current_db.q.out | 47 
 3 files changed, 48 insertions(+), 48 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/50f144d1/itests/src/test/resources/testconfiguration.properties
--
diff --git a/itests/src/test/resources/testconfiguration.properties 
b/itests/src/test/resources/testconfiguration.properties
index 48d62a8..c5d4e9f 100644
--- a/itests/src/test/resources/testconfiguration.properties
+++ b/itests/src/test/resources/testconfiguration.properties
@@ -285,7 +285,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\
   union8.q,\
   union9.q,\
   union_stats.q,\
-  update_access_time_non_current_db.q, \
   update_after_multiple_inserts.q,\
   update_all_non_partitioned.q,\
   update_all_partitioned.q,\
@@ -739,6 +738,7 @@ minillaplocal.query.files=\
   union_fast_stats.q,\
   union_remove_26.q,\
   union_top_level.q,\
+  update_access_time_non_current_db.q, \
   vector_annotate_stats_select.q,\
   vector_auto_smb_mapjoin_14.q,\
   vector_char_varchar_1.q,\

http://git-wip-us.apache.org/repos/asf/hive/blob/50f144d1/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out
 
b/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out
new file mode 100644
index 000..7fda3ac
--- /dev/null
+++ 
b/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out
@@ -0,0 +1,47 @@
+PREHOOK: query: create database temp1
+PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:temp1
+POSTHOOK: query: create database temp1
+POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:temp1
+PREHOOK: query: use temp1
+PREHOOK: type: SWITCHDATABASE
+PREHOOK: Input: database:temp1
+POSTHOOK: query: use temp1
+POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:temp1
+PREHOOK: query: create table test1(id int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:temp1
+PREHOOK: Output: temp1@test1
+POSTHOOK: query: create table test1(id int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:temp1
+POSTHOOK: Output: temp1@test1
+PREHOOK: query: create database temp2
+PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:temp2
+POSTHOOK: query: create database temp2
+POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:temp2
+PREHOOK: query: use temp2
+PREHOOK: type: SWITCHDATABASE
+PREHOOK: Input: database:temp2
+POSTHOOK: query: use temp2
+POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:temp2
+PREHOOK: query: create table test2(id int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:temp2
+PREHOOK: Output: temp2@test2
+POSTHOOK: query: create table test2(id int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:temp2
+POSTHOOK: Output: temp2@test2
+POSTHOOK: query: use temp1
+POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:temp1
+POSTHOOK: query: desc temp2.test2
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: temp2@test2
+id int 

http://git-wip-us.apache.org/repos/asf/hive/blob/50f144d1/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out 
b/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out
deleted file mode 100644
index 7fda3ac..000
--- a/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out
+++ /dev/null
@@ -1,47 +0,0 @@
-PREHOOK: query: create database temp1
-PREHOOK: type: CREATEDATABASE
-PREHOOK: Output: database:temp1
-POSTHOOK: query: create database temp1
-POSTHOOK: type: CREATEDATABASE
-POSTHOOK: Output: database:temp1
-PREHOOK: query: use temp1
-PREHOOK: type: SWITCHDATABASE
-PREHOOK: Input: database:temp1
-PO

[14/15] hive git commit: HIVE-19156 : TestMiniLlapLocalCliDriver.vectorized_dynamic_semijoin_reduction.q is broken (Jason Dere via Deepak Jaiswal)

2018-04-12 Thread vgarg
HIVE-19156 : TestMiniLlapLocalCliDriver.vectorized_dynamic_semijoin_reduction.q 
is broken (Jason Dere via Deepak Jaiswal)

Signed-off-by: Ashutosh Chauhan 


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/5a95a826
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/5a95a826
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/5a95a826

Branch: refs/heads/branch-3
Commit: 5a95a826a1ce6db56a5a721ef99e3f76a9360abf
Parents: 12a1784
Author: Jason Dere 
Authored: Thu Apr 12 08:47:25 2018 -0700
Committer: Vineet Garg 
Committed: Thu Apr 12 10:58:15 2018 -0700

--
 .../ql/exec/tez/DynamicValueRegistryTez.java | 19 +--
 1 file changed, 17 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/5a95a826/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java
index ec1e84b..2d99f50 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java
@@ -61,6 +61,11 @@ public class DynamicValueRegistryTez implements 
DynamicValueRegistry {
 }
   }
 
+  static class NullValue {
+  }
+
+  static final NullValue NULL_VALUE = new NullValue();
+
   protected Map values = new ConcurrentHashMap<>();
 
   public DynamicValueRegistryTez() {
@@ -71,11 +76,21 @@ public class DynamicValueRegistryTez implements 
DynamicValueRegistry {
 if (!values.containsKey(key)) {
   throw new NoDynamicValuesException("Value does not exist in registry: " 
+ key);
 }
-return values.get(key);
+Object val = values.get(key);
+
+if (val == NULL_VALUE) {
+  return null;
+}
+return val;
   }
 
   protected void setValue(String key, Object value) {
-values.put(key, value);
+if (value == null) {
+  // ConcurrentHashMap does not allow null - use a substitute value.
+  values.put(key, NULL_VALUE);
+} else {
+  values.put(key, value);
+}
   }
 
   @Override



[08/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query50.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query50.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query50.q.out
index b978201..fde0c4e 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query50.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query50.q.out
@@ -124,7 +124,7 @@ STAGE PLANS:
 Spark
  A masked pattern was here 
   Vertices:
-Map 6 
+Map 10 
 Map Operator Tree:
 TableScan
   alias: store
@@ -138,7 +138,7 @@ STAGE PLANS:
   Statistics: Num rows: 1704 Data size: 3256276 Basic 
stats: COMPLETE Column stats: NONE
   Spark HashTable Sink Operator
 keys:
-  0 _col3 (type: int)
+  0 _col10 (type: int)
   1 _col0 (type: int)
 Local Work:
   Map Reduce Local Work
@@ -146,63 +146,16 @@ STAGE PLANS:
   Stage: Stage-1
 Spark
   Edges:
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 476), Reducer 8 
(PARTITION-LEVEL SORT, 476)
-Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 481), Reducer 2 
(PARTITION-LEVEL SORT, 481)
-Reducer 4 <- Reducer 3 (GROUP, 529)
-Reducer 5 <- Reducer 4 (SORT, 1)
-Reducer 8 <- Map 7 (PARTITION-LEVEL SORT, 36), Map 9 (PARTITION-LEVEL 
SORT, 36)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 36), Map 7 (PARTITION-LEVEL 
SORT, 36)
+Reducer 3 <- Map 8 (PARTITION-LEVEL SORT, 436), Reducer 2 
(PARTITION-LEVEL SORT, 436)
+Reducer 4 <- Map 9 (PARTITION-LEVEL SORT, 438), Reducer 3 
(PARTITION-LEVEL SORT, 438)
+Reducer 5 <- Reducer 4 (GROUP, 529)
+Reducer 6 <- Reducer 5 (SORT, 1)
  A masked pattern was here 
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
-  alias: store_sales
-  Statistics: Num rows: 575995635 Data size: 50814502088 Basic 
stats: COMPLETE Column stats: NONE
-  Filter Operator
-predicate: (ss_customer_sk is not null and ss_item_sk is 
not null and ss_sold_date_sk is not null and ss_store_sk is not null and 
ss_ticket_number is not null) (type: boolean)
-Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
-Select Operator
-  expressions: ss_sold_date_sk (type: int), ss_item_sk 
(type: int), ss_customer_sk (type: int), ss_store_sk (type: int), 
ss_ticket_number (type: int)
-  outputColumnNames: _col0, _col1, _col2, _col3, _col4
-  Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
-  Map Join Operator
-condition map:
- Inner Join 0 to 1
-keys:
-  0 _col3 (type: int)
-  1 _col0 (type: int)
-outputColumnNames: _col0, _col1, _col2, _col4, _col6, 
_col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15
-input vertices:
-  1 Map 6
-Statistics: Num rows: 633595212 Data size: 55895953508 
Basic stats: COMPLETE Column stats: NONE
-Reduce Output Operator
-  key expressions: _col1 (type: int), _col2 (type: 
int), _col4 (type: int)
-  sort order: +++
-  Map-reduce partition columns: _col1 (type: int), 
_col2 (type: int), _col4 (type: int)
-  Statistics: Num rows: 633595212 Data size: 
55895953508 Basic stats: COMPLETE Column stats: NONE
-  value expressions: _col0 (type: int), _col6 (type: 
string), _col7 (type: int), _col8 (type: string), _col9 (type: string), _col10 
(type: string), _col11 (type: string), _col12 (type: string), _col13 (type: 
string), _col14 (type: string), _col15 (type: string)
-Local Work:
-  Map Reduce Local Work
-Map 10 
-Map Operator Tree:
-TableScan
-  alias: d1
-  Statistics: Num rows: 73049 Data size: 81741831 Basic stats: 
COMPLETE Column stats: NONE
-  Filter Operator
-predicate: d_date_sk is not null (type: boolean)
-Statistics: Num rows: 73049 Data size: 81741831 Basic 
stats: COMPLETE Column stats: NONE
-Select Operator
-  expressions: d_date_sk (type: int)
-  outputColumnNames: _col0
-  Statistics: Num rows: 

[03/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query81.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query81.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query81.q.out
index 23998d6..5b2d5b3 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query81.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query81.q.out
@@ -66,37 +66,72 @@ STAGE PLANS:
   Stage: Stage-1
 Spark
   Edges:
-Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 25), Map 14 
(PARTITION-LEVEL SORT, 25)
-Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 344), Reducer 11 
(PARTITION-LEVEL SORT, 344)
-Reducer 13 <- Reducer 12 (GROUP PARTITION-LEVEL SORT, 349)
-Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 697), Map 18 
(PARTITION-LEVEL SORT, 697)
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 25), Map 8 (PARTITION-LEVEL 
SORT, 25)
-Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 344), Reducer 2 
(PARTITION-LEVEL SORT, 344)
-Reducer 4 <- Reducer 3 (GROUP, 349)
-Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 262), Reducer 4 
(PARTITION-LEVEL SORT, 262)
-Reducer 6 <- Reducer 17 (PARTITION-LEVEL SORT, 656), Reducer 5 
(PARTITION-LEVEL SORT, 656)
-Reducer 7 <- Reducer 6 (SORT, 1)
+Reducer 10 <- Reducer 16 (PARTITION-LEVEL SORT, 262), Reducer 9 
(PARTITION-LEVEL SORT, 262)
+Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 25), Map 17 
(PARTITION-LEVEL SORT, 25)
+Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 344), Reducer 14 
(PARTITION-LEVEL SORT, 344)
+Reducer 16 <- Reducer 15 (GROUP PARTITION-LEVEL SORT, 349)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 697), Map 5 (PARTITION-LEVEL 
SORT, 697)
+Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 656), Reducer 2 
(PARTITION-LEVEL SORT, 656)
+Reducer 4 <- Reducer 3 (SORT, 1)
+Reducer 7 <- Map 11 (PARTITION-LEVEL SORT, 25), Map 6 (PARTITION-LEVEL 
SORT, 25)
+Reducer 8 <- Map 12 (PARTITION-LEVEL SORT, 344), Reducer 7 
(PARTITION-LEVEL SORT, 344)
+Reducer 9 <- Reducer 8 (GROUP, 349)
  A masked pattern was here 
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
-  alias: catalog_returns
-  Statistics: Num rows: 28798881 Data size: 3057234680 Basic 
stats: COMPLETE Column stats: NONE
+  alias: customer
+  Statistics: Num rows: 8000 Data size: 68801615852 Basic 
stats: COMPLETE Column stats: NONE
   Filter Operator
-predicate: (cr_returned_date_sk is not null and 
cr_returning_addr_sk is not null and cr_returning_customer_sk is not null) 
(type: boolean)
-Statistics: Num rows: 28798881 Data size: 3057234680 Basic 
stats: COMPLETE Column stats: NONE
+predicate: (c_current_addr_sk is not null and 
c_customer_sk is not null) (type: boolean)
+Statistics: Num rows: 8000 Data size: 68801615852 
Basic stats: COMPLETE Column stats: NONE
 Select Operator
-  expressions: cr_returned_date_sk (type: int), 
cr_returning_customer_sk (type: int), cr_returning_addr_sk (type: int), 
cr_return_amt_inc_tax (type: decimal(7,2))
-  outputColumnNames: _col0, _col1, _col2, _col3
-  Statistics: Num rows: 28798881 Data size: 3057234680 
Basic stats: COMPLETE Column stats: NONE
+  expressions: c_customer_sk (type: int), c_customer_id 
(type: string), c_current_addr_sk (type: int), c_salutation (type: string), 
c_first_name (type: string), c_last_name (type: string)
+  outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5
+  Statistics: Num rows: 8000 Data size: 68801615852 
Basic stats: COMPLETE Column stats: NONE
+  Reduce Output Operator
+key expressions: _col2 (type: int)
+sort order: +
+Map-reduce partition columns: _col2 (type: int)
+Statistics: Num rows: 8000 Data size: 68801615852 
Basic stats: COMPLETE Column stats: NONE
+value expressions: _col0 (type: int), _col1 (type: 
string), _col3 (type: string), _col4 (type: string), _col5 (type: string)
+Map 11 
+Map Operator Tree:
+TableScan
+  alias: date_dim
+  Statistics: Num rows: 73049 Data size: 81741831 Basic stats: 
COMPLETE Column stats: NONE
+  Filter Operator
+predicate: ((d_year = 1998) and d_date_sk is not null) 
(type: boolean)
+Statistics: Num rows: 36524 Data size: 40870356 Basic 
stats: COMPLETE Column stats: NONE
+Select

[11/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query29.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query29.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query29.q.out
index 456fce7..530a625 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query29.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query29.q.out
@@ -124,7 +124,7 @@ STAGE PLANS:
 Spark
  A masked pattern was here 
   Vertices:
-Map 14 
+Map 13 
 Map Operator Tree:
 TableScan
   alias: d2
@@ -146,51 +146,51 @@ STAGE PLANS:
   Stage: Stage-1
 Spark
   Edges:
-Reducer 10 <- Map 12 (PARTITION-LEVEL SORT, 306), Map 9 
(PARTITION-LEVEL SORT, 306)
-Reducer 11 <- Map 13 (PARTITION-LEVEL SORT, 374), Reducer 10 
(PARTITION-LEVEL SORT, 374)
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 398), Map 7 (PARTITION-LEVEL 
SORT, 398)
-Reducer 3 <- Map 8 (PARTITION-LEVEL SORT, 442), Reducer 2 
(PARTITION-LEVEL SORT, 442)
-Reducer 4 <- Reducer 11 (PARTITION-LEVEL SORT, 850), Reducer 3 
(PARTITION-LEVEL SORT, 850)
-Reducer 5 <- Reducer 4 (GROUP, 582)
-Reducer 6 <- Reducer 5 (SORT, 1)
+Reducer 10 <- Map 14 (PARTITION-LEVEL SORT, 486), Reducer 9 
(PARTITION-LEVEL SORT, 486)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 306), Map 6 (PARTITION-LEVEL 
SORT, 306)
+Reducer 3 <- Reducer 10 (PARTITION-LEVEL SORT, 917), Reducer 2 
(PARTITION-LEVEL SORT, 917)
+Reducer 4 <- Reducer 3 (GROUP, 640)
+Reducer 5 <- Reducer 4 (SORT, 1)
+Reducer 8 <- Map 11 (PARTITION-LEVEL SORT, 398), Map 7 
(PARTITION-LEVEL SORT, 398)
+Reducer 9 <- Map 12 (PARTITION-LEVEL SORT, 476), Reducer 8 
(PARTITION-LEVEL SORT, 476)
  A masked pattern was here 
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
-  alias: store_sales
-  Statistics: Num rows: 575995635 Data size: 50814502088 Basic 
stats: COMPLETE Column stats: NONE
+  alias: catalog_sales
+  Statistics: Num rows: 287989836 Data size: 38999608952 Basic 
stats: COMPLETE Column stats: NONE
   Filter Operator
-predicate: (ss_customer_sk is not null and ss_item_sk is 
not null and ss_sold_date_sk is not null and ss_store_sk is not null and 
ss_ticket_number is not null) (type: boolean)
-Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
+predicate: (cs_bill_customer_sk is not null and cs_item_sk 
is not null and cs_sold_date_sk is not null) (type: boolean)
+Statistics: Num rows: 287989836 Data size: 38999608952 
Basic stats: COMPLETE Column stats: NONE
 Select Operator
-  expressions: ss_sold_date_sk (type: int), ss_item_sk 
(type: int), ss_customer_sk (type: int), ss_store_sk (type: int), 
ss_ticket_number (type: int), ss_quantity (type: int)
-  outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5
-  Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
+  expressions: cs_sold_date_sk (type: int), 
cs_bill_customer_sk (type: int), cs_item_sk (type: int), cs_quantity (type: int)
+  outputColumnNames: _col0, _col1, _col2, _col3
+  Statistics: Num rows: 287989836 Data size: 38999608952 
Basic stats: COMPLETE Column stats: NONE
   Reduce Output Operator
 key expressions: _col0 (type: int)
 sort order: +
 Map-reduce partition columns: _col0 (type: int)
-Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
-value expressions: _col1 (type: int), _col2 (type: 
int), _col3 (type: int), _col4 (type: int), _col5 (type: int)
-Map 12 
+Statistics: Num rows: 287989836 Data size: 38999608952 
Basic stats: COMPLETE Column stats: NONE
+value expressions: _col1 (type: int), _col2 (type: 
int), _col3 (type: int)
+Map 11 
 Map Operator Tree:
 TableScan
-  alias: d3
+  alias: d1
   Statistics: Num rows: 73049 Data size: 81741831 Basic stats: 
COMPLETE Column stats: NONE
   Filter Operator
-predicate: ((d_year) IN (1999, 2000, 2001) and d_date_sk 
is not null) (type: boolean)
-Statistics: Num rows: 36525 Data size: 40871475 Basic 
stats: COMPLETE Column stats: NONE
+predicate: ((d

[12/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query19.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query19.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query19.q.out
index 453bc56..281445c 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query19.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query19.q.out
@@ -70,7 +70,7 @@ STAGE PLANS:
   Statistics: Num rows: 1704 Data size: 3256276 Basic 
stats: COMPLETE Column stats: NONE
   Spark HashTable Sink Operator
 keys:
-  0 _col3 (type: int)
+  0 _col7 (type: int)
   1 _col0 (type: int)
 Local Work:
   Map Reduce Local Work
@@ -78,35 +78,17 @@ STAGE PLANS:
   Stage: Stage-1
 Spark
   Edges:
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 398), Map 8 (PARTITION-LEVEL 
SORT, 398)
-Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 440), Reducer 2 
(PARTITION-LEVEL SORT, 440)
-Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 1009), Reducer 3 
(PARTITION-LEVEL SORT, 1009)
-Reducer 5 <- Map 11 (PARTITION-LEVEL SORT, 846), Reducer 4 
(PARTITION-LEVEL SORT, 846)
-Reducer 6 <- Reducer 5 (GROUP, 640)
-Reducer 7 <- Reducer 6 (SORT, 1)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 855), Map 6 (PARTITION-LEVEL 
SORT, 855)
+Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 1009), Reducer 9 
(PARTITION-LEVEL SORT, 1009)
+Reducer 4 <- Reducer 3 (GROUP, 582)
+Reducer 5 <- Reducer 4 (SORT, 1)
+Reducer 8 <- Map 10 (PARTITION-LEVEL SORT, 398), Map 7 
(PARTITION-LEVEL SORT, 398)
+Reducer 9 <- Map 11 (PARTITION-LEVEL SORT, 440), Reducer 8 
(PARTITION-LEVEL SORT, 440)
  A masked pattern was here 
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
-  alias: store_sales
-  Statistics: Num rows: 575995635 Data size: 50814502088 Basic 
stats: COMPLETE Column stats: NONE
-  Filter Operator
-predicate: (ss_customer_sk is not null and ss_item_sk is 
not null and ss_sold_date_sk is not null and ss_store_sk is not null) (type: 
boolean)
-Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
-Select Operator
-  expressions: ss_sold_date_sk (type: int), ss_item_sk 
(type: int), ss_customer_sk (type: int), ss_store_sk (type: int), 
ss_ext_sales_price (type: decimal(7,2))
-  outputColumnNames: _col0, _col1, _col2, _col3, _col4
-  Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
-  Reduce Output Operator
-key expressions: _col0 (type: int)
-sort order: +
-Map-reduce partition columns: _col0 (type: int)
-Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
-value expressions: _col1 (type: int), _col2 (type: 
int), _col3 (type: int), _col4 (type: decimal(7,2))
-Map 10 
-Map Operator Tree:
-TableScan
   alias: customer
   Statistics: Num rows: 8000 Data size: 68801615852 Basic 
stats: COMPLETE Column stats: NONE
   Filter Operator
@@ -117,30 +99,12 @@ STAGE PLANS:
   outputColumnNames: _col0, _col1
   Statistics: Num rows: 8000 Data size: 68801615852 
Basic stats: COMPLETE Column stats: NONE
   Reduce Output Operator
-key expressions: _col0 (type: int)
+key expressions: _col1 (type: int)
 sort order: +
-Map-reduce partition columns: _col0 (type: int)
+Map-reduce partition columns: _col1 (type: int)
 Statistics: Num rows: 8000 Data size: 68801615852 
Basic stats: COMPLETE Column stats: NONE
-value expressions: _col1 (type: int)
-Map 11 
-Map Operator Tree:
-TableScan
-  alias: customer_address
-  Statistics: Num rows: 4000 Data size: 40595195284 Basic 
stats: COMPLETE Column stats: NONE
-  Filter Operator
-predicate: ca_address_sk is not null (type: boolean)
-Statistics: Num rows: 4000 Data size: 40595195284 
Basic stats: COMPLETE Column stats: NONE
-Select Operator
-  expressions: ca_address_sk (type: int), ca_zip (type: 

[13/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan 
Haindrich via Ashutosh Chauhan)

Signed-off-by: Ashutosh Chauhan 


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/12a1784d
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/12a1784d
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/12a1784d

Branch: refs/heads/branch-3
Commit: 12a1784de44f185a963c34462a8f863db53920ce
Parents: 91d0a0a
Author: Zoltan Haindrich 
Authored: Wed Apr 11 20:05:01 2018 -0700
Committer: Vineet Garg 
Committed: Thu Apr 12 10:57:23 2018 -0700

--
 data/conf/perf-reg/spark/hive-site.xml  |   6 +
 data/conf/perf-reg/tez/hive-site.xml|   5 +
 .../hive/cli/control/CorePerfCliDriver.java |   6 +-
 .../hadoop/hive/ql/MetaStoreDumpUtility.java| 231 +
 .../org/apache/hadoop/hive/ql/QTestUtil.java| 221 +
 .../hive/ql/parse/BaseSemanticAnalyzer.java |   2 +-
 .../hadoop/hive/ql/parse/SemanticAnalyzer.java  |   2 +-
 .../clientpositive/perf/spark/query11.q.out |  48 +-
 .../clientpositive/perf/spark/query15.q.out | 164 +++
 .../clientpositive/perf/spark/query16.q.out |  70 +--
 .../clientpositive/perf/spark/query18.q.out | 216 
 .../clientpositive/perf/spark/query19.q.out | 218 -
 .../clientpositive/perf/spark/query21.q.out | 114 +++--
 .../clientpositive/perf/spark/query24.q.out | 282 +--
 .../clientpositive/perf/spark/query25.q.out | 118 ++---
 .../clientpositive/perf/spark/query29.q.out | 280 +--
 .../clientpositive/perf/spark/query30.q.out | 262 +-
 .../clientpositive/perf/spark/query32.q.out | 132 ++---
 .../clientpositive/perf/spark/query34.q.out |  34 +-
 .../clientpositive/perf/spark/query35.q.out |  74 ++-
 .../clientpositive/perf/spark/query37.q.out |  16 +-
 .../clientpositive/perf/spark/query4.q.out  | 214 
 .../clientpositive/perf/spark/query40.q.out | 116 +++--
 .../clientpositive/perf/spark/query44.q.out | 246 +-
 .../clientpositive/perf/spark/query45.q.out | 204 
 .../clientpositive/perf/spark/query46.q.out | 104 ++--
 .../clientpositive/perf/spark/query47.q.out |  92 ++--
 .../clientpositive/perf/spark/query48.q.out |  94 ++--
 .../clientpositive/perf/spark/query5.q.out  |  38 +-
 .../clientpositive/perf/spark/query50.q.out | 196 
 .../clientpositive/perf/spark/query53.q.out | 105 ++--
 .../clientpositive/perf/spark/query54.q.out | 277 ++-
 .../clientpositive/perf/spark/query57.q.out |  92 ++--
 .../clientpositive/perf/spark/query58.q.out | 490 +--
 .../clientpositive/perf/spark/query6.q.out  | 350 +++--
 .../clientpositive/perf/spark/query61.q.out |  40 +-
 .../clientpositive/perf/spark/query63.q.out | 105 ++--
 .../clientpositive/perf/spark/query65.q.out | 100 ++--
 .../clientpositive/perf/spark/query66.q.out |  20 +-
 .../clientpositive/perf/spark/query67.q.out | 137 +++---
 .../clientpositive/perf/spark/query68.q.out | 104 ++--
 .../clientpositive/perf/spark/query72.q.out | 461 +
 .../clientpositive/perf/spark/query73.q.out |  34 +-
 .../clientpositive/perf/spark/query75.q.out | 248 +-
 .../clientpositive/perf/spark/query76.q.out | 130 +++--
 .../clientpositive/perf/spark/query77.q.out |  66 +--
 .../clientpositive/perf/spark/query78.q.out | 273 +--
 .../clientpositive/perf/spark/query79.q.out |  10 +-
 .../clientpositive/perf/spark/query8.q.out  | 174 +++
 .../clientpositive/perf/spark/query80.q.out | 417 
 .../clientpositive/perf/spark/query81.q.out | 270 +-
 .../clientpositive/perf/spark/query82.q.out |  16 +-
 .../clientpositive/perf/spark/query83.q.out | 354 +++---
 .../clientpositive/perf/spark/query85.q.out | 309 ++--
 .../clientpositive/perf/spark/query88.q.out | 272 +-
 .../clientpositive/perf/spark/query89.q.out |  26 +-
 .../clientpositive/perf/spark/query90.q.out |  68 +--
 .../clientpositive/perf/spark/query91.q.out |  14 +-
 .../clientpositive/perf/spark/query92.q.out | 106 ++--
 .../clientpositive/perf/spark/query94.q.out |  70 +--
 .../clientpositive/perf/spark/query95.q.out | 100 ++--
 .../clientpositive/perf/spark/query97.q.out |  54 +-
 .../clientpositive/perf/spark/query99.q.out |  42 +-
 63 files changed, 4603 insertions(+), 4536 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/data/conf/perf-reg/spark/hive-site.xml
--
diff --git a/data/conf/perf-reg/spark/hive-site.xml 
b/data/conf/perf-reg/spark/hive-site.xml
index

[06/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query65.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query65.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query65.q.out
index 575fc5c..3b3baef 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query65.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query65.q.out
@@ -57,8 +57,7 @@ POSTHOOK: type: QUERY
 STAGE DEPENDENCIES:
   Stage-2 is a root stage
   Stage-3 depends on stages: Stage-2
-  Stage-4 depends on stages: Stage-3
-  Stage-1 depends on stages: Stage-4
+  Stage-1 depends on stages: Stage-3
   Stage-0 depends on stages: Stage-1
 
 STAGE PLANS:
@@ -66,29 +65,6 @@ STAGE PLANS:
 Spark
  A masked pattern was here 
   Vertices:
-Map 11 
-Map Operator Tree:
-TableScan
-  alias: store
-  Statistics: Num rows: 1704 Data size: 3256276 Basic stats: 
COMPLETE Column stats: NONE
-  Filter Operator
-predicate: s_store_sk is not null (type: boolean)
-Statistics: Num rows: 1704 Data size: 3256276 Basic stats: 
COMPLETE Column stats: NONE
-Select Operator
-  expressions: s_store_sk (type: int), s_store_name (type: 
string)
-  outputColumnNames: _col0, _col1
-  Statistics: Num rows: 1704 Data size: 3256276 Basic 
stats: COMPLETE Column stats: NONE
-  Spark HashTable Sink Operator
-keys:
-  0 _col0 (type: int)
-  1 _col0 (type: int)
-Local Work:
-  Map Reduce Local Work
-
-  Stage: Stage-3
-Spark
- A masked pattern was here 
-  Vertices:
 Map 6 
 Map Operator Tree:
 TableScan
@@ -108,7 +84,7 @@ STAGE PLANS:
 Local Work:
   Map Reduce Local Work
 
-  Stage: Stage-4
+  Stage: Stage-3
 Spark
  A masked pattern was here 
   Vertices:
@@ -135,8 +111,8 @@ STAGE PLANS:
 Spark
   Edges:
 Reducer 2 <- Map 1 (GROUP, 437)
-Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 328), Reducer 8 
(PARTITION-LEVEL SORT, 328)
-Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 86), Reducer 3 
(PARTITION-LEVEL SORT, 86)
+Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 328), Reducer 2 
(PARTITION-LEVEL SORT, 328), Reducer 8 (PARTITION-LEVEL SORT, 328)
+Reducer 4 <- Map 11 (PARTITION-LEVEL SORT, 166), Reducer 3 
(PARTITION-LEVEL SORT, 166)
 Reducer 5 <- Reducer 4 (SORT, 1)
 Reducer 8 <- Map 7 (GROUP PARTITION-LEVEL SORT, 437)
  A masked pattern was here 
@@ -180,6 +156,24 @@ STAGE PLANS:
 Map 10 
 Map Operator Tree:
 TableScan
+  alias: store
+  Statistics: Num rows: 1704 Data size: 3256276 Basic stats: 
COMPLETE Column stats: NONE
+  Filter Operator
+predicate: s_store_sk is not null (type: boolean)
+Statistics: Num rows: 1704 Data size: 3256276 Basic stats: 
COMPLETE Column stats: NONE
+Select Operator
+  expressions: s_store_sk (type: int), s_store_name (type: 
string)
+  outputColumnNames: _col0, _col1
+  Statistics: Num rows: 1704 Data size: 3256276 Basic 
stats: COMPLETE Column stats: NONE
+  Reduce Output Operator
+key expressions: _col0 (type: int)
+sort order: +
+Map-reduce partition columns: _col0 (type: int)
+Statistics: Num rows: 1704 Data size: 3256276 Basic 
stats: COMPLETE Column stats: NONE
+value expressions: _col1 (type: string)
+Map 11 
+Map Operator Tree:
+TableScan
   alias: item
   Statistics: Num rows: 462000 Data size: 663560457 Basic 
stats: COMPLETE Column stats: NONE
   Filter Operator
@@ -250,23 +244,23 @@ STAGE PLANS:
   Join Operator
 condition map:
  Inner Join 0 to 1
+ Inner Join 0 to 2
 keys:
   0 _col0 (type: int)
   1 _col0 (type: int)
-outputColumnNames: _col0, _col1, _col2, _col4
-Statistics: Num rows: 348477374 Data size: 30742775095 Basic 
stats: COMPLETE Column stats: NONE
+  2 _col0 (type: int)
+outputColumnNames: _col1, _col2, _col4, _col6
+Statistics: Num rows: 696954748 Data size: 61485550191 Basic 
stats: COMPLETE Column stats: NONE
 Filter Operator
   predica

[05/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query75.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query75.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query75.q.out
index f4fca34..4b76570 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query75.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query75.q.out
@@ -144,25 +144,25 @@ STAGE PLANS:
   Stage: Stage-1
 Spark
   Edges:
-Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 400), Map 16 
(PARTITION-LEVEL SORT, 400)
-Reducer 14 <- Map 10 (PARTITION-LEVEL SORT, 438), Reducer 13 
(PARTITION-LEVEL SORT, 438)
+Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 398), Map 16 
(PARTITION-LEVEL SORT, 398)
+Reducer 14 <- Map 10 (PARTITION-LEVEL SORT, 440), Reducer 13 
(PARTITION-LEVEL SORT, 440)
 Reducer 15 <- Map 18 (PARTITION-LEVEL SORT, 516), Reducer 14 
(PARTITION-LEVEL SORT, 516)
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 308), Map 16 
(PARTITION-LEVEL SORT, 308)
-Reducer 20 <- Map 16 (PARTITION-LEVEL SORT, 156), Map 19 
(PARTITION-LEVEL SORT, 156)
-Reducer 21 <- Map 10 (PARTITION-LEVEL SORT, 169), Reducer 20 
(PARTITION-LEVEL SORT, 169)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 306), Map 16 
(PARTITION-LEVEL SORT, 306)
+Reducer 20 <- Map 16 (PARTITION-LEVEL SORT, 154), Map 19 
(PARTITION-LEVEL SORT, 154)
+Reducer 21 <- Map 10 (PARTITION-LEVEL SORT, 171), Reducer 20 
(PARTITION-LEVEL SORT, 171)
 Reducer 22 <- Map 25 (PARTITION-LEVEL SORT, 196), Reducer 21 
(PARTITION-LEVEL SORT, 196)
-Reducer 27 <- Map 26 (PARTITION-LEVEL SORT, 308), Map 32 
(PARTITION-LEVEL SORT, 308)
-Reducer 28 <- Map 33 (PARTITION-LEVEL SORT, 336), Reducer 27 
(PARTITION-LEVEL SORT, 336)
+Reducer 27 <- Map 26 (PARTITION-LEVEL SORT, 306), Map 32 
(PARTITION-LEVEL SORT, 306)
+Reducer 28 <- Map 33 (PARTITION-LEVEL SORT, 338), Reducer 27 
(PARTITION-LEVEL SORT, 338)
 Reducer 29 <- Map 34 (PARTITION-LEVEL SORT, 393), Reducer 28 
(PARTITION-LEVEL SORT, 393)
-Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 336), Reducer 2 
(PARTITION-LEVEL SORT, 336)
+Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 338), Reducer 2 
(PARTITION-LEVEL SORT, 338)
 Reducer 30 <- Reducer 29 (GROUP, 934), Reducer 38 (GROUP, 934)
 Reducer 31 <- Reducer 30 (GROUP PARTITION-LEVEL SORT, 671), Reducer 45 
(GROUP PARTITION-LEVEL SORT, 671)
-Reducer 36 <- Map 32 (PARTITION-LEVEL SORT, 400), Map 35 
(PARTITION-LEVEL SORT, 400)
-Reducer 37 <- Map 33 (PARTITION-LEVEL SORT, 438), Reducer 36 
(PARTITION-LEVEL SORT, 438)
+Reducer 36 <- Map 32 (PARTITION-LEVEL SORT, 398), Map 35 
(PARTITION-LEVEL SORT, 398)
+Reducer 37 <- Map 33 (PARTITION-LEVEL SORT, 440), Reducer 36 
(PARTITION-LEVEL SORT, 440)
 Reducer 38 <- Map 41 (PARTITION-LEVEL SORT, 516), Reducer 37 
(PARTITION-LEVEL SORT, 516)
 Reducer 4 <- Map 11 (PARTITION-LEVEL SORT, 393), Reducer 3 
(PARTITION-LEVEL SORT, 393)
-Reducer 43 <- Map 32 (PARTITION-LEVEL SORT, 156), Map 42 
(PARTITION-LEVEL SORT, 156)
-Reducer 44 <- Map 33 (PARTITION-LEVEL SORT, 169), Reducer 43 
(PARTITION-LEVEL SORT, 169)
+Reducer 43 <- Map 32 (PARTITION-LEVEL SORT, 154), Map 42 
(PARTITION-LEVEL SORT, 154)
+Reducer 44 <- Map 33 (PARTITION-LEVEL SORT, 171), Reducer 43 
(PARTITION-LEVEL SORT, 171)
 Reducer 45 <- Map 48 (PARTITION-LEVEL SORT, 196), Reducer 44 
(PARTITION-LEVEL SORT, 196)
 Reducer 5 <- Reducer 15 (GROUP, 934), Reducer 4 (GROUP, 934)
 Reducer 6 <- Reducer 22 (GROUP PARTITION-LEVEL SORT, 671), Reducer 5 
(GROUP PARTITION-LEVEL SORT, 671)
@@ -183,28 +183,29 @@ STAGE PLANS:
   outputColumnNames: _col0, _col1, _col2, _col3, _col4
   Statistics: Num rows: 287989836 Data size: 38999608952 
Basic stats: COMPLETE Column stats: NONE
   Reduce Output Operator
-key expressions: _col1 (type: int)
+key expressions: _col0 (type: int)
 sort order: +
-Map-reduce partition columns: _col1 (type: int)
+Map-reduce partition columns: _col0 (type: int)
 Statistics: Num rows: 287989836 Data size: 38999608952 
Basic stats: COMPLETE Column stats: NONE
-value expressions: _col0 (type: int), _col2 (type: 
int), _col3 (type: int), _col4 (type: decimal(7,2))
+value expressions: _col1 (type: int), _col2 (type: 
int), _col3 (type: int), _col4 (type: decimal(7,2))
 Map 10 
 Map Operator Tree:
 TableScan
-  alias: date_dim
-  Statistics: Num rows: 73049 Data size: 81741831 Basic stats:

[07/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query58.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query58.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query58.q.out
index eb5dffe..f06cbef 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query58.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query58.q.out
@@ -1,6 +1,6 @@
-Warning: Map Join MAPJOIN[180][bigTable=?] in task 'Stage-1:MAPRED' is a cross 
product
-Warning: Map Join MAPJOIN[181][bigTable=?] in task 'Stage-1:MAPRED' is a cross 
product
-Warning: Map Join MAPJOIN[182][bigTable=?] in task 'Stage-1:MAPRED' is a cross 
product
+Warning: Map Join MAPJOIN[183][bigTable=?] in task 'Stage-1:MAPRED' is a cross 
product
+Warning: Map Join MAPJOIN[184][bigTable=?] in task 'Stage-1:MAPRED' is a cross 
product
+Warning: Map Join MAPJOIN[185][bigTable=?] in task 'Stage-1:MAPRED' is a cross 
product
 PREHOOK: query: explain
 with ss_items as
  (select i_item_id item_id
@@ -140,10 +140,10 @@ STAGE PLANS:
   Stage: Stage-2
 Spark
   Edges:
-Reducer 12 <- Map 11 (GROUP, 1)
+Reducer 11 <- Map 10 (GROUP, 1)
  A masked pattern was here 
   Vertices:
-Map 11 
+Map 10 
 Map Operator Tree:
 TableScan
   alias: date_dim
@@ -162,7 +162,7 @@ STAGE PLANS:
   sort order: 
   Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
   value expressions: _col0 (type: bigint)
-Reducer 12 
+Reducer 11 
 Local Work:
   Map Reduce Local Work
 Reduce Operator Tree:
@@ -184,10 +184,10 @@ STAGE PLANS:
   Stage: Stage-3
 Spark
   Edges:
-Reducer 25 <- Map 24 (GROUP, 1)
+Reducer 24 <- Map 23 (GROUP, 1)
  A masked pattern was here 
   Vertices:
-Map 24 
+Map 23 
 Map Operator Tree:
 TableScan
   alias: date_dim
@@ -206,7 +206,7 @@ STAGE PLANS:
   sort order: 
   Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
   value expressions: _col0 (type: bigint)
-Reducer 25 
+Reducer 24 
 Local Work:
   Map Reduce Local Work
 Reduce Operator Tree:
@@ -228,10 +228,10 @@ STAGE PLANS:
   Stage: Stage-4
 Spark
   Edges:
-Reducer 38 <- Map 37 (GROUP, 1)
+Reducer 37 <- Map 36 (GROUP, 1)
  A masked pattern was here 
   Vertices:
-Map 37 
+Map 36 
 Map Operator Tree:
 TableScan
   alias: date_dim
@@ -250,7 +250,7 @@ STAGE PLANS:
   sort order: 
   Statistics: Num rows: 1 Data size: 8 Basic stats: 
COMPLETE Column stats: NONE
   value expressions: _col0 (type: bigint)
-Reducer 38 
+Reducer 37 
 Local Work:
   Map Reduce Local Work
 Reduce Operator Tree:
@@ -272,47 +272,76 @@ STAGE PLANS:
   Stage: Stage-1
 Spark
   Edges:
-Reducer 10 <- Reducer 9 (GROUP, 2)
-Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 2), Reducer 23 
(PARTITION-LEVEL SORT, 2)
-Reducer 18 <- Map 27 (PARTITION-LEVEL SORT, 398), Reducer 17 
(PARTITION-LEVEL SORT, 398)
-Reducer 19 <- Map 28 (PARTITION-LEVEL SORT, 442), Reducer 18 
(PARTITION-LEVEL SORT, 442)
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2), Reducer 10 
(PARTITION-LEVEL SORT, 2)
-Reducer 20 <- Reducer 19 (GROUP, 481)
-Reducer 22 <- Map 21 (PARTITION-LEVEL SORT, 2), Map 26 
(PARTITION-LEVEL SORT, 2)
-Reducer 23 <- Reducer 22 (GROUP, 2)
-Reducer 3 <- Map 14 (PARTITION-LEVEL SORT, 306), Reducer 2 
(PARTITION-LEVEL SORT, 306)
-Reducer 30 <- Map 29 (PARTITION-LEVEL SORT, 2), Reducer 36 
(PARTITION-LEVEL SORT, 2)
-Reducer 31 <- Map 40 (PARTITION-LEVEL SORT, 154), Reducer 30 
(PARTITION-LEVEL SORT, 154)
-Reducer 32 <- Map 41 (PARTITION-LEVEL SORT, 174), Reducer 31 
(PARTITION-LEVEL SORT, 174)
-Reducer 33 <- Reducer 32 (GROUP, 186)
-Reducer 35 <- Map 34 (PARTITION-LEVEL SORT, 2), Map 39 
(PARTITION-LEVEL SORT, 2)
-Reducer 36 <- Reducer 35 (GROUP, 2)
-Reducer 4 <- Map 15 (PARTITION-LEVEL SORT, 341), Reducer 3 
(PARTITION-LEVEL SORT, 341)
-Reducer 5 <- Reducer 4 (GROUP, 369)
-Reducer 6 <- Reducer 20 (PARTITION-LEVEL SORT, 518), Reducer 33 
(PARTITION-LEVEL SORT, 518), Reducer 5 (PARTITION-LEVEL SORT, 518)
-Reducer 7 <- Reducer 6 (SORT, 1)
-Reducer 9 <- Map 13 (PARTITION-LEVEL SORT, 2), Map 8 (PARTITION-LEVEL 
SORT, 2)
+Reducer 13 <- Map 12 (P

[02/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query85.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query85.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query85.q.out
index c82dcf7..d60751c 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query85.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query85.q.out
@@ -175,7 +175,7 @@ STAGE PLANS:
 Spark
  A masked pattern was here 
   Vertices:
-Map 15 
+Map 13 
 Map Operator Tree:
 TableScan
   alias: reason
@@ -212,21 +212,21 @@ STAGE PLANS:
   Statistics: Num rows: 4602 Data size: 2696178 Basic 
stats: COMPLETE Column stats: NONE
   Spark HashTable Sink Operator
 keys:
-  0 _col0 (type: int)
-  1 _col2 (type: int)
+  0 _col10 (type: int)
+  1 _col0 (type: int)
 Local Work:
   Map Reduce Local Work
 
   Stage: Stage-1
 Spark
   Edges:
-Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 20), Map 14 
(PARTITION-LEVEL SORT, 20)
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 90), Map 8 (PARTITION-LEVEL 
SORT, 90)
-Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 17), Reducer 2 
(PARTITION-LEVEL SORT, 17)
-Reducer 4 <- Map 10 (PARTITION-LEVEL SORT, 19), Reducer 3 
(PARTITION-LEVEL SORT, 19)
-Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 35), Reducer 4 
(PARTITION-LEVEL SORT, 35)
-Reducer 6 <- Reducer 5 (GROUP, 2)
-Reducer 7 <- Reducer 6 (SORT, 1)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 28), Map 9 (PARTITION-LEVEL 
SORT, 28)
+Reducer 3 <- Map 10 (PARTITION-LEVEL SORT, 98), Reducer 2 
(PARTITION-LEVEL SORT, 98)
+Reducer 4 <- Map 12 (PARTITION-LEVEL SORT, 5), Reducer 3 
(PARTITION-LEVEL SORT, 5)
+Reducer 5 <- Map 14 (PARTITION-LEVEL SORT, 11), Reducer 4 
(PARTITION-LEVEL SORT, 11)
+Reducer 6 <- Map 15 (PARTITION-LEVEL SORT, 7), Reducer 5 
(PARTITION-LEVEL SORT, 7)
+Reducer 7 <- Reducer 6 (GROUP, 7)
+Reducer 8 <- Reducer 7 (SORT, 1)
  A masked pattern was here 
   Vertices:
 Map 1 
@@ -242,61 +242,32 @@ STAGE PLANS:
   outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6, _col7
   Statistics: Num rows: 14398467 Data size: 1325194184 
Basic stats: COMPLETE Column stats: NONE
   Reduce Output Operator
-key expressions: _col2 (type: int)
-sort order: +
-Map-reduce partition columns: _col2 (type: int)
+key expressions: _col0 (type: int), _col5 (type: int)
+sort order: ++
+Map-reduce partition columns: _col0 (type: int), _col5 
(type: int)
 Statistics: Num rows: 14398467 Data size: 1325194184 
Basic stats: COMPLETE Column stats: NONE
-value expressions: _col0 (type: int), _col1 (type: 
int), _col3 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: 
decimal(7,2)), _col7 (type: decimal(7,2))
+value expressions: _col1 (type: int), _col2 (type: 
int), _col3 (type: int), _col4 (type: int), _col6 (type: decimal(7,2)), _col7 
(type: decimal(7,2))
 Map 10 
 Map Operator Tree:
 TableScan
-  alias: cd2
-  Statistics: Num rows: 1861800 Data size: 717186159 Basic 
stats: COMPLETE Column stats: NONE
+  alias: customer_address
+  Statistics: Num rows: 4000 Data size: 40595195284 Basic 
stats: COMPLETE Column stats: NONE
   Filter Operator
-predicate: (((cd_education_status = '4 yr Degree') or 
(cd_education_status = 'Primary') or (cd_education_status = 'Advanced Degree')) 
and ((cd_marital_status = 'M') or (cd_marital_status = 'D') or 
(cd_marital_status = 'U')) and cd_demo_sk is not null and cd_education_status 
is not null and cd_marital_status is not null) (type: boolean)
-Statistics: Num rows: 1861800 Data size: 717186159 Basic 
stats: COMPLETE Column stats: NONE
+predicate: ((ca_country = 'United States') and (ca_state) 
IN ('KY', 'GA', 'NM', 'MT', 'OR', 'IN', 'WI', 'MO', 'WV') and ca_address_sk is 
not null) (type: boolean)
+Statistics: Num rows: 1000 Data size: 10148798821 
Basic stats: COMPLETE Column stats: NONE
 Select Operator
-  expressions: cd_demo_sk (type: int), cd_marital_status 
(type: string), cd_education_status (type: string)
-  outputColumnNames: _col0, _col

[01/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
Repository: hive
Updated Branches:
  refs/heads/branch-3 91d0a0a2f -> 50f144d1f


http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query92.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query92.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query92.q.out
index 70c4c5a..e7b8632 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query92.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query92.q.out
@@ -67,7 +67,7 @@ STAGE PLANS:
 Spark
  A masked pattern was here 
   Vertices:
-Map 7 
+Map 5 
 Map Operator Tree:
 TableScan
   alias: date_dim
@@ -90,7 +90,7 @@ STAGE PLANS:
 Spark
  A masked pattern was here 
   Vertices:
-Map 10 
+Map 9 
 Map Operator Tree:
 TableScan
   alias: date_dim
@@ -112,11 +112,11 @@ STAGE PLANS:
   Stage: Stage-1
 Spark
   Edges:
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 156), Map 6 (PARTITION-LEVEL 
SORT, 156)
-Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 270), Reducer 9 
(PARTITION-LEVEL SORT, 270)
-Reducer 4 <- Reducer 3 (GROUP, 1)
-Reducer 5 <- Reducer 4 (SORT, 1)
-Reducer 9 <- Map 8 (GROUP, 169)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 261), Reducer 8 
(PARTITION-LEVEL SORT, 261)
+Reducer 3 <- Reducer 2 (GROUP, 1)
+Reducer 4 <- Reducer 3 (SORT, 1)
+Reducer 7 <- Map 6 (GROUP, 169)
+Reducer 8 <- Map 10 (PARTITION-LEVEL SORT, 87), Reducer 7 
(PARTITION-LEVEL SORT, 87)
  A masked pattern was here 
   Vertices:
 Map 1 
@@ -131,13 +131,25 @@ STAGE PLANS:
   expressions: ws_sold_date_sk (type: int), ws_item_sk 
(type: int), ws_ext_discount_amt (type: decimal(7,2))
   outputColumnNames: _col0, _col1, _col2
   Statistics: Num rows: 144002668 Data size: 19580198212 
Basic stats: COMPLETE Column stats: NONE
-  Reduce Output Operator
-key expressions: _col1 (type: int)
-sort order: +
-Map-reduce partition columns: _col1 (type: int)
-Statistics: Num rows: 144002668 Data size: 19580198212 
Basic stats: COMPLETE Column stats: NONE
-value expressions: _col0 (type: int), _col2 (type: 
decimal(7,2))
-Map 6 
+  Map Join Operator
+condition map:
+ Inner Join 0 to 1
+keys:
+  0 _col0 (type: int)
+  1 _col0 (type: int)
+outputColumnNames: _col1, _col2
+input vertices:
+  1 Map 5
+Statistics: Num rows: 158402938 Data size: 21538218500 
Basic stats: COMPLETE Column stats: NONE
+Reduce Output Operator
+  key expressions: _col1 (type: int)
+  sort order: +
+  Map-reduce partition columns: _col1 (type: int)
+  Statistics: Num rows: 158402938 Data size: 
21538218500 Basic stats: COMPLETE Column stats: NONE
+  value expressions: _col2 (type: decimal(7,2))
+Local Work:
+  Map Reduce Local Work
+Map 10 
 Map Operator Tree:
 TableScan
   alias: item
@@ -154,7 +166,7 @@ STAGE PLANS:
 sort order: +
 Map-reduce partition columns: _col0 (type: int)
 Statistics: Num rows: 231000 Data size: 331780228 
Basic stats: COMPLETE Column stats: NONE
-Map 8 
+Map 6 
 Map Operator Tree:
 TableScan
   alias: web_sales
@@ -174,7 +186,7 @@ STAGE PLANS:
   1 _col0 (type: int)
 outputColumnNames: _col1, _col2
 input vertices:
-  1 Map 10
+  1 Map 9
 Statistics: Num rows: 158402938 Data size: 21538218500 
Basic stats: COMPLETE Column stats: NONE
 Group By Operator
   aggregations: sum(_col2), count(_col2)
@@ -191,50 +203,22 @@ STAGE PLANS:
 Local Work:
   Map Reduce Local Work
 Reducer 2 
-Local Work:
-  Map Reduce Local Work
 Reduce Operator Tree:
   Join Operator
 condition map:
  Inner Join 0 to 1
 keys:
   0 _col1 (type: int)
-  1 _col0 (type: int)
-outp

[04/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query78.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query78.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query78.q.out
index c2e8577..1467c5f 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query78.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query78.q.out
@@ -120,41 +120,57 @@ STAGE PLANS:
   Stage: Stage-1
 Spark
   Edges:
-Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 164), Map 14 
(PARTITION-LEVEL SORT, 164)
-Reducer 12 <- Map 15 (PARTITION-LEVEL SORT, 85), Reducer 11 
(PARTITION-LEVEL SORT, 85)
-Reducer 13 <- Reducer 12 (GROUP, 93)
-Reducer 17 <- Map 16 (PARTITION-LEVEL SORT, 85), Reducer 20 
(PARTITION-LEVEL SORT, 85)
+Reducer 11 <- Map 10 (PARTITION-LEVEL SORT, 85), Reducer 14 
(PARTITION-LEVEL SORT, 85)
+Reducer 12 <- Reducer 11 (GROUP, 93)
+Reducer 14 <- Map 13 (PARTITION-LEVEL SORT, 164), Map 15 
(PARTITION-LEVEL SORT, 164)
+Reducer 17 <- Map 10 (PARTITION-LEVEL SORT, 85), Reducer 20 
(PARTITION-LEVEL SORT, 85)
 Reducer 18 <- Reducer 17 (GROUP, 93)
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 432), Map 8 (PARTITION-LEVEL 
SORT, 432)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 219), Reducer 8 
(PARTITION-LEVEL SORT, 219)
 Reducer 20 <- Map 19 (PARTITION-LEVEL SORT, 177), Map 21 
(PARTITION-LEVEL SORT, 177)
-Reducer 3 <- Map 9 (PARTITION-LEVEL SORT, 219), Reducer 2 
(PARTITION-LEVEL SORT, 219)
-Reducer 4 <- Reducer 3 (GROUP, 241)
-Reducer 5 <- Reducer 13 (PARTITION-LEVEL SORT, 167), Reducer 4 
(PARTITION-LEVEL SORT, 167)
-Reducer 6 <- Reducer 18 (PARTITION-LEVEL SORT, 91), Reducer 5 
(PARTITION-LEVEL SORT, 91)
-Reducer 7 <- Reducer 6 (SORT, 1)
+Reducer 3 <- Reducer 2 (GROUP, 241)
+Reducer 4 <- Reducer 12 (PARTITION-LEVEL SORT, 167), Reducer 3 
(PARTITION-LEVEL SORT, 167)
+Reducer 5 <- Reducer 18 (PARTITION-LEVEL SORT, 91), Reducer 4 
(PARTITION-LEVEL SORT, 91)
+Reducer 6 <- Reducer 5 (SORT, 1)
+Reducer 8 <- Map 7 (PARTITION-LEVEL SORT, 432), Map 9 (PARTITION-LEVEL 
SORT, 432)
  A masked pattern was here 
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
-  alias: store_sales
-  Statistics: Num rows: 575995635 Data size: 50814502088 Basic 
stats: COMPLETE Column stats: NONE
+  alias: date_dim
+  Statistics: Num rows: 73049 Data size: 81741831 Basic stats: 
COMPLETE Column stats: NONE
   Filter Operator
-predicate: ss_sold_date_sk is not null (type: boolean)
-Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
+predicate: ((d_year = 2000) and d_date_sk is not null) 
(type: boolean)
+Statistics: Num rows: 36524 Data size: 40870356 Basic 
stats: COMPLETE Column stats: NONE
 Select Operator
-  expressions: ss_sold_date_sk (type: int), ss_item_sk 
(type: int), ss_customer_sk (type: int), ss_ticket_number (type: int), 
ss_quantity (type: int), ss_wholesale_cost (type: decimal(7,2)), ss_sales_price 
(type: decimal(7,2))
-  outputColumnNames: _col0, _col1, _col2, _col3, _col4, 
_col5, _col6
-  Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
+  expressions: d_date_sk (type: int)
+  outputColumnNames: _col0
+  Statistics: Num rows: 36524 Data size: 40870356 Basic 
stats: COMPLETE Column stats: NONE
   Reduce Output Operator
-key expressions: _col1 (type: int), _col3 (type: int)
-sort order: ++
-Map-reduce partition columns: _col1 (type: int), _col3 
(type: int)
-Statistics: Num rows: 575995635 Data size: 50814502088 
Basic stats: COMPLETE Column stats: NONE
-value expressions: _col0 (type: int), _col2 (type: 
int), _col4 (type: int), _col5 (type: decimal(7,2)), _col6 (type: decimal(7,2))
+key expressions: _col0 (type: int)
+sort order: +
+Map-reduce partition columns: _col0 (type: int)
+Statistics: Num rows: 36524 Data size: 40870356 Basic 
stats: COMPLETE Column stats: NONE
 Map 10 
 Map Operator Tree:
 TableScan
+  alias: date_dim
+  Statistics: Num rows: 73049 Data size: 81741831 Basic stats: 
COMPLETE Column stats: NONE
+  Filter Operator
+   

[09/15] hive git commit: HIVE-19147 : Fix PerfCliDrivers: Tpcds30T missed CAT_NAME change (Zoltan Haindrich via Ashutosh Chauhan)

2018-04-12 Thread vgarg
http://git-wip-us.apache.org/repos/asf/hive/blob/12a1784d/ql/src/test/results/clientpositive/perf/spark/query45.q.out
--
diff --git a/ql/src/test/results/clientpositive/perf/spark/query45.q.out 
b/ql/src/test/results/clientpositive/perf/spark/query45.q.out
index 7e1cc88..b674400 100644
--- a/ql/src/test/results/clientpositive/perf/spark/query45.q.out
+++ b/ql/src/test/results/clientpositive/perf/spark/query45.q.out
@@ -46,10 +46,10 @@ STAGE PLANS:
   Stage: Stage-2
 Spark
   Edges:
-Reducer 6 <- Map 5 (GROUP, 1)
+Reducer 16 <- Map 15 (GROUP, 1)
  A masked pattern was here 
   Vertices:
-Map 5 
+Map 15 
 Map Operator Tree:
 TableScan
   alias: item
@@ -70,7 +70,7 @@ STAGE PLANS:
   sort order: 
   Statistics: Num rows: 1 Data size: 16 Basic stats: 
COMPLETE Column stats: NONE
   value expressions: _col0 (type: bigint), _col1 
(type: bigint)
-Reducer 6 
+Reducer 16 
 Local Work:
   Map Reduce Local Work
 Reduce Operator Tree:
@@ -88,45 +88,33 @@ STAGE PLANS:
 Spark
   Edges:
 Reducer 11 <- Map 10 (GROUP, 3)
-Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 154), Map 15 
(PARTITION-LEVEL SORT, 154)
-Reducer 14 <- Map 16 (PARTITION-LEVEL SORT, 706), Reducer 13 
(PARTITION-LEVEL SORT, 706)
-Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 527), Reducer 9 
(PARTITION-LEVEL SORT, 527)
-Reducer 3 <- Reducer 2 (GROUP, 224)
-Reducer 4 <- Reducer 3 (SORT, 1)
+Reducer 13 <- Map 12 (PARTITION-LEVEL SORT, 154), Map 14 
(PARTITION-LEVEL SORT, 154)
+Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 855), Map 6 (PARTITION-LEVEL 
SORT, 855)
+Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 777), Reducer 9 
(PARTITION-LEVEL SORT, 777)
+Reducer 4 <- Reducer 3 (GROUP, 230)
+Reducer 5 <- Reducer 4 (SORT, 1)
 Reducer 8 <- Map 7 (PARTITION-LEVEL SORT, 7), Reducer 11 
(PARTITION-LEVEL SORT, 7)
-Reducer 9 <- Reducer 14 (PARTITION-LEVEL SORT, 191), Reducer 8 
(PARTITION-LEVEL SORT, 191)
+Reducer 9 <- Reducer 13 (PARTITION-LEVEL SORT, 174), Reducer 8 
(PARTITION-LEVEL SORT, 174)
  A masked pattern was here 
   Vertices:
 Map 1 
 Map Operator Tree:
 TableScan
-  alias: customer_address
-  Statistics: Num rows: 4000 Data size: 40595195284 Basic 
stats: COMPLETE Column stats: NONE
+  alias: customer
+  Statistics: Num rows: 8000 Data size: 68801615852 Basic 
stats: COMPLETE Column stats: NONE
   Filter Operator
-predicate: ca_address_sk is not null (type: boolean)
-Statistics: Num rows: 4000 Data size: 40595195284 
Basic stats: COMPLETE Column stats: NONE
+predicate: (c_current_addr_sk is not null and 
c_customer_sk is not null) (type: boolean)
+Statistics: Num rows: 8000 Data size: 68801615852 
Basic stats: COMPLETE Column stats: NONE
 Select Operator
-  expressions: ca_address_sk (type: int), ca_county (type: 
string), ca_zip (type: string)
-  outputColumnNames: _col0, _col1, _col2
-  Statistics: Num rows: 4000 Data size: 40595195284 
Basic stats: COMPLETE Column stats: NONE
-  Map Join Operator
-condition map:
- Inner Join 0 to 1
-keys:
-  0 
-  1 
-outputColumnNames: _col0, _col1, _col2, _col3, _col4
-input vertices:
-  1 Reducer 6
-Statistics: Num rows: 4000 Data size: 41275195284 
Basic stats: COMPLETE Column stats: NONE
-Reduce Output Operator
-  key expressions: _col0 (type: int)
-  sort order: +
-  Map-reduce partition columns: _col0 (type: int)
-  Statistics: Num rows: 4000 Data size: 
41275195284 Basic stats: COMPLETE Column stats: NONE
-  value expressions: _col1 (type: string), _col2 
(type: string), _col3 (type: bigint), _col4 (type: bigint)
-Local Work:
-  Map Reduce Local Work
+  expressions: c_customer_sk (type: int), 
c_current_addr_sk (type: int)
+  outputColumnNames: _col0, _col1
+  Statistics: Num rows: 8000 Data size: 68801615852 
Basic stats: COMPLETE Column stats: NONE
+  Reduce Output Operator
+key expressions: _col1 (

hive git commit: HIVE-19175 : TestMiniLlapLocalCliDriver.testCliDriver update_access_time_non_current_db failing (Vineet Garg via Peter Vary)

2018-04-12 Thread hashutosh
Repository: hive
Updated Branches:
  refs/heads/master a2dd09f86 -> b7c64b15c


HIVE-19175 : TestMiniLlapLocalCliDriver.testCliDriver 
update_access_time_non_current_db failing (Vineet Garg via Peter Vary)

Signed-off-by: Ashutosh Chauhan 


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/b7c64b15
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/b7c64b15
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/b7c64b15

Branch: refs/heads/master
Commit: b7c64b15c28b78c889e44f58f5e61374696ab9a5
Parents: a2dd09f
Author: Vineet Garg 
Authored: Thu Apr 12 09:34:31 2018 -0700
Committer: Ashutosh Chauhan 
Committed: Thu Apr 12 09:34:31 2018 -0700

--
 .../test/resources/testconfiguration.properties |  2 +-
 .../update_access_time_non_current_db.q.out | 47 
 .../update_access_time_non_current_db.q.out | 47 
 3 files changed, 48 insertions(+), 48 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/b7c64b15/itests/src/test/resources/testconfiguration.properties
--
diff --git a/itests/src/test/resources/testconfiguration.properties 
b/itests/src/test/resources/testconfiguration.properties
index 48d62a8..c5d4e9f 100644
--- a/itests/src/test/resources/testconfiguration.properties
+++ b/itests/src/test/resources/testconfiguration.properties
@@ -285,7 +285,6 @@ minillaplocal.shared.query.files=alter_merge_2_orc.q,\
   union8.q,\
   union9.q,\
   union_stats.q,\
-  update_access_time_non_current_db.q, \
   update_after_multiple_inserts.q,\
   update_all_non_partitioned.q,\
   update_all_partitioned.q,\
@@ -739,6 +738,7 @@ minillaplocal.query.files=\
   union_fast_stats.q,\
   union_remove_26.q,\
   union_top_level.q,\
+  update_access_time_non_current_db.q, \
   vector_annotate_stats_select.q,\
   vector_auto_smb_mapjoin_14.q,\
   vector_char_varchar_1.q,\

http://git-wip-us.apache.org/repos/asf/hive/blob/b7c64b15/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out
 
b/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out
new file mode 100644
index 000..7fda3ac
--- /dev/null
+++ 
b/ql/src/test/results/clientpositive/llap/update_access_time_non_current_db.q.out
@@ -0,0 +1,47 @@
+PREHOOK: query: create database temp1
+PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:temp1
+POSTHOOK: query: create database temp1
+POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:temp1
+PREHOOK: query: use temp1
+PREHOOK: type: SWITCHDATABASE
+PREHOOK: Input: database:temp1
+POSTHOOK: query: use temp1
+POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:temp1
+PREHOOK: query: create table test1(id int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:temp1
+PREHOOK: Output: temp1@test1
+POSTHOOK: query: create table test1(id int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:temp1
+POSTHOOK: Output: temp1@test1
+PREHOOK: query: create database temp2
+PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:temp2
+POSTHOOK: query: create database temp2
+POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:temp2
+PREHOOK: query: use temp2
+PREHOOK: type: SWITCHDATABASE
+PREHOOK: Input: database:temp2
+POSTHOOK: query: use temp2
+POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:temp2
+PREHOOK: query: create table test2(id int)
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:temp2
+PREHOOK: Output: temp2@test2
+POSTHOOK: query: create table test2(id int)
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:temp2
+POSTHOOK: Output: temp2@test2
+POSTHOOK: query: use temp1
+POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:temp1
+POSTHOOK: query: desc temp2.test2
+POSTHOOK: type: DESCTABLE
+POSTHOOK: Input: temp2@test2
+id int 

http://git-wip-us.apache.org/repos/asf/hive/blob/b7c64b15/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out
--
diff --git 
a/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out 
b/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out
deleted file mode 100644
index 7fda3ac..000
--- a/ql/src/test/results/clientpositive/update_access_time_non_current_db.q.out
+++ /dev/null
@@ -1,47 +0,0 @@
-PREHOOK: query: create database temp1
-PREHOOK: type: CREATEDATABASE
-PREHOOK: Output: database:temp1
-POSTHOOK: query: create database temp1
-POSTHOOK: type: CREATEDATABASE
-POSTHOOK: Output: database:temp1
-PREHOOK: 

hive git commit: HIVE-19156 : TestMiniLlapLocalCliDriver.vectorized_dynamic_semijoin_reduction.q is broken (Jason Dere via Deepak Jaiswal)

2018-04-12 Thread hashutosh
Repository: hive
Updated Branches:
  refs/heads/master 2e027cff7 -> a2dd09f86


HIVE-19156 : TestMiniLlapLocalCliDriver.vectorized_dynamic_semijoin_reduction.q 
is broken (Jason Dere via Deepak Jaiswal)

Signed-off-by: Ashutosh Chauhan 


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/a2dd09f8
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/a2dd09f8
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/a2dd09f8

Branch: refs/heads/master
Commit: a2dd09f86386ca3f51eca90948d6ec1963ece982
Parents: 2e027cf
Author: Jason Dere 
Authored: Thu Apr 12 08:47:25 2018 -0700
Committer: Ashutosh Chauhan 
Committed: Thu Apr 12 08:47:25 2018 -0700

--
 .../ql/exec/tez/DynamicValueRegistryTez.java | 19 +--
 1 file changed, 17 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hive/blob/a2dd09f8/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java
--
diff --git 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java
index ec1e84b..2d99f50 100644
--- 
a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java
+++ 
b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/DynamicValueRegistryTez.java
@@ -61,6 +61,11 @@ public class DynamicValueRegistryTez implements 
DynamicValueRegistry {
 }
   }
 
+  static class NullValue {
+  }
+
+  static final NullValue NULL_VALUE = new NullValue();
+
   protected Map values = new ConcurrentHashMap<>();
 
   public DynamicValueRegistryTez() {
@@ -71,11 +76,21 @@ public class DynamicValueRegistryTez implements 
DynamicValueRegistry {
 if (!values.containsKey(key)) {
   throw new NoDynamicValuesException("Value does not exist in registry: " 
+ key);
 }
-return values.get(key);
+Object val = values.get(key);
+
+if (val == NULL_VALUE) {
+  return null;
+}
+return val;
   }
 
   protected void setValue(String key, Object value) {
-values.put(key, value);
+if (value == null) {
+  // ConcurrentHashMap does not allow null - use a substitute value.
+  values.put(key, NULL_VALUE);
+} else {
+  values.put(key, value);
+}
   }
 
   @Override