[jira] [Commented] (HIVE-19950) Hive ACID NOT LOCK LockComponent Correctly

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16520052#comment-16520052
 ] 

Hive QA commented on HIVE-19950:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
54s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
39s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
23s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m  
3s{color} | {color:blue} standalone-metastore in master has 227 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
41s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
20s{color} | {color:red} standalone-metastore: The patch generated 4 new + 531 
unchanged - 11 fixed = 535 total (was 542) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  3m 
16s{color} | {color:red} standalone-metastore generated 1 new + 227 unchanged - 
0 fixed = 228 total (was 227) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
58s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 19m 36s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:standalone-metastore |
|  |  Private method 
org.apache.hadoop.hive.metastore.txn.TxnHandler.shouldNeverHappen(long, long, 
long) is never called  At TxnHandler.java:is never called  At 
TxnHandler.java:[lines 2497-2498] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-11989/dev-support/hive-personality.sh
 |
| git revision | master / 24e16cc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11989/yetus/diff-checkstyle-standalone-metastore.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11989/yetus/new-findbugs-standalone-metastore.html
 |
| modules | C: standalone-metastore U: standalone-metastore |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11989/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Hive ACID NOT LOCK LockComponent Correctly
> --
>
> Key: HIVE-19950
> URL: https://issues.apache.org/jira/browse/HIVE-19950
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.3.2
>Reporter: nickSoul
>Priority: Blocker
> Attachments: patch.patch
>
>
> Hi,
>     When using Streaming Mutation recently, I found LockComponents were not 
> locked correctly by current transaction.  Below is my test case:
> Step1: Begin a transaction with transactionId 126, and the transaction locks 
> a table. Then hangs the transaction. The lock information were correctly 
> restored in mariaDB
> {code:java}
> MariaDB [hive]> select 
> HL_LOCK_EXT_ID,HL_LOCK_INT_ID,HL_TXNID,HL_DB,HL_TABLE,HL_PARTITION,HL_LOCK_STATE,HL_LOCK_TYPE,HL_ACQUIRED_AT,HL_BLOCKEDBY_EXT_ID,HL_BLOCKEDBY_INT_ID
>  from HIVE_LOCKS;
> +++--+---

[jira] [Commented] (HIVE-19956) Include yarn registry classes to jdbc standalone jar

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16520046#comment-16520046
 ] 

Hive QA commented on HIVE-19956:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928593/HIVE-19956.1.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 1 failed/errored test(s), 14589 tests 
executed
*Failed tests:*
{noformat}
org.apache.hive.jdbc.TestJdbcWithMiniHS2.testParallelCompilation3 (batchId=246)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11988/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11988/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11988/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 1 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928593 - PreCommit-HIVE-Build

> Include yarn registry classes to jdbc standalone jar
> 
>
> Key: HIVE-19956
> URL: https://issues.apache.org/jira/browse/HIVE-19956
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19956.1.patch
>
>
> HS2 Active/Passive HA requires some yarn registry classes. Include it in JDBC 
> standalone jar. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HIVE-19867) Test and verify Concurrent INSERTS

2018-06-21 Thread Steve Yeom (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16516053#comment-16516053
 ] 

Steve Yeom edited comment on HIVE-19867 at 6/22/18 5:40 AM:


I found the patch 03 does not address the cases of concurrent writes to a 
table. 
If writeId for transactional stats is saved into TBLS/PARTITIONS, then 
it can be used to figure out concurrent writes by comparing a write with its 
writeIdList.
Simply by checking it is in the list or not. 

I.e., if we assume two concurrent NSERTs are subsequently committed and updated 
stats 
of, let's say, table1. Then first INSERT's writeId is now saved and so can be 
used to see if it is
in second INSERT's writeIdList when second INSERT comes in. Also second 
INSERT's writeId
can be checked to see it is in first INSERT's writeIdList that is saved in 
TBLS/PARTITIONS.
If each writeId is contained in the other's writeIdList, we can say it is 
concurrent INSERT case
and turn the COLUMN_STATS_ACCURATE flag off.


was (Author: steveyeom2017):
I found the patch 03 does not address the cases of concurrent writes to a 
table. 
If writeId for transactional stats is saved into TBLS/PARTITIONS, then 
it can be used to figure out concurrent writes by comparing a write with its 
writeIdList.
Simply by checking it is in the list or not. 

> Test and verify Concurrent INSERTS  
> 
>
> Key: HIVE-19867
> URL: https://issues.apache.org/jira/browse/HIVE-19867
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Affects Versions: 4.0.0
>Reporter: Steve Yeom
>Assignee: Steve Yeom
>Priority: Major
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HIVE-19867) Test and verify Concurrent INSERTS

2018-06-21 Thread Steve Yeom (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519787#comment-16519787
 ] 

Steve Yeom edited comment on HIVE-19867 at 6/22/18 5:33 AM:


Sergey and I talked about this. He mentioned several cases from the perspective 
of Readers.

If I summarize cases of both Stats updaters (esp., concurrent INSERTs) and 
readers, then,
let's assume the txnId of a stats entity(table or partition) is 1 and its write 
id on table1 is 11.
Then we may have
case1: Suppose concurrent writes 12 (with txnId 2) and 13 (txnId 3) and 
concurrent reader 14 (txnId 4). Each write has 12, 13 in its writeIdList.
  Here concurrent reader 14 has 12,13 as open writes in its writeIdList.
  1) Write 12 comes and updates the stats of table1 and its transaction is 
committed. 
So now txnid of the stats is 2.
  2) Then write 13 comes in and, if we have the code for this jira, checks 
itself (number 13) 
from stats's writeIdList (written by txnId 2). 13 should be there in the 
writeIdList. 
Also, if we save 12 into TXNS/PARTITIONS, we can check whether 12 is in 
write 13's writeIdList. This detects the case of concurrent writes and can 
turn the flag
(COLUMN_STATS_ACCURATE) off.
  3) Now reader comes in and finds the stats is not valid by simply checking 
the flag.
(the reader also can determine the stats' invalidity by comparing 
writeIdLists of itself 
and the stats. But the next case2 shows the reason why we need to turn the 
flag off.)
case2: Suppose concurrent writes 12 and 13. But assume we have a reader 14 
(txnId 4) that 
  started its transaction after writes 12 and 13 are committed. 
  If the flag is still on and the txnId in TBLS/PARTITIONS is 3, then reader 14 
does not have
  a way to figure out the stats are invalid due to concurrent writes since its 
own writeIdList for
  table1 does not have 12, 13 as open writes and both are committed.


was (Author: steveyeom2017):
Sergey and I talked about this. He mentioned several cases from the perspective 
of Readers like:

Let's assume the txnId of a stats entity(table or partition) is 1 and its write 
id on table1 is 11.
Then we may have
case1: Suppose concurrent writes 12 (with txnId 2) and 13 (txnId 3) and 
concurrent reader 14 (txnId 4).
  Here concurrent reader 14 has 12,13 as open writes in its writeIdList.
  1) Write 12 comes and updates the stats of table1 and its transaction is 
committed. 
So now txnid of the stats is txnid is 2.
  2) Then write 13 comes in and checks itself (number 13) from stats's 
writeIdList by txnId 2. 
13 should be there in the writeIdList. So it detects concurrent writes and 
can turn the flag
(COLUMN_STATS_ACCURATE) off.
  3) Now reader comes in and finds the stats is not valid by simply checking 
the flag.
(the reader also can determine the stats' validity by comparing 
writeIdLists of itself and the stats)
case2: Suppose concurrent writes 12 and 13. But assume we have a reader 14 
(txnId 4) that started its 
  transaction after writes 12 and 13 are done. 
  If the flag is still on and the txnId in TBLS/PARTITIONS is 3, then reader 14 
does not have  a way 
  to figure out the stats are invalid due to concurrent writes since its own 
writeIdList for table1 does not have
  12, 13 as open writes and both are committed.

> Test and verify Concurrent INSERTS  
> 
>
> Key: HIVE-19867
> URL: https://issues.apache.org/jira/browse/HIVE-19867
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Affects Versions: 4.0.0
>Reporter: Steve Yeom
>Assignee: Steve Yeom
>Priority: Major
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19967) SMB Join : Need Optraits for PTFOperator ala GBY Op

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19967:
--
Description: 
The SMB join on one or more PTF Ops should reset the optraits keys just like 
GBY Op does.

Currently there is no implementation of PTFOp optraits.

  was:
The optraits for ReduceSinkOp used to use the key columns as bucket and sort 
columns which worked fine for SMB, however, to enable prefix in Bucket Map 
Join, this logic was updated to use the bucket columns from parent operators. 
However, this may break reduce side SMB in a scenario like this,

 

Task1 (TS bucketed by col0), passes it down to RS which ignores the key columns 
and uses col0 as bucket key.

Task2 (Set of ops work such that data is sorted by a set of columns), however, 
with current logic, the bucketing column set in Task1 keeps getting pushed in 
Optraits, thus losing the real flow.

Task3(Join op) The physical optimizer looks at the parent RS ops which 
incidentally are sorted by same column as the original Task1's bucket column, 
however, in the meantime lost the meaning.

 

 

Summary: SMB Join : Need Optraits for PTFOperator ala GBY Op  (was: SMB 
Join : ReduceSink should use correct keys in optraits)

> SMB Join : Need Optraits for PTFOperator ala GBY Op
> ---
>
> Key: HIVE-19967
> URL: https://issues.apache.org/jira/browse/HIVE-19967
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
>
> The SMB join on one or more PTF Ops should reset the optraits keys just like 
> GBY Op does.
> Currently there is no implementation of PTFOp optraits.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19956) Include yarn registry classes to jdbc standalone jar

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16520024#comment-16520024
 ] 

Hive QA commented on HIVE-19956:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
 1s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
19s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
12s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
41s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch 1 line(s) with tabs. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
12s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 10m 14s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  xml  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-11988/dev-support/hive-personality.sh
 |
| git revision | master / 24e16cc |
| Default Java | 1.8.0_111 |
| whitespace | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11988/yetus/whitespace-tabs.txt
 |
| modules | C: jdbc U: jdbc |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11988/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Include yarn registry classes to jdbc standalone jar
> 
>
> Key: HIVE-19956
> URL: https://issues.apache.org/jira/browse/HIVE-19956
> Project: Hive
>  Issue Type: Bug
>  Components: JDBC
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Prasanth Jayachandran
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19956.1.patch
>
>
> HS2 Active/Passive HA requires some yarn registry classes. Include it in JDBC 
> standalone jar. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19783) Retrieve only locations in HiveMetaStore.dropPartitionsAndGetLocations

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16520019#comment-16520019
 ] 

Hive QA commented on HIVE-19783:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928592/HIVE-19783.5.patch

{color:green}SUCCESS:{color} +1 due to 2 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14589 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11987/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11987/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11987/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928592 - PreCommit-HIVE-Build

> Retrieve only locations in HiveMetaStore.dropPartitionsAndGetLocations
> --
>
> Key: HIVE-19783
> URL: https://issues.apache.org/jira/browse/HIVE-19783
> Project: Hive
>  Issue Type: Improvement
>Reporter: Peter Vary
>Assignee: Peter Vary
>Priority: Major
> Attachments: HIVE-19783.2.patch, HIVE-19783.4.patch, 
> HIVE-19783.5.patch, HIVE-19783.patch
>
>
> Optimize further the dropTable command.
> Currently {{HiveMetaStore.dropPartitionsAndGetLocations}} retrieves the whole 
> partition object, but we need only the locations instead.
> Create a RawStore method to retrieve only the locations.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18725) Improve error handling for subqueries if there is wrong column reference

2018-06-21 Thread Igor Kryvenko (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18725?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16520011#comment-16520011
 ] 

Igor Kryvenko commented on HIVE-18725:
--

[~vgarg] [~ashutoshc] Hi guys. Could review, please? 

> Improve error handling for subqueries if there is wrong column reference
> 
>
> Key: HIVE-18725
> URL: https://issues.apache.org/jira/browse/HIVE-18725
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Igor Kryvenko
>Priority: Major
> Attachments: HIVE-18725.01.patch, HIVE-18725.02.patch, 
> HIVE-18725.03.patch, HIVE-18725.04.patch, HIVE-18725.05.patch, 
> HIVE-18725.06.patch, HIVE-18725.07.patch, HIVE-18725.08.patch
>
>
> If there is a column reference within subquery which doesn't exist Hive 
> throws misleading error message.
> e.g. 
> {code:sql}
> select * from table1 where table1.col1 IN (select col2 from table2 where 
> table2.col1=table1.non_existing_column) and table1.col1 IN (select 4);
> {code}
> The above query, assuming table1 doesn't have non_existing_column, will throw 
> following misleading error:
> {noformat}
> FAILED: SemanticException Line 0:-1 Unsupported SubQuery Expression 'col1': 
> Only 1 SubQuery expression is supported.
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19966) Compile error when using open-jdk

2018-06-21 Thread Vihang Karajgaonkar (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19966?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16520008#comment-16520008
 ] 

Vihang Karajgaonkar commented on HIVE-19966:


It errors out for me. I was playing around using dockers containers and tried 
to build hive code and it errors out.

> Compile error when using open-jdk
> -
>
> Key: HIVE-19966
> URL: https://issues.apache.org/jira/browse/HIVE-19966
> Project: Hive
>  Issue Type: Bug
>Reporter: Vihang Karajgaonkar
>Priority: Major
>
> When you compile Hive using open-jdk-8 you see this error
> {noformat}
> hive/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/EncodedReaderImpl.java:[72,16]
>  sun.misc.Cleaner is internal proprietary API and may be removed in a future 
> release
> [ERROR] -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR]   mvn  -rf :hive-exec
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19481) sample10.q returns wrong results

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19481?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19481:
--
Status: Patch Available  (was: In Progress)

> sample10.q returns wrong results 
> -
>
> Key: HIVE-19481
> URL: https://issues.apache.org/jira/browse/HIVE-19481
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Steve Yeom
>Assignee: Deepak Jaiswal
>Priority: Major
> Fix For: 3.1.0
>
>
> Ran "mvn test -Dtest=TestMiniLlapLocalCliDriver -Dqfile=sample10.q " after 
> changing the table to be 
> insert-only transactional. 
> The following queries returns couple of rows whereis no row results returns 
> for non-ACID table. 
> query: select ds, count(1) from srcpartbucket tablesample (bucket 2 out of 4 
> on key) where ds is not null group by ds ORDER BY ds ASC
> 2008-04-08  14
> 2008-04-09  14
> ..
> query: select ds, count(1) from srcpartbucket tablesample (bucket 1 out of 2 
> on key) where ds is not null group by ds ORDER BY ds ASC
> 2008-04-08  4
> 2008-04-09  4



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19966) Compile error when using open-jdk

2018-06-21 Thread Prasanth Jayachandran (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19966?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16520004#comment-16520004
 ] 

Prasanth Jayachandran commented on HIVE-19966:
--

[~vihangk1] Is that warning or error? open-jdk should have sun.misc.cleaner

> Compile error when using open-jdk
> -
>
> Key: HIVE-19966
> URL: https://issues.apache.org/jira/browse/HIVE-19966
> Project: Hive
>  Issue Type: Bug
>Reporter: Vihang Karajgaonkar
>Priority: Major
>
> When you compile Hive using open-jdk-8 you see this error
> {noformat}
> hive/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/EncodedReaderImpl.java:[72,16]
>  sun.misc.Cleaner is internal proprietary API and may be removed in a future 
> release
> [ERROR] -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR]   mvn  -rf :hive-exec
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19783) Retrieve only locations in HiveMetaStore.dropPartitionsAndGetLocations

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1652#comment-1652
 ] 

Hive QA commented on HIVE-19783:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
45s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
15s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
36s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
12s{color} | {color:blue} standalone-metastore in master has 227 extant 
Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
7s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
9s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  0m 
21s{color} | {color:red} hcatalog-unit in the patch failed. {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
2s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
11s{color} | {color:green} The patch hcatalog-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
27s{color} | {color:green} standalone-metastore: The patch generated 0 new + 
1020 unchanged - 1 fixed = 1020 total (was 1021) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  3m 
20s{color} | {color:red} standalone-metastore generated 1 new + 227 unchanged - 
0 fixed = 228 total (was 227) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
8s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 22m  6s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:standalone-metastore |
|  |  
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.dropPartitionsAndGetLocations(RawStore,
 String, String, String, Path, boolean) makes inefficient use of keySet 
iterator instead of entrySet iterator  At HiveMetaStore.java:boolean) makes 
inefficient use of keySet iterator instead of entrySet iterator  At 
HiveMetaStore.java:[line 2596] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-11987/dev-support/hive-personality.sh
 |
| git revision | master / 24e16cc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| mvninstall | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11987/yetus/patch-mvninstall-itests_hcatalog-unit.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11987/yetus/new-findbugs-standalone-metastore.html
 |
| modules | C: itests/hcatalog-unit standalone-metastore U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11987/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Retrieve only locations in HiveMetaStore.dropPartitionsAndGetLocations
> --
>
> Key: HIVE-19783
> URL: https://issues.apache.org/jira/browse/HIVE-19783
> Project: Hive
>  Issue Type: Improvement
>Reporter: Peter Vary
>Assignee: Peter V

[jira] [Updated] (HIVE-19967) SMB Join : ReduceSink should use correct keys in optraits

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19967:
--
Status: Patch Available  (was: In Progress)

> SMB Join : ReduceSink should use correct keys in optraits
> -
>
> Key: HIVE-19967
> URL: https://issues.apache.org/jira/browse/HIVE-19967
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
>
> The optraits for ReduceSinkOp used to use the key columns as bucket and sort 
> columns which worked fine for SMB, however, to enable prefix in Bucket Map 
> Join, this logic was updated to use the bucket columns from parent operators. 
> However, this may break reduce side SMB in a scenario like this,
>  
> Task1 (TS bucketed by col0), passes it down to RS which ignores the key 
> columns and uses col0 as bucket key.
> Task2 (Set of ops work such that data is sorted by a set of columns), 
> however, with current logic, the bucketing column set in Task1 keeps getting 
> pushed in Optraits, thus losing the real flow.
> Task3(Join op) The physical optimizer looks at the parent RS ops which 
> incidentally are sorted by same column as the original Task1's bucket column, 
> however, in the meantime lost the meaning.
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19967) SMB Join : ReduceSink should use correct keys in optraits

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal reassigned HIVE-19967:
-


> SMB Join : ReduceSink should use correct keys in optraits
> -
>
> Key: HIVE-19967
> URL: https://issues.apache.org/jira/browse/HIVE-19967
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
>
> The optraits for ReduceSinkOp used to use the key columns as bucket and sort 
> columns which worked fine for SMB, however, to enable prefix in Bucket Map 
> Join, this logic was updated to use the bucket columns from parent operators. 
> However, this may break reduce side SMB in a scenario like this,
>  
> Task1 (TS bucketed by col0), passes it down to RS which ignores the key 
> columns and uses col0 as bucket key.
> Task2 (Set of ops work such that data is sorted by a set of columns), 
> however, with current logic, the bucketing column set in Task1 keeps getting 
> pushed in Optraits, thus losing the real flow.
> Task3(Join op) The physical optimizer looks at the parent RS ops which 
> incidentally are sorted by same column as the original Task1's bucket column, 
> however, in the meantime lost the meaning.
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Work started] (HIVE-19967) SMB Join : ReduceSink should use correct keys in optraits

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-19967 started by Deepak Jaiswal.
-
> SMB Join : ReduceSink should use correct keys in optraits
> -
>
> Key: HIVE-19967
> URL: https://issues.apache.org/jira/browse/HIVE-19967
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
>
> The optraits for ReduceSinkOp used to use the key columns as bucket and sort 
> columns which worked fine for SMB, however, to enable prefix in Bucket Map 
> Join, this logic was updated to use the bucket columns from parent operators. 
> However, this may break reduce side SMB in a scenario like this,
>  
> Task1 (TS bucketed by col0), passes it down to RS which ignores the key 
> columns and uses col0 as bucket key.
> Task2 (Set of ops work such that data is sorted by a set of columns), 
> however, with current logic, the bucketing column set in Task1 keeps getting 
> pushed in Optraits, thus losing the real flow.
> Task3(Join op) The physical optimizer looks at the parent RS ops which 
> incidentally are sorted by same column as the original Task1's bucket column, 
> however, in the meantime lost the meaning.
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19966) Compile error when using open-jdk

2018-06-21 Thread Vihang Karajgaonkar (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19966?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519996#comment-16519996
 ] 

Vihang Karajgaonkar commented on HIVE-19966:


[~prasanth_j] [~sershe] Any idea how to fix this? Does this mean we can build 
Hive only using Oracle jdk?

> Compile error when using open-jdk
> -
>
> Key: HIVE-19966
> URL: https://issues.apache.org/jira/browse/HIVE-19966
> Project: Hive
>  Issue Type: Bug
>Reporter: Vihang Karajgaonkar
>Priority: Major
>
> When you compile Hive using open-jdk-8 you see this error
> {noformat}
> hive/ql/src/java/org/apache/hadoop/hive/ql/io/orc/encoded/EncodedReaderImpl.java:[72,16]
>  sun.misc.Cleaner is internal proprietary API and may be removed in a future 
> release
> [ERROR] -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR]   mvn  -rf :hive-exec
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-12342) Set default value of hive.optimize.index.filter to true

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-12342?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519983#comment-16519983
 ] 

Hive QA commented on HIVE-12342:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928590/HIVE-12342.19.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 18 failed/errored test(s), 14589 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[cbo_ppd_non_deterministic]
 (batchId=74)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[partition_wise_fileformat2]
 (batchId=90)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_udf_col] (batchId=34)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[smb_mapjoin_20] 
(batchId=76)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[union33] (batchId=26)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[union] (batchId=4)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[union_offcbo] 
(batchId=47)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[unionDistinct_1] 
(batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_opt_vectorization]
 (batchId=166)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_optimization]
 (batchId=167)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[bucket_mapjoin_mismatch1]
 (batchId=95)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[sortmerge_mapjoin_mismatch_1]
 (batchId=95)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_sortmerge_join_16]
 (batchId=131)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[bucketmapjoin7] 
(batchId=120)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[smb_mapjoin_20] 
(batchId=140)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[union33] 
(batchId=118)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[union] (batchId=107)
org.apache.hadoop.hive.ql.TestTxnCommandsWithSplitUpdateAndVectorization.testSetClauseFakeColumn
 (batchId=290)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11986/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11986/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11986/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 18 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928590 - PreCommit-HIVE-Build

> Set default value of hive.optimize.index.filter to true
> ---
>
> Key: HIVE-12342
> URL: https://issues.apache.org/jira/browse/HIVE-12342
> Project: Hive
>  Issue Type: Task
>  Components: Configuration
>Reporter: Ashutosh Chauhan
>Assignee: Igor Kryvenko
>Priority: Major
> Attachments: HIVE-12342.05.patch, HIVE-12342.06.patch, 
> HIVE-12342.07.patch, HIVE-12342.08.patch, HIVE-12342.09.patch, 
> HIVE-12342.1.patch, HIVE-12342.10.patch, HIVE-12342.11.patch, 
> HIVE-12342.12.patch, HIVE-12342.13.patch, HIVE-12342.14.patch, 
> HIVE-12342.15.patch, HIVE-12342.16.patch, HIVE-12342.17.patch, 
> HIVE-12342.18.patch, HIVE-12342.19.patch, HIVE-12342.2.patch, 
> HIVE-12342.3.patch, HIVE-12342.4.patch, HIVE-12342.patch
>
>
> This configuration governs ppd for storage layer. When applicable, it will 
> always help. It should be on by default.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-12342) Set default value of hive.optimize.index.filter to true

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-12342?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519975#comment-16519975
 ] 

Hive QA commented on HIVE-12342:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
11s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
53s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  3m  
0s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
58s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
39s{color} | {color:blue} common in master has 62 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
37s{color} | {color:blue} accumulo-handler in master has 21 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
27s{color} | {color:blue} contrib in master has 13 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
46s{color} | {color:blue} hbase-handler in master has 15 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  5m 
38s{color} | {color:blue} ql in master has 2282 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m 
54s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
11s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  4m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  3m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  3m 
20s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
58s{color} | {color:red} ql: The patch generated 1 new + 30 unchanged - 0 fixed 
= 31 total (was 30) {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 4 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch 2 line(s) with tabs. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  9m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m 
41s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} asflicense {color} | {color:red}  0m 
19s{color} | {color:red} The patch generated 2 ASF License warnings. {color} |
| {color:black}{color} | {color:black} {color} | {color:black} 58m 38s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-11986/dev-support/hive-personality.sh
 |
| git revision | master / 24e16cc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11986/yetus/diff-checkstyle-ql.txt
 |
| whitespace | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11986/yetus/whitespace-eol.txt
 |
| whitespace | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11986/yetus/whitespace-tabs.txt
 |
| asflicense | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11986/yetus/patch-asflicense-problems.txt
 |
| modules | C: common accumulo-handler contrib hbase-handler ql U: . |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11986/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Set default value of hive.optim

[jira] [Commented] (HIVE-12192) Hive should carry out timestamp computations in UTC

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-12192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519977#comment-16519977
 ] 

Hive QA commented on HIVE-12192:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
21s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
17s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 11m 
13s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 11m 
23s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
25s{color} | {color:blue} storage-api in master has 48 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
33s{color} | {color:blue} common in master has 62 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
38s{color} | {color:blue} serde in master has 190 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
29s{color} | {color:blue} accumulo-handler in master has 21 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
25s{color} | {color:blue} druid-handler in master has 13 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
31s{color} | {color:blue} hbase-handler in master has 15 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
37s{color} | {color:blue} hcatalog/core in master has 33 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
26s{color} | {color:blue} hcatalog/hcatalog-pig-adapter in master has 2 extant 
Findbugs warnings. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  4m  
5s{color} | {color:red} branch/itests/hive-jmh cannot run convertXmlToText from 
findbugs {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
37s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
30s{color} | {color:blue} llap-common in master has 71 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
40s{color} | {color:blue} llap-server in master has 83 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
8s{color} | {color:blue} ql in master has 2282 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m  
7s{color} | {color:blue} standalone-metastore in master has 227 extant Findbugs 
warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 10m 
37s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
7s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
12s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  0m 
36s{color} | {color:red} core in the patch failed. {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  0m 
48s{color} | {color:red} hive-unit in the patch failed. {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  0m 
19s{color} | {color:red} llap-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} compile {color} | {color:red}  0m 
41s{color} | {color:red} ql in the patch failed. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  0m 36s{color} 
| {color:red} core in the patch failed. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  0m 48s{color} 
| {color:red} hive-unit in the patch failed. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  0m 19s{color} 
| {color:red} llap-common in the patch failed. {color} |
| {color:red}-1{color} | {color:red} javac {color} | {color:red}  0m 41s{color

[jira] [Updated] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19928:
--
Attachment: HIVE-19928.02-branch-3.patch

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Fix For: 3.1.0, 4.0.0
>
> Attachments: HIVE-19928.01-branch-3.patch, 
> HIVE-19928.02-branch-3.patch, HIVE-19928.1.patch, HIVE-19928.2.patch, 
> HIVE-19928.3.patch, HIVE-19928.4.patch, HIVE-19928.5.patch, 
> HIVE-19928.6.patch, HIVE-19928.7.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19027) Make materializations invalidation cache work with multiple active remote metastores

2018-06-21 Thread Jesus Camacho Rodriguez (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-19027:
---
Attachment: HIVE-19027.01.patch

> Make materializations invalidation cache work with multiple active remote 
> metastores
> 
>
> Key: HIVE-19027
> URL: https://issues.apache.org/jira/browse/HIVE-19027
> Project: Hive
>  Issue Type: Improvement
>  Components: Materialized views
>Affects Versions: 3.0.0
>Reporter: Jesus Camacho Rodriguez
>Assignee: Jesus Camacho Rodriguez
>Priority: Critical
> Attachments: HIVE-19027.01.patch, HIVE-19027.01.patch
>
>
> The main points:
>  - Only MVs stored in transactional tables can have a time window value of 0. 
> Those are the only MVs that can be guaranteed to not be outdated when a query 
> is executed, if we use custom storage handlers to store the materialized 
> view, we cannot make any promises.
>  - For MVs that +cannot be outdated+, we do not check the metastore. Instead, 
> comparison is based on valid write id lists.
>  - For MVs that +can be outdated+, we still rely on the invalidation cache.
>  ** The window for valid outdated MVs can be specified in intervals of 1 
> minute (less than that, it is difficult to have any guarantees about whether 
> the MV is actually outdated by less than a minute or not).
>  ** The async loading is done every interval / 2 (or probably better, we can 
> make it configurable).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-12192) Hive should carry out timestamp computations in UTC

2018-06-21 Thread Jesus Camacho Rodriguez (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-12192?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-12192:
---
Attachment: HIVE-12192.20.patch

> Hive should carry out timestamp computations in UTC
> ---
>
> Key: HIVE-12192
> URL: https://issues.apache.org/jira/browse/HIVE-12192
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Reporter: Ryan Blue
>Assignee: Jesus Camacho Rodriguez
>Priority: Blocker
>  Labels: timestamp
> Attachments: HIVE-12192.01.patch, HIVE-12192.02.patch, 
> HIVE-12192.03.patch, HIVE-12192.04.patch, HIVE-12192.05.patch, 
> HIVE-12192.06.patch, HIVE-12192.07.patch, HIVE-12192.08.patch, 
> HIVE-12192.09.patch, HIVE-12192.10.patch, HIVE-12192.11.patch, 
> HIVE-12192.12.patch, HIVE-12192.13.patch, HIVE-12192.14.patch, 
> HIVE-12192.15.patch, HIVE-12192.16.patch, HIVE-12192.17.patch, 
> HIVE-12192.18.patch, HIVE-12192.19.patch, HIVE-12192.20.patch, 
> HIVE-12192.patch
>
>
> Hive currently uses the "local" time of a java.sql.Timestamp to represent the 
> SQL data type TIMESTAMP WITHOUT TIME ZONE. The purpose is to be able to use 
> {{Timestamp#getYear()}} and similar methods to implement SQL functions like 
> {{year}}.
> When the SQL session's time zone is a DST zone, such as America/Los_Angeles 
> that alternates between PST and PDT, there are times that cannot be 
> represented because the effective zone skips them.
> {code}
> hive> select TIMESTAMP '2015-03-08 02:10:00.101';
> 2015-03-08 03:10:00.101
> {code}
> Using UTC instead of the SQL session time zone as the underlying zone for a 
> java.sql.Timestamp avoids this bug, while still returning correct values for 
> {{getYear}} etc. Using UTC as the convenience representation (timestamp 
> without time zone has no real zone) would make timestamp calculations more 
> consistent and avoid similar problems in the future.
> Notably, this would break the {{unix_timestamp}} UDF that specifies the 
> result is with respect to ["the default timezone and default 
> locale"|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+UDF#LanguageManualUDF-DateFunctions].
>  That function would need to be updated to use the 
> {{System.getProperty("user.timezone")}} zone.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-16391) Publish proper Hive 1.2 jars (without including all dependencies in uber jar)

2018-06-21 Thread Saisai Shao (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-16391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519931#comment-16519931
 ] 

Saisai Shao commented on HIVE-16391:


Gently ping [~hagleitn], would you please help to review the current proposed 
patch and suggest the next step. Thanks a lot.

> Publish proper Hive 1.2 jars (without including all dependencies in uber jar)
> -
>
> Key: HIVE-16391
> URL: https://issues.apache.org/jira/browse/HIVE-16391
> Project: Hive
>  Issue Type: Task
>  Components: Build Infrastructure
>Affects Versions: 1.2.2
>Reporter: Reynold Xin
>Assignee: Saisai Shao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.2.3
>
> Attachments: HIVE-16391.1.patch, HIVE-16391.2.patch, HIVE-16391.patch
>
>
> Apache Spark currently depends on a forked version of Apache Hive. AFAIK, the 
> only change in the fork is to work around the issue that Hive publishes only 
> two sets of jars: one set with no dependency declared, and another with all 
> the dependencies included in the published uber jar. That is to say, Hive 
> doesn't publish a set of jars with the proper dependencies declared.
> There is general consensus on both sides that we should remove the forked 
> Hive.
> The change in the forked version is recorded here 
> https://github.com/JoshRosen/hive/tree/release-1.2.1-spark2
> Note that the fork in the past included other fixes but those have all become 
> unnecessary.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-12192) Hive should carry out timestamp computations in UTC

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-12192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519919#comment-16519919
 ] 

Hive QA commented on HIVE-12192:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928682/HIVE-12192.19.patch

{color:green}SUCCESS:{color} +1 due to 79 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 6 failed/errored test(s), 14573 tests 
executed
*Failed tests:*
{noformat}
TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=258)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=258)
TestMinimrCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=258)
org.apache.hadoop.hive.cli.TestMiniDruidKafkaCliDriver.testCliDriver[druidkafkamini_basic]
 (batchId=258)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[llap_uncompressed]
 (batchId=170)
org.apache.hadoop.hive.cli.TestMiniTezCliDriver.testCliDriver[acid_vectorization_original_tez]
 (batchId=106)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11984/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11984/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11984/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 6 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928682 - PreCommit-HIVE-Build

> Hive should carry out timestamp computations in UTC
> ---
>
> Key: HIVE-12192
> URL: https://issues.apache.org/jira/browse/HIVE-12192
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Reporter: Ryan Blue
>Assignee: Jesus Camacho Rodriguez
>Priority: Blocker
>  Labels: timestamp
> Attachments: HIVE-12192.01.patch, HIVE-12192.02.patch, 
> HIVE-12192.03.patch, HIVE-12192.04.patch, HIVE-12192.05.patch, 
> HIVE-12192.06.patch, HIVE-12192.07.patch, HIVE-12192.08.patch, 
> HIVE-12192.09.patch, HIVE-12192.10.patch, HIVE-12192.11.patch, 
> HIVE-12192.12.patch, HIVE-12192.13.patch, HIVE-12192.14.patch, 
> HIVE-12192.15.patch, HIVE-12192.16.patch, HIVE-12192.17.patch, 
> HIVE-12192.18.patch, HIVE-12192.19.patch, HIVE-12192.patch
>
>
> Hive currently uses the "local" time of a java.sql.Timestamp to represent the 
> SQL data type TIMESTAMP WITHOUT TIME ZONE. The purpose is to be able to use 
> {{Timestamp#getYear()}} and similar methods to implement SQL functions like 
> {{year}}.
> When the SQL session's time zone is a DST zone, such as America/Los_Angeles 
> that alternates between PST and PDT, there are times that cannot be 
> represented because the effective zone skips them.
> {code}
> hive> select TIMESTAMP '2015-03-08 02:10:00.101';
> 2015-03-08 03:10:00.101
> {code}
> Using UTC instead of the SQL session time zone as the underlying zone for a 
> java.sql.Timestamp avoids this bug, while still returning correct values for 
> {{getYear}} etc. Using UTC as the convenience representation (timestamp 
> without time zone has no real zone) would make timestamp calculations more 
> consistent and avoid similar problems in the future.
> Notably, this would break the {{unix_timestamp}} UDF that specifies the 
> result is with respect to ["the default timezone and default 
> locale"|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+UDF#LanguageManualUDF-DateFunctions].
>  That function would need to be updated to use the 
> {{System.getProperty("user.timezone")}} zone.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19946) VectorizedRowBatchCtx.recordIdColumnVector cannot be shared between different JVMs

2018-06-21 Thread Teddy Choi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Teddy Choi updated HIVE-19946:
--
Attachment: HIVE-19946.2.patch

> VectorizedRowBatchCtx.recordIdColumnVector cannot be shared between different 
> JVMs
> --
>
> Key: HIVE-19946
> URL: https://issues.apache.org/jira/browse/HIVE-19946
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0
>Reporter: Teddy Choi
>Assignee: Teddy Choi
>Priority: Major
> Attachments: HIVE-19946.1.patch, HIVE-19946.2.patch
>
>
> VectorizedRowBatchCtx.recordIdColumnVector was used temporarily to pass 
> record id column, which is virtual, between a reducer and a mapper. However, 
> when the reducer and the mapper are not in a same JVM, it makes incorrect 
> results.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19921) Fix perf duration and queue name in HiveProtoLoggingHook

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519895#comment-16519895
 ] 

Hive QA commented on HIVE-19921:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928582/HIVE-19921.02-branch-3.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 9 failed/errored test(s), 14346 tests 
executed
*Failed tests:*
{noformat}
TestBeeLineDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=256)
TestDummy - did not produce a TEST-*.xml file (likely timed out) (batchId=256)
TestMiniDruidCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=256)
TestMiniDruidKafkaCliDriver - did not produce a TEST-*.xml file (likely timed 
out) (batchId=256)
TestMinimrCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=94)

[infer_bucket_sort_num_buckets.q,infer_bucket_sort_reducers_power_two.q,parallel_orderby.q,bucket_num_reducers_acid.q,scriptfile1.q,infer_bucket_sort_map_operators.q,infer_bucket_sort_merge.q,root_dir_external_table.q,infer_bucket_sort_dyn_part.q,udf_using.q]
TestTezPerfCliDriver - did not produce a TEST-*.xml file (likely timed out) 
(batchId=256)
org.apache.hadoop.hive.ql.TestWarehouseExternalDir.testManagedPaths 
(batchId=233)
org.apache.hive.service.TestHS2ImpersonationWithRemoteMS.testImpersonation 
(batchId=242)
org.apache.hive.spark.client.rpc.TestRpc.testServerPort (batchId=308)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11983/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11983/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11983/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 9 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928582 - PreCommit-HIVE-Build

> Fix perf duration and queue name in HiveProtoLoggingHook
> 
>
> Key: HIVE-19921
> URL: https://issues.apache.org/jira/browse/HIVE-19921
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Harish Jaiprakash
>Assignee: Harish Jaiprakash
>Priority: Major
> Attachments: HIVE-19921.01-branch-3.patch, HIVE-19921.01.patch, 
> HIVE-19921.02-branch-3.patch
>
>
> The perf log should return duration instead of end time.
> The queue name should be llap queue for llap queries.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-17751) Separate HMS Client and HMS server into separate sub-modules

2018-06-21 Thread Alexander Kolbasov (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-17751?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519889#comment-16519889
 ] 

Alexander Kolbasov commented on HIVE-17751:
---

There is an interesting problem with HIVE-19418. Because of this Hive QL now 
depends on server implementation since all the Txn part lives on the server 
side. I am wondering what is the intend here and how it works now - or this 
{{TxnStore}} is really a client-side thing?

[~sershe] [~alangates] any comment?

> Separate HMS Client and HMS server into separate sub-modules
> 
>
> Key: HIVE-17751
> URL: https://issues.apache.org/jira/browse/HIVE-17751
> Project: Hive
>  Issue Type: Sub-task
>  Components: Standalone Metastore
>Reporter: Vihang Karajgaonkar
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-17751.01.patch, 
> HIVE-17751.06-standalone-metastore.patch
>
>
> external applications which are interfacing with HMS should ideally only 
> include HMSClient library instead of one big library containing server as 
> well. We should ideally have a thin client library so that cross version 
> support for external applications is easier. We should sub-divide the 
> standalone module into possibly 3 modules (one for common classes, one for 
> client classes and one for server) or 2 sub-modules (one for client and one 
> for server) so that we can generate separate jars for HMS client and server.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19564) Vectorization: Fix NULL / Wrong Results issues in Arithmetic

2018-06-21 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-19564:

Status: In Progress  (was: Patch Available)

> Vectorization: Fix NULL / Wrong Results issues in Arithmetic
> 
>
> Key: HIVE-19564
> URL: https://issues.apache.org/jira/browse/HIVE-19564
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Fix For: 4.0.0
>
> Attachments: HIVE-19564.01.patch, HIVE-19564.02.patch
>
>
> Write new UT tests that use random data and intentional isRepeating batches 
> to checks for NULL and Wrong Results for vectorized arithmetic (+,-,*,/,%)
> Divide by 0 bug found in Long Scalar % Column.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19564) Vectorization: Fix NULL / Wrong Results issues in Arithmetic

2018-06-21 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-19564:

Status: Patch Available  (was: In Progress)

> Vectorization: Fix NULL / Wrong Results issues in Arithmetic
> 
>
> Key: HIVE-19564
> URL: https://issues.apache.org/jira/browse/HIVE-19564
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Fix For: 4.0.0
>
> Attachments: HIVE-19564.01.patch, HIVE-19564.02.patch
>
>
> Write new UT tests that use random data and intentional isRepeating batches 
> to checks for NULL and Wrong Results for vectorized arithmetic (+,-,*,/,%)
> Divide by 0 bug found in Long Scalar % Column.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19564) Vectorization: Fix NULL / Wrong Results issues in Arithmetic

2018-06-21 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-19564:

Attachment: HIVE-19564.02.patch

> Vectorization: Fix NULL / Wrong Results issues in Arithmetic
> 
>
> Key: HIVE-19564
> URL: https://issues.apache.org/jira/browse/HIVE-19564
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Fix For: 4.0.0
>
> Attachments: HIVE-19564.01.patch, HIVE-19564.02.patch
>
>
> Write new UT tests that use random data and intentional isRepeating batches 
> to checks for NULL and Wrong Results for vectorized arithmetic (+,-,*,/,%)
> Divide by 0 bug found in Long Scalar % Column.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19921) Fix perf duration and queue name in HiveProtoLoggingHook

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519858#comment-16519858
 ] 

Hive QA commented on HIVE-19921:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m 10s{color} 
| {color:red} 
/data/hiveptest/logs/PreCommit-HIVE-Build-11983/patches/PreCommit-HIVE-Build-11983.patch
 does not apply to master. Rebase required? Wrong Branch? See 
http://cwiki.apache.org/confluence/display/Hive/HowToContribute for help. 
{color} |
\\
\\
|| Subsystem || Report/Notes ||
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11983/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Fix perf duration and queue name in HiveProtoLoggingHook
> 
>
> Key: HIVE-19921
> URL: https://issues.apache.org/jira/browse/HIVE-19921
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Harish Jaiprakash
>Assignee: Harish Jaiprakash
>Priority: Major
> Attachments: HIVE-19921.01-branch-3.patch, HIVE-19921.01.patch, 
> HIVE-19921.02-branch-3.patch
>
>
> The perf log should return duration instead of end time.
> The queue name should be llap queue for llap queries.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19890) ACID: Inherit bucket-id from original ROW_ID for delete deltas

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19890?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519853#comment-16519853
 ] 

Hive QA commented on HIVE-19890:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928579/HIVE-19890.3.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14589 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11982/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11982/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11982/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928579 - PreCommit-HIVE-Build

> ACID: Inherit bucket-id from original ROW_ID for delete deltas
> --
>
> Key: HIVE-19890
> URL: https://issues.apache.org/jira/browse/HIVE-19890
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Major
> Attachments: HIVE-19890.1.patch, HIVE-19890.2.patch, 
> HIVE-19890.3.patch
>
>
> The ACID delete deltas for unbucketed tables are written to arbitrary files, 
> which should instead be shuffled using the bucket-id instead of hash(ROW__ID).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19841) Upgrade commons-collections to commons-collections4

2018-06-21 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman reassigned HIVE-19841:
-

Assignee: Andrew Sherman

> Upgrade commons-collections to commons-collections4
> ---
>
> Key: HIVE-19841
> URL: https://issues.apache.org/jira/browse/HIVE-19841
> Project: Hive
>  Issue Type: Task
>Affects Versions: 3.0.0, 4.0.0
>Reporter: BELUGA BEHR
>Assignee: Andrew Sherman
>Priority: Minor
>
> Perhaps time to drink the Apache champagne (eat the Apache dog food) and 
> upgrade the commons-collections library from 3.x to 4.x.
> {code}
> 3.2.2
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19404) Revise DDL Task Result Logging

2018-06-21 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman reassigned HIVE-19404:
-

Assignee: Andrew Sherman

> Revise DDL Task Result Logging
> --
>
> Key: HIVE-19404
> URL: https://issues.apache.org/jira/browse/HIVE-19404
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Affects Versions: 3.0.0, 2.4.0
>Reporter: BELUGA BEHR
>Assignee: Andrew Sherman
>Priority: Trivial
>  Labels: noob
>
> There is some logging in {{DDLTask}} that can be made better:
> {code}
> 2018-05-03 03:08:32,524 INFO  hive.ql.exec.DDLTask: 
> [HiveServer2-Background-Pool: Thread-101980]: results : 706
> {code}
> This logging should either be demoted to _debug_ level logging and/or 
> requires additional context.
> {code}
> 2018-05-03 03:08:32,524 INFO  hive.ql.exec.DDLTask: 
> [HiveServer2-Background-Pool: Thread-101980]: Found 706 tables that match the 
> SHOW DATABASE statement
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19391) RowContainer.setupWriter hangs when the file creation failed

2018-06-21 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19391?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman reassigned HIVE-19391:
-

Assignee: Andrew Sherman

> RowContainer.setupWriter hangs when the file creation failed
> 
>
> Key: HIVE-19391
> URL: https://issues.apache.org/jira/browse/HIVE-19391
> Project: Hive
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 1.0.0
>Reporter: John Doe
>Assignee: Andrew Sherman
>Priority: Major
>
> The while loop in the RowContainer.setupWriter function hangs endlessly when 
> the file creation failed, causing the delete() returns false.
> The file creation failure can be caused by different reasons, e.g., disk full.
> Here is the code snippet.
>  
> {code:java}
>   protected void setupWriter() throws HiveException {
> try {
>   if ( tmpFile != null ) {
> return;
>   }
>   String suffix = ".tmp";
>   if (this.keyObject != null) {
> suffix = "." + this.keyObject.toString() + suffix;
>   }
>   while (true) {
> parentFile = File.createTempFile("hive-rowcontainer", ""); //file 
> creation failed
> boolean success = parentFile.delete() && parentFile.mkdir(); //delete 
> returns false
> if (success) {
>   break;
> }
> LOG.debug("retry creating tmp row-container directory...");
>   }
>   ...
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19395) OutStream.write hangs with misconfigured bufferSize

2018-06-21 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19395?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman reassigned HIVE-19395:
-

Assignee: Andrew Sherman

> OutStream.write hangs with misconfigured bufferSize
> ---
>
> Key: HIVE-19395
> URL: https://issues.apache.org/jira/browse/HIVE-19395
> Project: Hive
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 1.0.0
>Reporter: John Doe
>Assignee: Andrew Sherman
>Priority: Minor
>
> When the bufferSize is configured to be 0 in the class initialization, the 
> while loop in OutStream.write function hangs endlessly.
> This is because when the bufferSize is 0, current.remaining is 0, length will 
> always > 0.
> Here is the code snippet.
> {code:java}
>OutStream(String name, int bufferSize, CompressionCodec codec, 
> OutputReceiver receiver) throws IOException {
> ...
> this.bufferSize = bufferSize; //bufferSize can be configured with 0
> ...
>   }
> private void getNewInputBuffer() throws IOException {
>   ...
>   current = ByteBuffer.allocate(bufferSize);
>   ...
>   }
>   public void write(byte[] bytes, int offset, int length) throws IOException {
> if (current == null) {
>   getNewInputBuffer();
> }
> int remaining = Math.min(current.remaining(), length);
> current.put(bytes, offset, remaining);
> uncompressedBytes += remaining;
> length -= remaining;
> while (length != 0) {//length > 0
>   spill();
>   offset += remaining;
>   remaining = Math.min(current.remaining(), length);//current.remaining() 
> == 0
>   current.put(bytes, offset, remaining);
>   uncompressedBytes += remaining;
>   length -= remaining;
> }
>   }
> {code}
> The similar case is HDFS-13513, HDFS-13514



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19392) unsynchronized index causes DataInputByteBuffer$Buffer.read hangs

2018-06-21 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19392?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman reassigned HIVE-19392:
-

Assignee: Andrew Sherman

> unsynchronized index causes DataInputByteBuffer$Buffer.read hangs
> -
>
> Key: HIVE-19392
> URL: https://issues.apache.org/jira/browse/HIVE-19392
> Project: Hive
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 1.0.0
>Reporter: John Doe
>Assignee: Andrew Sherman
>Priority: Minor
>
> In DataInputByteBuffer$Buffer class, the fields bidx and buffers, etc are 
> unsynchronized when used in read() and reset() function. In certain 
> circumstances, e.g., the reset() is invoked in a loop, the unsynchronized 
> bidx and buffers can trigger a concurrency bug.
> This bug also shows in 
> [Hadoop-15429|https://issues.apache.org/jira/browse/HADOOP-15429].



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19703) GenericUDTFGetSplits never uses num splits argument

2018-06-21 Thread Jaume M (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19703?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaume M updated HIVE-19703:
---
Attachment: HIVE-19703.1.patch
Status: Patch Available  (was: Open)

> GenericUDTFGetSplits never uses num splits argument
> ---
>
> Key: HIVE-19703
> URL: https://issues.apache.org/jira/browse/HIVE-19703
> Project: Hive
>  Issue Type: Bug
>  Components: UDF
>Reporter: Eric Wohlstadter
>Assignee: Jaume M
>Priority: Major
>  Labels: pull-request-available
> Attachments: HIVE-19703.1.patch
>
>
> The description for GenericUDTFGetSplits says
> {code}
> Returns an array of length int serialized splits for the referenced tables 
> string.
> {code}
> but the argument to control the number of splits is DOA.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19703) GenericUDTFGetSplits never uses num splits argument

2018-06-21 Thread ASF GitHub Bot (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519838#comment-16519838
 ] 

ASF GitHub Bot commented on HIVE-19703:
---

GitHub user beltran opened a pull request:

https://github.com/apache/hive/pull/377

HIVE-19703: GenericUDTFGetSplits never uses num splits argument



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/beltran/hive HIVE-19703

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/hive/pull/377.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #377


commit bc8848d7093e5047a08a720342f7abe93cf9675d
Author: Jaume Marhuenda 
Date:   2018-06-21T21:34:00Z

HIVE-19703: GenericUDTFGetSplits never uses num splits argument




> GenericUDTFGetSplits never uses num splits argument
> ---
>
> Key: HIVE-19703
> URL: https://issues.apache.org/jira/browse/HIVE-19703
> Project: Hive
>  Issue Type: Bug
>  Components: UDF
>Reporter: Eric Wohlstadter
>Assignee: Jaume M
>Priority: Major
>  Labels: pull-request-available
>
> The description for GenericUDTFGetSplits says
> {code}
> Returns an array of length int serialized splits for the referenced tables 
> string.
> {code}
> but the argument to control the number of splits is DOA.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19703) GenericUDTFGetSplits never uses num splits argument

2018-06-21 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19703?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-19703:
--
Labels: pull-request-available  (was: )

> GenericUDTFGetSplits never uses num splits argument
> ---
>
> Key: HIVE-19703
> URL: https://issues.apache.org/jira/browse/HIVE-19703
> Project: Hive
>  Issue Type: Bug
>  Components: UDF
>Reporter: Eric Wohlstadter
>Assignee: Jaume M
>Priority: Major
>  Labels: pull-request-available
>
> The description for GenericUDTFGetSplits says
> {code}
> Returns an array of length int serialized splits for the referenced tables 
> string.
> {code}
> but the argument to control the number of splits is DOA.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19902) Provide Metastore micro-benchmarks

2018-06-21 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-19902:
--
Status: Patch Available  (was: Open)

> Provide Metastore micro-benchmarks
> --
>
> Key: HIVE-19902
> URL: https://issues.apache.org/jira/browse/HIVE-19902
> Project: Hive
>  Issue Type: Improvement
>  Components: Standalone Metastore
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-19902.01.patch
>
>
> It would be very useful to have metastore benchmarks to be able to track perf 
> issues.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Issue Comment Deleted] (HIVE-19902) Provide Metastore micro-benchmarks

2018-06-21 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-19902:
--
Comment: was deleted

(was: Patch is applied on top of aea5908341c7b7084551216efc6c89d64b1cb06a)

> Provide Metastore micro-benchmarks
> --
>
> Key: HIVE-19902
> URL: https://issues.apache.org/jira/browse/HIVE-19902
> Project: Hive
>  Issue Type: Improvement
>  Components: Standalone Metastore
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-19902.01.patch
>
>
> It would be very useful to have metastore benchmarks to be able to track perf 
> issues.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19902) Provide Metastore micro-benchmarks

2018-06-21 Thread Alexander Kolbasov (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Kolbasov updated HIVE-19902:
--
Attachment: HIVE-19902.01.patch

> Provide Metastore micro-benchmarks
> --
>
> Key: HIVE-19902
> URL: https://issues.apache.org/jira/browse/HIVE-19902
> Project: Hive
>  Issue Type: Improvement
>  Components: Standalone Metastore
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Alexander Kolbasov
>Assignee: Alexander Kolbasov
>Priority: Major
> Attachments: HIVE-19902.01.patch
>
>
> It would be very useful to have metastore benchmarks to be able to track perf 
> issues.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19336) Disable SMB/Bucketmap join for external tables

2018-06-21 Thread Jason Dere (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19336?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-19336:
--
Attachment: HIVE-19336.01-branch-3.patch

> Disable SMB/Bucketmap join for external tables
> --
>
> Key: HIVE-19336
> URL: https://issues.apache.org/jira/browse/HIVE-19336
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Jason Dere
>Assignee: Jason Dere
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-19336.01-branch-3.patch, HIVE-19336.1.patch, 
> HIVE-19336.2.patch, HIVE-19336.3.patch, HIVE-19336.4.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19336) Disable SMB/Bucketmap join for external tables

2018-06-21 Thread Jason Dere (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19336?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Dere updated HIVE-19336:
--
Target Version/s: 3.1.0

> Disable SMB/Bucketmap join for external tables
> --
>
> Key: HIVE-19336
> URL: https://issues.apache.org/jira/browse/HIVE-19336
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Jason Dere
>Assignee: Jason Dere
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-19336.1.patch, HIVE-19336.2.patch, 
> HIVE-19336.3.patch, HIVE-19336.4.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19890) ACID: Inherit bucket-id from original ROW_ID for delete deltas

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19890?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519836#comment-16519836
 ] 

Hive QA commented on HIVE-19890:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
13s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
6s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
46s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m  
0s{color} | {color:blue} ql in master has 2282 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
59s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
4s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
4s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
46s{color} | {color:red} ql: The patch generated 4 new + 691 unchanged - 4 
fixed = 695 total (was 695) {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 1 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
15s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
14s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 16s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-11982/dev-support/hive-personality.sh
 |
| git revision | master / 24e16cc |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11982/yetus/diff-checkstyle-ql.txt
 |
| whitespace | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11982/yetus/whitespace-eol.txt
 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11982/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> ACID: Inherit bucket-id from original ROW_ID for delete deltas
> --
>
> Key: HIVE-19890
> URL: https://issues.apache.org/jira/browse/HIVE-19890
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0
>Reporter: Gopal V
>Assignee: Gopal V
>Priority: Major
> Attachments: HIVE-19890.1.patch, HIVE-19890.2.patch, 
> HIVE-19890.3.patch
>
>
> The ACID delete deltas for unbucketed tables are written to arbitrary files, 
> which should instead be shuffled using the bucket-id instead of hash(ROW__ID).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19963) metadata_only_queries.q fails

2018-06-21 Thread Steve Yeom (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19963?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519832#comment-16519832
 ] 

Steve Yeom commented on HIVE-19963:
---

Hi [~sershe] please look at the patch for this jira. 
Thanks, 
Steve.

> metadata_only_queries.q fails
> -
>
> Key: HIVE-19963
> URL: https://issues.apache.org/jira/browse/HIVE-19963
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Affects Versions: 4.0.0
>Reporter: Steve Yeom
>Assignee: Steve Yeom
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-19963.01.master-txnstats.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19671) Distribute by rand() can lead to data inconsistency

2018-06-21 Thread Xuefu Zhang (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19671?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519830#comment-16519830
 ] 

Xuefu Zhang commented on HIVE-19671:


Printing a warning is good, but we may not know if a partitioning is 
non-deterministic. Let me know your idea. Thanks.

> Distribute by rand() can lead to data inconsistency
> ---
>
> Key: HIVE-19671
> URL: https://issues.apache.org/jira/browse/HIVE-19671
> Project: Hive
>  Issue Type: Bug
>Reporter: Rui Li
>Assignee: Rui Li
>Priority: Major
>
> Noticed the following queries can give different results:
> {code}
> select count(*) from tbl;
> select count(*) from (select * from tbl distribute by rand()) a;
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19963) metadata_only_queries.q fails

2018-06-21 Thread Steve Yeom (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Yeom updated HIVE-19963:
--
Attachment: HIVE-19963.01.master-txnstats.patch

> metadata_only_queries.q fails
> -
>
> Key: HIVE-19963
> URL: https://issues.apache.org/jira/browse/HIVE-19963
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Affects Versions: 4.0.0
>Reporter: Steve Yeom
>Assignee: Steve Yeom
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-19963.01.master-txnstats.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19963) metadata_only_queries.q fails

2018-06-21 Thread Steve Yeom (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Yeom updated HIVE-19963:
--
Status: Patch Available  (was: In Progress)

> metadata_only_queries.q fails
> -
>
> Key: HIVE-19963
> URL: https://issues.apache.org/jira/browse/HIVE-19963
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Affects Versions: 4.0.0
>Reporter: Steve Yeom
>Assignee: Steve Yeom
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-19963.01.master-txnstats.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19581) view do not support unicode characters well

2018-06-21 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman updated HIVE-19581:
--
Attachment: HIVE-19581.1.patch
Status: Patch Available  (was: Open)

This works on master. I am adding a test to avoid future regressions, 
especially as our unicode coverage seems limited.

> view do not support unicode characters well
> ---
>
> Key: HIVE-19581
> URL: https://issues.apache.org/jira/browse/HIVE-19581
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.1.0
>Reporter: kai
>Assignee: Andrew Sherman
>Priority: Major
> Attachments: HIVE-19581.1.patch, explain.png, metastore.png
>
>
> create table t_test (name ,string) ;
>  insert into table t_test VALUES ('李四');
>  create view t_view_test as select * from t_test where name='李四';
> when select  * from t_view_test   no  records return



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19378) "hive.lock.numretries" Is Misleading

2018-06-21 Thread Alice Fan (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19378?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519821#comment-16519821
 ] 

Alice Fan commented on HIVE-19378:
--

Hi [~belugabehr],
I made change for 'hive.lock.numretries' and 'hive.unlock.numretries' at Hive 
Conf. properties at the 
[wiki|https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties]
 page. Is there a way to add a highlight or #anchor? Just curious :) 



> "hive.lock.numretries" Is Misleading
> 
>
> Key: HIVE-19378
> URL: https://issues.apache.org/jira/browse/HIVE-19378
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Affects Versions: 3.0.0, 2.4.0
>Reporter: BELUGA BEHR
>Assignee: Alice Fan
>Priority: Minor
> Attachments: HIVE-19378.1.patch
>
>
> Configuration 'hive.lock.numretries' is confusing.  It's not actually a 
> 'retry' count, it's the total number of attempt to try:
>  
> {code:java|title=ZooKeeperHiveLockManager.java}
> do {
>   lastException = null;
>   tryNum++;
>   try {
> if (tryNum > 1) {
>   Thread.sleep(sleepTime);
>   prepareRetry();
> }
> ret = lockPrimitive(key, mode, keepAlive, parentCreated, 
> conflictingLocks);
> ...
> } while (tryNum < numRetriesForLock);
> {code}
> So, from this code you can see that on the first loop, {{tryNum}} is set to 
> 1, in which case, if the configuration num*retries* is set to 1, there will 
> be one attempt total.  With a *retry* value of 1, I would assume one initial 
> attempt and one additional retry.  Please change to:
> {code}
> while (tryNum <= numRetriesForLock);
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Work started] (HIVE-19963) metadata_only_queries.q fails

2018-06-21 Thread Steve Yeom (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-19963 started by Steve Yeom.
-
> metadata_only_queries.q fails
> -
>
> Key: HIVE-19963
> URL: https://issues.apache.org/jira/browse/HIVE-19963
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Affects Versions: 4.0.0
>Reporter: Steve Yeom
>Assignee: Steve Yeom
>Priority: Major
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519819#comment-16519819
 ] 

Hive QA commented on HIVE-19928:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928678/HIVE-19928.01-branch-3.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11981/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11981/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11981/

Messages:
{noformat}
 This message was trimmed, see log for full details 
+ git reset --hard HEAD
HEAD is now at 6adab1c HIVE-19928 : Load Data for managed tables should set the 
owner of loaded files to a configurable user (Deepak Jaiswal, reviewed by Jason 
Dere)
+ git clean -f -d
+ git checkout branch-3
Switched to branch 'branch-3'
Your branch is behind 'origin/branch-3' by 6 commits, and can be fast-forwarded.
  (use "git pull" to update your local branch)
+ git reset --hard origin/branch-3
HEAD is now at aaeb38b HIVE-19892: Disable query results cache for for 
HiveServer2 doAs=true (Jason Dere, reviewed by Sergey Shelukhin)
+ git merge --ff-only origin/branch-3
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2018-06-21 22:40:10.376
+ rm -rf ../yetus_PreCommit-HIVE-Build-11981
+ mkdir ../yetus_PreCommit-HIVE-Build-11981
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-11981
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-11981/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
error: a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java: does not 
exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/MoveTask.java: does not 
exist in index
error: a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java: does not 
exist in index
error: a/ql/src/test/org/apache/hadoop/hive/ql/metadata/TestHiveCopyFiles.java: 
does not exist in index
Going to apply patch with: git apply -p1
+ [[ maven == \m\a\v\e\n ]]
+ rm -rf /data/hiveptest/working/maven/org/apache/hive
+ mvn -B clean install -DskipTests -T 4 -q 
-Dmaven.repo.local=/data/hiveptest/working/maven
protoc-jar: executing: [/tmp/protoc3291016345283480419.exe, --version]
libprotoc 2.5.0
protoc-jar: executing: [/tmp/protoc3291016345283480419.exe, 
-I/data/hiveptest/working/apache-github-source-source/standalone-metastore/src/main/protobuf/org/apache/hadoop/hive/metastore,
 
--java_out=/data/hiveptest/working/apache-github-source-source/standalone-metastore/target/generated-sources,
 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto]
ANTLR Parser Generator  Version 3.5.2
Output file 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/target/generated-sources/org/apache/hadoop/hive/metastore/parser/FilterParser.java
 does not exist: must build 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/src/main/java/org/apache/hadoop/hive/metastore/parser/Filter.g
org/apache/hadoop/hive/metastore/parser/Filter.g
[ERROR] COMPILATION ERROR : 
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:[5576,17]
 class, interface, or enum expected
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:[5578,3]
 class, interface, or enum expected
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:[5580,17]
 class, interface, or enum expected
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:[5582,5]
 class, interface, or enum expected
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:[5583,5]
 class, interface, or enum expected
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:[5585,5]
 class, interface, or enum expected
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:[5587,5]
 class, interface, or enum expected
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java:[5594,10]
 class, interface, or enum expected
[ERROR] 
/data/hiveptest/working/apache-github-source-source/common/src/java/org/apache/hadoop/hive/conf/HiveC

[jira] [Commented] (HIVE-18725) Improve error handling for subqueries if there is wrong column reference

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18725?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519818#comment-16519818
 ] 

Hive QA commented on HIVE-18725:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928569/HIVE-18725.08.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14542 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11980/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11980/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11980/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928569 - PreCommit-HIVE-Build

> Improve error handling for subqueries if there is wrong column reference
> 
>
> Key: HIVE-18725
> URL: https://issues.apache.org/jira/browse/HIVE-18725
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Igor Kryvenko
>Priority: Major
> Attachments: HIVE-18725.01.patch, HIVE-18725.02.patch, 
> HIVE-18725.03.patch, HIVE-18725.04.patch, HIVE-18725.05.patch, 
> HIVE-18725.06.patch, HIVE-18725.07.patch, HIVE-18725.08.patch
>
>
> If there is a column reference within subquery which doesn't exist Hive 
> throws misleading error message.
> e.g. 
> {code:sql}
> select * from table1 where table1.col1 IN (select col2 from table2 where 
> table2.col1=table1.non_existing_column) and table1.col1 IN (select 4);
> {code}
> The above query, assuming table1 doesn't have non_existing_column, will throw 
> following misleading error:
> {noformat}
> FAILED: SemanticException Line 0:-1 Unsupported SubQuery Expression 'col1': 
> Only 1 SubQuery expression is supported.
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18725) Improve error handling for subqueries if there is wrong column reference

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18725?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519804#comment-16519804
 ] 

Hive QA commented on HIVE-18725:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  8m 
 4s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
7s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
42s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m 
11s{color} | {color:blue} ql in master has 2282 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
57s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
25s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m  
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
41s{color} | {color:green} ql: The patch generated 0 new + 166 unchanged - 5 
fixed = 166 total (was 171) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m 
14s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 24m 13s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-11980/dev-support/hive-personality.sh
 |
| git revision | master / 6adab1c |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| modules | C: ql U: ql |
| Console output | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11980/yetus.txt |
| Powered by | Apache Yetushttp://yetus.apache.org |


This message was automatically generated.



> Improve error handling for subqueries if there is wrong column reference
> 
>
> Key: HIVE-18725
> URL: https://issues.apache.org/jira/browse/HIVE-18725
> Project: Hive
>  Issue Type: Improvement
>  Components: Query Planning
>Reporter: Vineet Garg
>Assignee: Igor Kryvenko
>Priority: Major
> Attachments: HIVE-18725.01.patch, HIVE-18725.02.patch, 
> HIVE-18725.03.patch, HIVE-18725.04.patch, HIVE-18725.05.patch, 
> HIVE-18725.06.patch, HIVE-18725.07.patch, HIVE-18725.08.patch
>
>
> If there is a column reference within subquery which doesn't exist Hive 
> throws misleading error message.
> e.g. 
> {code:sql}
> select * from table1 where table1.col1 IN (select col2 from table2 where 
> table2.col1=table1.non_existing_column) and table1.col1 IN (select 4);
> {code}
> The above query, assuming table1 doesn't have non_existing_column, will throw 
> following misleading error:
> {noformat}
> FAILED: SemanticException Line 0:-1 Unsupported SubQuery Expression 'col1': 
> Only 1 SubQuery expression is supported.
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-12192) Hive should carry out timestamp computations in UTC

2018-06-21 Thread Jesus Camacho Rodriguez (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-12192?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-12192:
---
Attachment: HIVE-12192.19.patch

> Hive should carry out timestamp computations in UTC
> ---
>
> Key: HIVE-12192
> URL: https://issues.apache.org/jira/browse/HIVE-12192
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Reporter: Ryan Blue
>Assignee: Jesus Camacho Rodriguez
>Priority: Blocker
>  Labels: timestamp
> Attachments: HIVE-12192.01.patch, HIVE-12192.02.patch, 
> HIVE-12192.03.patch, HIVE-12192.04.patch, HIVE-12192.05.patch, 
> HIVE-12192.06.patch, HIVE-12192.07.patch, HIVE-12192.08.patch, 
> HIVE-12192.09.patch, HIVE-12192.10.patch, HIVE-12192.11.patch, 
> HIVE-12192.12.patch, HIVE-12192.13.patch, HIVE-12192.14.patch, 
> HIVE-12192.15.patch, HIVE-12192.16.patch, HIVE-12192.17.patch, 
> HIVE-12192.18.patch, HIVE-12192.19.patch, HIVE-12192.patch
>
>
> Hive currently uses the "local" time of a java.sql.Timestamp to represent the 
> SQL data type TIMESTAMP WITHOUT TIME ZONE. The purpose is to be able to use 
> {{Timestamp#getYear()}} and similar methods to implement SQL functions like 
> {{year}}.
> When the SQL session's time zone is a DST zone, such as America/Los_Angeles 
> that alternates between PST and PDT, there are times that cannot be 
> represented because the effective zone skips them.
> {code}
> hive> select TIMESTAMP '2015-03-08 02:10:00.101';
> 2015-03-08 03:10:00.101
> {code}
> Using UTC instead of the SQL session time zone as the underlying zone for a 
> java.sql.Timestamp avoids this bug, while still returning correct values for 
> {{getYear}} etc. Using UTC as the convenience representation (timestamp 
> without time zone has no real zone) would make timestamp calculations more 
> consistent and avoid similar problems in the future.
> Notably, this would break the {{unix_timestamp}} UDF that specifies the 
> result is with respect to ["the default timezone and default 
> locale"|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+UDF#LanguageManualUDF-DateFunctions].
>  That function would need to be updated to use the 
> {{System.getProperty("user.timezone")}} zone.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19581) view do not support unicode characters well

2018-06-21 Thread Andrew Sherman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Sherman reassigned HIVE-19581:
-

Assignee: Andrew Sherman

> view do not support unicode characters well
> ---
>
> Key: HIVE-19581
> URL: https://issues.apache.org/jira/browse/HIVE-19581
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 1.1.0
>Reporter: kai
>Assignee: Andrew Sherman
>Priority: Major
> Attachments: explain.png, metastore.png
>
>
> create table t_test (name ,string) ;
>  insert into table t_test VALUES ('李四');
>  create view t_view_test as select * from t_test where name='李四';
> when select  * from t_view_test   no  records return



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19899) Support stored as JsonFile

2018-06-21 Thread Aihua Xu (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aihua Xu updated HIVE-19899:

   Resolution: Fixed
Fix Version/s: 4.0.0
 Release Note: Support "create table ... stored as JsonFile" syntax.
   Status: Resolved  (was: Patch Available)

Pushed to master. Thanks [~ychena] and [~belugabehr] for reviewing.

> Support stored as JsonFile 
> ---
>
> Key: HIVE-19899
> URL: https://issues.apache.org/jira/browse/HIVE-19899
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Affects Versions: 3.0.0
>Reporter: Aihua Xu
>Assignee: Aihua Xu
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-19899.1.patch, HIVE-19899.2.patch, 
> HIVE-19899.3.patch, HIVE-19899.4.patch, HIVE-19899.5.patch
>
>
> This is to support "Create table ... stored as JsonFile" syntax. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18118) Explain Extended should indicate if a file being read is an EC file

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519789#comment-16519789
 ] 

Hive QA commented on HIVE-18118:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928553/HIVE-18118.3.patch

{color:green}SUCCESS:{color} +1 due to 6 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 11 failed/errored test(s), 14543 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestBeeLineDriver.testCliDriver[materialized_view_create_rewrite]
 (batchId=257)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[show_tblproperties] 
(batchId=87)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[unset_table_view_property]
 (batchId=54)
org.apache.hadoop.hive.cli.TestMiniDruidCliDriver.testCliDriver[druidmini_mv] 
(batchId=257)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[materialized_view_create]
 (batchId=172)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[materialized_view_create_rewrite]
 (batchId=154)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[materialized_view_create_rewrite_dummy]
 (batchId=167)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[materialized_view_create_rewrite_multi_db]
 (batchId=171)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[materialized_view_describe]
 (batchId=171)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[unset_table_property]
 (batchId=95)
org.apache.hadoop.hive.metastore.client.TestRuntimeStats.testCleanup[Embedded] 
(batchId=212)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11978/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11978/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11978/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 11 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928553 - PreCommit-HIVE-Build

> Explain Extended should indicate if a file being read is an EC file
> ---
>
> Key: HIVE-18118
> URL: https://issues.apache.org/jira/browse/HIVE-18118
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sahil Takiar
>Assignee: Andrew Sherman
>Priority: Major
> Attachments: HIVE-18118.1.patch, HIVE-18118.2.patch, 
> HIVE-18118.3.patch
>
>
> We already print out the files Hive will read in the explain extended 
> command, we just have to modify it to say whether or not its an EC file.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19867) Test and verify Concurrent INSERTS

2018-06-21 Thread Steve Yeom (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519787#comment-16519787
 ] 

Steve Yeom commented on HIVE-19867:
---

Sergey and I talked about this. He mentioned several cases from the perspective 
of Readers like:

Let's assume the txnId of a stats entity(table or partition) is 1 and its write 
id on table1 is 11.
Then we may have
case1: Suppose concurrent writes 12 (with txnId 2) and 13 (txnId 3) and 
concurrent reader 14 (txnId 4).
  Here concurrent reader 14 has 12,13 as open writes in its writeIdList.
  1) Write 12 comes and updates the stats of table1 and its transaction is 
committed. 
So now txnid of the stats is txnid is 2.
  2) Then write 13 comes in and checks itself (number 13) from stats's 
writeIdList by txnId 2. 
13 should be there in the writeIdList. So it detects concurrent writes and 
can turn the flag
(COLUMN_STATS_ACCURATE) off.
  3) Now reader comes in and finds the stats is not valid by simply checking 
the flag.
(the reader also can determine the stats' validity by comparing 
writeIdLists of itself and the stats)
case2: Suppose concurrent writes 12 and 13. But assume we have a reader 14 
(txnId 4) that started its 
  transaction after writes 12 and 13 are done. 
  If the flag is still on and the txnId in TBLS/PARTITIONS is 3, then reader 14 
does not have  a way 
  to figure out the stats are invalid due to concurrent writes since its own 
writeIdList for table1 does not have
  12, 13 as open writes and both are committed.

> Test and verify Concurrent INSERTS  
> 
>
> Key: HIVE-19867
> URL: https://issues.apache.org/jira/browse/HIVE-19867
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Affects Versions: 4.0.0
>Reporter: Steve Yeom
>Assignee: Steve Yeom
>Priority: Major
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19965) Make HiveEndPoint use IMetaStoreClient.add_partition

2018-06-21 Thread Eugene Koifman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman reassigned HIVE-19965:
-


> Make HiveEndPoint use IMetaStoreClient.add_partition
> 
>
> Key: HIVE-19965
> URL: https://issues.apache.org/jira/browse/HIVE-19965
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Blocker
>
> it currently uses "alter table add partition if exists..."
> which since HIVE-18814 requires X lock on the table which blocks other 
> streaming writers from making progress.  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19928:
--
Attachment: HIVE-19928.01-branch-3.patch

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Fix For: 3.1.0, 4.0.0
>
> Attachments: HIVE-19928.01-branch-3.patch, HIVE-19928.1.patch, 
> HIVE-19928.2.patch, HIVE-19928.3.patch, HIVE-19928.4.patch, 
> HIVE-19928.5.patch, HIVE-19928.6.patch, HIVE-19928.7.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19965) Make HiveEndPoint use IMetaStoreClient.add_partition

2018-06-21 Thread Eugene Koifman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-19965:
--
Status: Patch Available  (was: Open)

[~prasanth_j] could you review please

> Make HiveEndPoint use IMetaStoreClient.add_partition
> 
>
> Key: HIVE-19965
> URL: https://issues.apache.org/jira/browse/HIVE-19965
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Blocker
> Attachments: HIVE-19965.01.patch
>
>
> it currently uses "alter table add partition if exists..."
> which since HIVE-18814 requires X lock on the table which blocks other 
> streaming writers from making progress.  
> HIVE-19961 to investigate if X lock can be relaxed, etc.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19965) Make HiveEndPoint use IMetaStoreClient.add_partition

2018-06-21 Thread Eugene Koifman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-19965:
--
Attachment: HIVE-19965.01.patch

> Make HiveEndPoint use IMetaStoreClient.add_partition
> 
>
> Key: HIVE-19965
> URL: https://issues.apache.org/jira/browse/HIVE-19965
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Blocker
> Attachments: HIVE-19965.01.patch
>
>
> it currently uses "alter table add partition if exists..."
> which since HIVE-18814 requires X lock on the table which blocks other 
> streaming writers from making progress.  
> HIVE-19961 to investigate if X lock can be relaxed, etc.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19965) Make HiveEndPoint use IMetaStoreClient.add_partition

2018-06-21 Thread Eugene Koifman (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eugene Koifman updated HIVE-19965:
--
Description: 
it currently uses "alter table add partition if exists..."

which since HIVE-18814 requires X lock on the table which blocks other 
streaming writers from making progress.  

HIVE-19961 to investigate if X lock can be relaxed, etc.

  was:
it currently uses "alter table add partition if exists..."

which since HIVE-18814 requires X lock on the table which blocks other 
streaming writers from making progress.  


> Make HiveEndPoint use IMetaStoreClient.add_partition
> 
>
> Key: HIVE-19965
> URL: https://issues.apache.org/jira/browse/HIVE-19965
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 3.0.0
>Reporter: Eugene Koifman
>Assignee: Eugene Koifman
>Priority: Blocker
>
> it currently uses "alter table add partition if exists..."
> which since HIVE-18814 requires X lock on the table which blocks other 
> streaming writers from making progress.  
> HIVE-19961 to investigate if X lock can be relaxed, etc.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18118) Explain Extended should indicate if a file being read is an EC file

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519774#comment-16519774
 ] 

Hive QA commented on HIVE-18118:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
40s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  7m 
23s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  3m 
14s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
48s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
31s{color} | {color:blue} common in master has 62 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
40s{color} | {color:blue} itests/hive-unit in master has 2 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
48s{color} | {color:blue} itests/util in master has 52 extant Findbugs 
warnings. {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m 
14s{color} | {color:blue} ql in master has 2280 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  3m 
16s{color} | {color:blue} standalone-metastore in master has 227 extant 
Findbugs warnings. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m 
58s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
9s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  0m 
37s{color} | {color:red} hive-unit in the patch failed. {color} |
| {color:red}-1{color} | {color:red} mvninstall {color} | {color:red}  0m 
51s{color} | {color:red} ql in the patch failed. {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  3m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  3m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
13s{color} | {color:green} common: The patch generated 0 new + 96 unchanged - 2 
fixed = 96 total (was 98) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
17s{color} | {color:green} The patch hive-unit passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
14s{color} | {color:green} The patch util passed checkstyle {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
46s{color} | {color:green} ql: The patch generated 0 new + 501 unchanged - 35 
fixed = 501 total (was 536) {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
22s{color} | {color:green} standalone-metastore: The patch generated 0 new + 
455 unchanged - 12 fixed = 455 total (was 467) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
40s{color} | {color:green} common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
46s{color} | {color:green} hive-unit in the patch passed. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
53s{color} | {color:green} util in the patch passed. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  4m  
8s{color} | {color:green} ql in the patch passed. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m 
25s{color} | {color:green} standalone-metastore generated 0 new + 225 unchanged 
- 2 fixed = 225 total (was 227) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m 
53s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not 

[jira] [Assigned] (HIVE-19316) StatsTask fails due to ClassCastException

2018-06-21 Thread Jaume M (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19316?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaume M reassigned HIVE-19316:
--

Assignee: Jaume M

> StatsTask fails due to ClassCastException
> -
>
> Key: HIVE-19316
> URL: https://issues.apache.org/jira/browse/HIVE-19316
> Project: Hive
>  Issue Type: Bug
>  Components: Statistics
>Reporter: Rui Li
>Assignee: Jaume M
>Priority: Major
>
> The stack trace:
> {noformat}
> 2018-04-26T20:17:37,674 ERROR [pool-7-thread-11] 
> metastore.RetryingHMSHandler: java.lang.ClassCastException: 
> org.apache.hadoop.hive.metastore.api.LongColumnStatsData cannot be cast to 
> org.apache.hadoop.hive.metastore.columnstats.cache.LongColumnStatsDataInspector
> at 
> org.apache.hadoop.hive.metastore.columnstats.merge.LongColumnStatsMerger.merge(LongColumnStatsMerger.java:30)
> at 
> org.apache.hadoop.hive.metastore.utils.MetaStoreUtils.mergeColStats(MetaStoreUtils.java:1052)
> at 
> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.set_aggr_stats_for(HiveMetaStore.java:7202)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:147)
> at 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:108)
> at com.sun.proxy.$Proxy26.set_aggr_stats_for(Unknown Source)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$set_aggr_stats_for.getResult(ThriftHiveMetastore.java:16795)
> at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$set_aggr_stats_for.getResult(ThriftHiveMetastore.java:16779)
> at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
> at 
> org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:111)
> at 
> org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:107)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
> at 
> org.apache.hadoop.hive.metastore.TUGIBasedProcessor.process(TUGIBasedProcessor.java:119)
> at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19928:
--
Fix Version/s: 4.0.0
   3.1.0

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Fix For: 3.1.0, 4.0.0
>
> Attachments: HIVE-19928.1.patch, HIVE-19928.2.patch, 
> HIVE-19928.3.patch, HIVE-19928.4.patch, HIVE-19928.5.patch, 
> HIVE-19928.6.patch, HIVE-19928.7.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Deepak Jaiswal (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519755#comment-16519755
 ] 

Deepak Jaiswal edited comment on HIVE-19928 at 6/21/18 8:33 PM:


Thanks [~jdere] for the review. committed to master, preparing for branch-3


was (Author: djaiswal):
committed to master, preparing for branch-3

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Attachments: HIVE-19928.1.patch, HIVE-19928.2.patch, 
> HIVE-19928.3.patch, HIVE-19928.4.patch, HIVE-19928.5.patch, 
> HIVE-19928.6.patch, HIVE-19928.7.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Deepak Jaiswal (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519755#comment-16519755
 ] 

Deepak Jaiswal commented on HIVE-19928:
---

committed to master, preparing for branch-3

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Attachments: HIVE-19928.1.patch, HIVE-19928.2.patch, 
> HIVE-19928.3.patch, HIVE-19928.4.patch, HIVE-19928.5.patch, 
> HIVE-19928.6.patch, HIVE-19928.7.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19964) Apply resource plan fails if trigger expression has quotes

2018-06-21 Thread Prasanth Jayachandran (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19964?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519744#comment-16519744
 ] 

Prasanth Jayachandran commented on HIVE-19964:
--

Uploading [~sershe]'s patch.

lgtm, +1. Pending tests. 

> Apply resource plan fails if trigger expression has quotes
> --
>
> Key: HIVE-19964
> URL: https://issues.apache.org/jira/browse/HIVE-19964
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Aswathy Chellammal Sreekumar
>Assignee: Sergey Shelukhin
>Priority: Major
> Attachments: HIVE-19964.1.patch
>
>
> {code:java}
> 0: jdbc:hive2://localhost:1> CREATE TRIGGER global.big_hdfs_read WHEN 
> HDFS_BYTES_READ > '300kb' DO KILL;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.015 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.025 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER TRIGGER global.big_hdfs_read ADD TO 
> UNMANAGED;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.014 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.029 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ENABLE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.012 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.021 seconds
> INFO : OK
> No rows affected (0.045 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ACTIVATE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.017 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Starting task [Stage-0:DDL] in serial mode
> ERROR : FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: HDFS_BYTES_READ > 
> 300kb
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.037 seconds
> Error: Error while processing statement: FAILED: Execution Error, return code 
> 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: 
> HDFS_BYTES_READ > 300kb (state=08S01,code=1){code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19964) Apply resource plan fails if trigger expression has quotes

2018-06-21 Thread Prasanth Jayachandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-19964:
-
Status: Patch Available  (was: Open)

> Apply resource plan fails if trigger expression has quotes
> --
>
> Key: HIVE-19964
> URL: https://issues.apache.org/jira/browse/HIVE-19964
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Aswathy Chellammal Sreekumar
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19964.1.patch
>
>
> {code:java}
> 0: jdbc:hive2://localhost:1> CREATE TRIGGER global.big_hdfs_read WHEN 
> HDFS_BYTES_READ > '300kb' DO KILL;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.015 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.025 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER TRIGGER global.big_hdfs_read ADD TO 
> UNMANAGED;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.014 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.029 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ENABLE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.012 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.021 seconds
> INFO : OK
> No rows affected (0.045 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ACTIVATE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.017 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Starting task [Stage-0:DDL] in serial mode
> ERROR : FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: HDFS_BYTES_READ > 
> 300kb
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.037 seconds
> Error: Error while processing statement: FAILED: Execution Error, return code 
> 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: 
> HDFS_BYTES_READ > 300kb (state=08S01,code=1){code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19964) Apply resource plan fails if trigger expression has quotes

2018-06-21 Thread Prasanth Jayachandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran updated HIVE-19964:
-
Attachment: HIVE-19964.1.patch

> Apply resource plan fails if trigger expression has quotes
> --
>
> Key: HIVE-19964
> URL: https://issues.apache.org/jira/browse/HIVE-19964
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Aswathy Chellammal Sreekumar
>Assignee: Prasanth Jayachandran
>Priority: Major
> Attachments: HIVE-19964.1.patch
>
>
> {code:java}
> 0: jdbc:hive2://localhost:1> CREATE TRIGGER global.big_hdfs_read WHEN 
> HDFS_BYTES_READ > '300kb' DO KILL;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.015 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.025 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER TRIGGER global.big_hdfs_read ADD TO 
> UNMANAGED;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.014 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.029 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ENABLE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.012 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.021 seconds
> INFO : OK
> No rows affected (0.045 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ACTIVATE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.017 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Starting task [Stage-0:DDL] in serial mode
> ERROR : FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: HDFS_BYTES_READ > 
> 300kb
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.037 seconds
> Error: Error while processing statement: FAILED: Execution Error, return code 
> 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: 
> HDFS_BYTES_READ > 300kb (state=08S01,code=1){code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19964) Apply resource plan fails if trigger expression has quotes

2018-06-21 Thread Prasanth Jayachandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran reassigned HIVE-19964:


Assignee: Sergey Shelukhin  (was: Prasanth Jayachandran)

> Apply resource plan fails if trigger expression has quotes
> --
>
> Key: HIVE-19964
> URL: https://issues.apache.org/jira/browse/HIVE-19964
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Aswathy Chellammal Sreekumar
>Assignee: Sergey Shelukhin
>Priority: Major
> Attachments: HIVE-19964.1.patch
>
>
> {code:java}
> 0: jdbc:hive2://localhost:1> CREATE TRIGGER global.big_hdfs_read WHEN 
> HDFS_BYTES_READ > '300kb' DO KILL;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.015 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.025 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER TRIGGER global.big_hdfs_read ADD TO 
> UNMANAGED;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.014 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.029 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ENABLE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.012 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.021 seconds
> INFO : OK
> No rows affected (0.045 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ACTIVATE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.017 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Starting task [Stage-0:DDL] in serial mode
> ERROR : FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: HDFS_BYTES_READ > 
> 300kb
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.037 seconds
> Error: Error while processing statement: FAILED: Execution Error, return code 
> 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: 
> HDFS_BYTES_READ > 300kb (state=08S01,code=1){code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19964) Apply resource plan fails if trigger expression has quotes

2018-06-21 Thread Prasanth Jayachandran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prasanth Jayachandran reassigned HIVE-19964:


Assignee: Prasanth Jayachandran

> Apply resource plan fails if trigger expression has quotes
> --
>
> Key: HIVE-19964
> URL: https://issues.apache.org/jira/browse/HIVE-19964
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.0, 4.0.0
>Reporter: Aswathy Chellammal Sreekumar
>Assignee: Prasanth Jayachandran
>Priority: Major
>
> {code:java}
> 0: jdbc:hive2://localhost:1> CREATE TRIGGER global.big_hdfs_read WHEN 
> HDFS_BYTES_READ > '300kb' DO KILL;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.015 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890):
>  CREATE TRIGGER global.big_hdfs_read WHEN HDFS_BYTES_READ > '300kb' DO KILL
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131017_72b1441b-d790-4db7-83ca-479735843890);
>  Time taken: 0.025 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER TRIGGER global.big_hdfs_read ADD TO 
> UNMANAGED;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.014 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5):
>  ALTER TRIGGER global.big_hdfs_read ADD TO UNMANAGED
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131031_dd489324-db23-412f-9409-32ba697a10e5);
>  Time taken: 0.029 seconds
> INFO : OK
> No rows affected (0.054 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ENABLE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.012 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e):
>  ALTER RESOURCE PLAN global ENABLE
> INFO : Starting task [Stage-0:DDL] in serial mode
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131036_26a5f4f3-91e3-4bec-ab42-800adb90104e);
>  Time taken: 0.021 seconds
> INFO : OK
> No rows affected (0.045 seconds)
> 0: jdbc:hive2://localhost:1> ALTER RESOURCE PLAN global ACTIVATE;
> INFO : Compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Semantic Analysis Completed (retrial = false)
> INFO : Returning Hive schema: Schema(fieldSchemas:null, properties:null)
> INFO : Completed compiling 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.017 seconds
> INFO : Executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b):
>  ALTER RESOURCE PLAN global ACTIVATE
> INFO : Starting task [Stage-0:DDL] in serial mode
> ERROR : FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: HDFS_BYTES_READ > 
> 300kb
> INFO : Completed executing 
> command(queryId=pjayachandran_20180621131037_551b2af0-321b-4638-8ac0-76771a159f4b);
>  Time taken: 0.037 seconds
> Error: Error while processing statement: FAILED: Execution Error, return code 
> 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Invalid expression: 
> HDFS_BYTES_READ > 300kb (state=08S01,code=1){code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19889) Wrong results due to PPD of non deterministic functions with CBO

2018-06-21 Thread Naveen Gangam (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naveen Gangam updated HIVE-19889:
-
  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

Fix has been committed to master. Thank you for your contribution [~janulatha]

> Wrong results due to PPD of non deterministic functions with CBO
> 
>
> Key: HIVE-19889
> URL: https://issues.apache.org/jira/browse/HIVE-19889
> Project: Hive
>  Issue Type: Bug
>Reporter: Janaki Lahorani
>Assignee: Janaki Lahorani
>Priority: Major
> Fix For: 4.0.0
>
> Attachments: HIVE-19889.1.patch, HIVE-19889.2.patch
>
>
> The following query can give wrong results when CBO is on:
> {code}
> select * from (
> select part1,randum123
> from (SELECT *, cast(rand() as double) AS randum123 FROM testA where 
> part1='CA' and part2 = 'ABC') a
> where randum123 <= 0.5) s where s.randum123 > 0.25 limit 20;
> The plan of the query is as follows:
> STAGE PLANS:
>   Stage: Stage-1
> Map Reduce
>   Map Operator Tree:
>   TableScan
> alias: testa
> Statistics: Num rows: 2 Data size: 4580 Basic stats: COMPLETE 
> Column stats: NONE
> Filter Operator
>   predicate: ((rand() <= 0.5D) and (rand() > 0.25D)) (type: 
> boolean)
>   Statistics: Num rows: 1 Data size: 2290 Basic stats: COMPLETE 
> Column stats: NONE
>   Select Operator
> expressions: 'CA' (type: string), rand() (type: double)
> outputColumnNames: _col0, _col1
> Statistics: Num rows: 1 Data size: 2290 Basic stats: COMPLETE 
> Column stats: NONE
> Limit
>   Number of rows: 20
>   Statistics: Num rows: 1 Data size: 2290 Basic stats: 
> COMPLETE Column stats: NONE
>   File Output Operator
> compressed: false
> Statistics: Num rows: 1 Data size: 2290 Basic stats: 
> COMPLETE Column stats: NONE
> table:
> input format: 
> org.apache.hadoop.mapred.SequenceFileInputFormat
> output format: 
> org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat
> serde: 
> org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
>   Stage: Stage-0
> Fetch Operator
>   limit: 20
>   Processor Tree:
> ListSink
> {code}
> The relevant part in the plan is the filter:
> {code}
> Filter Operator
>   predicate: ((rand() <= 0.5D) and (rand() > 0.25D)) (type: 
> boolean)
> {code}
> The predicates randum123 <= 0.5 and s.randum123 > 0.25 were pushed down.  And 
> randum123 was resolved to rand().  This is bad because it will result in 
> invocation of rand() two times and rand() UDF is non-deterministic.  Both the 
> rand calls can generate values that can satisfy the predicates independently, 
> but not together, whereas the original intention of the query is to give 
> results when rand falls between 0.25 and 0.5.
> A sample result:
> {code}
> CA0.9191984370369802
> CA0.397933021566812
> {code}
> where the condition was not satisfied.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-17751) Separate HMS Client and HMS server into separate sub-modules

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-17751?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519729#comment-16519729
 ] 

Hive QA commented on HIVE-17751:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928545/HIVE-17751.01.patch

{color:red}ERROR:{color} -1 due to build exiting with an error

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11977/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11977/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11977/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Tests exited with: NonZeroExitCodeException
Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit 
status 1 and output '+ date '+%Y-%m-%d %T.%3N'
2018-06-21 20:10:59.762
+ [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]]
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ 
PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games
+ export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m '
+ ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m '
+ export 'MAVEN_OPTS=-Xmx1g '
+ MAVEN_OPTS='-Xmx1g '
+ cd /data/hiveptest/working/
+ tee /data/hiveptest/logs/PreCommit-HIVE-Build-11977/source-prep.txt
+ [[ false == \t\r\u\e ]]
+ mkdir -p maven ivy
+ [[ git = \s\v\n ]]
+ [[ git = \g\i\t ]]
+ [[ -z master ]]
+ [[ -d apache-github-source-source ]]
+ [[ ! -d apache-github-source-source/.git ]]
+ [[ ! -d apache-github-source-source ]]
+ date '+%Y-%m-%d %T.%3N'
2018-06-21 20:10:59.766
+ cd apache-github-source-source
+ git fetch origin
+ git reset --hard HEAD
HEAD is now at 9cd3960 HIVE-19892: Disable query results cache for for 
HiveServer2 doAs=true (Jason Dere, reviewed by Sergey Shelukhin)
+ git clean -f -d
+ git checkout master
Already on 'master'
Your branch is up-to-date with 'origin/master'.
+ git reset --hard origin/master
HEAD is now at 9cd3960 HIVE-19892: Disable query results cache for for 
HiveServer2 doAs=true (Jason Dere, reviewed by Sergey Shelukhin)
+ git merge --ff-only origin/master
Already up-to-date.
+ date '+%Y-%m-%d %T.%3N'
2018-06-21 20:11:01.115
+ rm -rf ../yetus_PreCommit-HIVE-Build-11977
+ mkdir ../yetus_PreCommit-HIVE-Build-11977
+ git gc
+ cp -R . ../yetus_PreCommit-HIVE-Build-11977
+ mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-11977/yetus
+ patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh
+ patchFilePath=/data/hiveptest/working/scratch/build.patch
+ [[ -f /data/hiveptest/working/scratch/build.patch ]]
+ chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh
+ /data/hiveptest/working/scratch/smart-apply-patch.sh 
/data/hiveptest/working/scratch/build.patch
fatal: git apply: bad git-diff - inconsistent old filename on line 1887
Going to apply patch with: git apply -p1
/data/hiveptest/working/scratch/build.patch:5175: new blank line at EOF.
+
warning: 1 line adds whitespace errors.
+ [[ maven == \m\a\v\e\n ]]
+ rm -rf /data/hiveptest/working/maven/org/apache/hive
+ mvn -B clean install -DskipTests -T 4 -q 
-Dmaven.repo.local=/data/hiveptest/working/maven
ANTLR Parser Generator  Version 3.5.2
protoc-jar: executing: [/tmp/protoc5090090272804891313.exe, --version]
libprotoc 2.5.0
protoc-jar: executing: [/tmp/protoc5090090272804891313.exe, 
-I/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/src/main/protobuf/org/apache/hadoop/hive/metastore,
 
--java_out=/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/target/generated-sources,
 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto]
ANTLR Parser Generator  Version 3.5.2
Output file 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/target/generated-sources/org/apache/hadoop/hive/metastore/parser/FilterParser.java
 does not exist: must build 
/data/hiveptest/working/apache-github-source-source/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/parser/Filter.g
org/apache/hadoop/hive/metastore/parser/Filter.g
[ERROR] Failed to execute goal on project hive-metastore: Could not resolve 
dependencies for project org.apache.hive:hive-metastore:jar:4.0.0-SNAPSHOT: 
Could not find artifact 
org.apache.hive:hive-standalone-metastore:jar:4.0.0-SNAPSHOT in 
apache.snapshots (https://repository.apache.org/snapshots) -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[

[jira] [Commented] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519727#comment-16519727
 ] 

Hive QA commented on HIVE-19928:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928668/HIVE-19928.7.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:green}SUCCESS:{color} +1 due to 14540 tests passed

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11976/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11976/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11976/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928668 - PreCommit-HIVE-Build

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Attachments: HIVE-19928.1.patch, HIVE-19928.2.patch, 
> HIVE-19928.3.patch, HIVE-19928.4.patch, HIVE-19928.5.patch, 
> HIVE-19928.6.patch, HIVE-19928.7.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-18785) Make JSON SerDe First-Class SerDe

2018-06-21 Thread Vihang Karajgaonkar (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-18785?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vihang Karajgaonkar reassigned HIVE-18785:
--

Assignee: Aihua Xu  (was: Vihang Karajgaonkar)

> Make JSON SerDe First-Class SerDe
> -
>
> Key: HIVE-18785
> URL: https://issues.apache.org/jira/browse/HIVE-18785
> Project: Hive
>  Issue Type: New Feature
>  Components: Serializers/Deserializers
>Affects Versions: 3.0.0
>Reporter: BELUGA BEHR
>Assignee: Aihua Xu
>Priority: Major
>
> According to the [Hive SerDe 
> Docs|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-RowFormats&SerDe],
>  there are some extra steps involved in getting the JSON SerDe to work:
> {quote}
> ROW FORMAT SERDE 
> 'org.apache.hive.hcatalog.data.JsonSerDe' 
> STORED AS TEXTFILE
> In some distributions, a reference to hive-hcatalog-core.jar is required.
> ADD JAR /usr/lib/hive-hcatalog/lib/hive-hcatalog-core.jar;
> {quote}
> I would like to propose that we move this SerDe into first-class status:
> {{STORED AS JSONFILE}}
> The user should have to perform no additional steps to use this SerDe.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-18785) Make JSON SerDe First-Class SerDe

2018-06-21 Thread Vihang Karajgaonkar (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-18785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519719#comment-16519719
 ] 

Vihang Karajgaonkar commented on HIVE-18785:


Looks like [~aihuaxu] already did most of the work. I will not able to work on 
this soon. Reassigning it to you [~aihuaxu] like we discussed offline.

> Make JSON SerDe First-Class SerDe
> -
>
> Key: HIVE-18785
> URL: https://issues.apache.org/jira/browse/HIVE-18785
> Project: Hive
>  Issue Type: New Feature
>  Components: Serializers/Deserializers
>Affects Versions: 3.0.0
>Reporter: BELUGA BEHR
>Assignee: Vihang Karajgaonkar
>Priority: Major
>
> According to the [Hive SerDe 
> Docs|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-RowFormats&SerDe],
>  there are some extra steps involved in getting the JSON SerDe to work:
> {quote}
> ROW FORMAT SERDE 
> 'org.apache.hive.hcatalog.data.JsonSerDe' 
> STORED AS TEXTFILE
> In some distributions, a reference to hive-hcatalog-core.jar is required.
> ADD JAR /usr/lib/hive-hcatalog/lib/hive-hcatalog-core.jar;
> {quote}
> I would like to propose that we move this SerDe into first-class status:
> {{STORED AS JSONFILE}}
> The user should have to perform no additional steps to use this SerDe.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519706#comment-16519706
 ] 

Hive QA commented on HIVE-19928:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} master Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
48s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  6m 
51s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
23s{color} | {color:green} master passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
57s{color} | {color:green} master passed {color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  0m 
34s{color} | {color:blue} common in master has 62 extant Findbugs warnings. 
{color} |
| {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue}  4m 
14s{color} | {color:blue} ql in master has 2280 extant Findbugs warnings. 
{color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
13s{color} | {color:green} master passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m  
9s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
48s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m 
22s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  1m 
22s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} checkstyle {color} | {color:red}  0m 
42s{color} | {color:red} ql: The patch generated 9 new + 256 unchanged - 14 
fixed = 265 total (was 270) {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:red}-1{color} | {color:red} findbugs {color} | {color:red}  4m 
19s{color} | {color:red} ql generated 3 new + 2279 unchanged - 1 fixed = 2282 
total (was 2280) {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m  
9s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
13s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 27m 19s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| FindBugs | module:ql |
|  |  Call to org.apache.hadoop.hive.metastore.TableType.equals(String) in 
org.apache.hadoop.hive.ql.metadata.Hive.loadPartition(Path, Table, Map, 
LoadTableDesc$LoadFileType, boolean, boolean, boolean, boolean, boolean, Long, 
int, boolean)  At Hive.java:Table, Map, LoadTableDesc$LoadFileType, boolean, 
boolean, boolean, boolean, boolean, Long, int, boolean)  At Hive.java:[line 
1757] |
|  |  Call to org.apache.hadoop.hive.metastore.TableType.equals(String) in 
org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Path, String, 
LoadTableDesc$LoadFileType, boolean, boolean, boolean, boolean, Long, int, 
boolean)  At Hive.java:String, LoadTableDesc$LoadFileType, boolean, boolean, 
boolean, boolean, Long, int, boolean)  At Hive.java:[line 2341] |
|  |  Exception is caught when Exception is not thrown in 
org.apache.hadoop.hive.ql.metadata.Hive.moveFile(HiveConf, Path, Path, boolean, 
boolean, boolean)  At Hive.java:is not thrown in 
org.apache.hadoop.hive.ql.metadata.Hive.moveFile(HiveConf, Path, Path, boolean, 
boolean, boolean)  At Hive.java:[line 3735] |
\\
\\
|| Subsystem || Report/Notes ||
| Optional Tests |  asflicense  javac  javadoc  findbugs  checkstyle  compile  |
| uname | Linux hiveptest-server-upstream 3.16.0-4-amd64 #1 SMP Debian 
3.16.36-1+deb8u1 (2016-09-03) x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 
/data/hiveptest/working/yetus_PreCommit-HIVE-Build-11976/dev-support/hive-personality.sh
 |
| git revision | master / 9cd3960 |
| Default Java | 1.8.0_111 |
| findbugs | v3.0.0 |
| checkstyle | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11976/yetus/diff-checkstyle-ql.txt
 |
| findbugs | 
http://104.198.109.242/logs//PreCommit-HIVE-Build-11976/yetus/new-findbugs-ql.html
 |
| modules | C: common ql U: . |
| Console outpu

[jira] [Commented] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Jason Dere (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519695#comment-16519695
 ] 

Jason Dere commented on HIVE-19928:
---

+1

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Attachments: HIVE-19928.1.patch, HIVE-19928.2.patch, 
> HIVE-19928.3.patch, HIVE-19928.4.patch, HIVE-19928.5.patch, 
> HIVE-19928.6.patch, HIVE-19928.7.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19911) Hive delete queries fail with Invalid table alias or column reference

2018-06-21 Thread Mykhailo Kysliuk (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mykhailo Kysliuk updated HIVE-19911:

Attachment: HIVE-19911.2-branch-2.3.patch

> Hive delete queries fail with Invalid table alias or column reference
> -
>
> Key: HIVE-19911
> URL: https://issues.apache.org/jira/browse/HIVE-19911
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.3.3
>Reporter: Mykhailo Kysliuk
>Assignee: Mykhailo Kysliuk
>Priority: Major
> Attachments: HIVE-19911.1-branch-2.3.patch, 
> HIVE-19911.2-branch-2.3.patch
>
>
> Env:
> hadoop-2.7.0
> hive-2.3.3
> OS:
> centos-release-7-5.1804.el7.centos.x86_64
> Steps to reproduce (at hive cli):
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> DROP TABLE IF EXISTS detaillineitem_all;
> DROP TABLE IF EXISTS detaillineitem_all_delete_1526330755128;
> CREATE TABLE `detaillineitem_all`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3)
> ) PARTITIONED BY (
>   `branchnumber_p` varchar(3))
> CLUSTERED BY (
>   detailid)
> INTO 25 BUCKETS
> STORED AS ORC
> TBLPROPERTIES (
>   'orc.compress'='NONE',
>   'transactional'='true');
> CREATE TABLE `detaillineitem_all_delete_1526330755128`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3),
>   `branchnumber_p` varchar(3));
> DELETE from detaillineitem_all WHERE EXISTS (
> SELECT
> 1
> FROM
> detaillineitem_all_delete_1526330755128 AS t1
> WHERE
> (detaillineitem_all.detailid = t1.detailid)
>   AND
> (detaillineitem_all.branchnumber = CAST(t1.branchnumber AS STRING)));
> {code}
> Exception:
> {code}
> 2018-06-15T16:51:48,625 ERROR [f6bd86a7-04e5-4284-9031-3b9a0ccc80f3 main] 
> ql.Driver: FAILED: SemanticException Line 0:-1 Invalid table alias or column 
> reference 'sq_1': (possible column names are: mber)) sq_corr_1)) (tok_where 
> (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 0:-1 Invalid table 
> alias or column reference 'sq_1': (possible column names are: mber)) 
> sq_corr_1)) (tok_where (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11620)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11568)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11536)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11514)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapGroupByForSemijoin(SemanticAnalyzer.java:8416)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:8305)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3278)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:9592)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10549)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10427)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:11125)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:11138)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10807)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:73)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:462)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeDelete(UpdateDeleteSemanticAnalyzer.java:111)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:81)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:512)
> at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1317)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1457)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
> at 
> org.apache.hado

[jira] [Updated] (HIVE-19911) Hive delete queries fail with Invalid table alias or column reference

2018-06-21 Thread Mykhailo Kysliuk (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mykhailo Kysliuk updated HIVE-19911:

Status: Open  (was: Patch Available)

> Hive delete queries fail with Invalid table alias or column reference
> -
>
> Key: HIVE-19911
> URL: https://issues.apache.org/jira/browse/HIVE-19911
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.3.3
>Reporter: Mykhailo Kysliuk
>Assignee: Mykhailo Kysliuk
>Priority: Major
> Attachments: HIVE-19911.1-branch-2.3.patch, 
> HIVE-19911.2-branch-2.3.patch
>
>
> Env:
> hadoop-2.7.0
> hive-2.3.3
> OS:
> centos-release-7-5.1804.el7.centos.x86_64
> Steps to reproduce (at hive cli):
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> DROP TABLE IF EXISTS detaillineitem_all;
> DROP TABLE IF EXISTS detaillineitem_all_delete_1526330755128;
> CREATE TABLE `detaillineitem_all`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3)
> ) PARTITIONED BY (
>   `branchnumber_p` varchar(3))
> CLUSTERED BY (
>   detailid)
> INTO 25 BUCKETS
> STORED AS ORC
> TBLPROPERTIES (
>   'orc.compress'='NONE',
>   'transactional'='true');
> CREATE TABLE `detaillineitem_all_delete_1526330755128`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3),
>   `branchnumber_p` varchar(3));
> DELETE from detaillineitem_all WHERE EXISTS (
> SELECT
> 1
> FROM
> detaillineitem_all_delete_1526330755128 AS t1
> WHERE
> (detaillineitem_all.detailid = t1.detailid)
>   AND
> (detaillineitem_all.branchnumber = CAST(t1.branchnumber AS STRING)));
> {code}
> Exception:
> {code}
> 2018-06-15T16:51:48,625 ERROR [f6bd86a7-04e5-4284-9031-3b9a0ccc80f3 main] 
> ql.Driver: FAILED: SemanticException Line 0:-1 Invalid table alias or column 
> reference 'sq_1': (possible column names are: mber)) sq_corr_1)) (tok_where 
> (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 0:-1 Invalid table 
> alias or column reference 'sq_1': (possible column names are: mber)) 
> sq_corr_1)) (tok_where (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11620)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11568)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11536)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11514)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapGroupByForSemijoin(SemanticAnalyzer.java:8416)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:8305)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3278)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:9592)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10549)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10427)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:11125)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:11138)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10807)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:73)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:462)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeDelete(UpdateDeleteSemanticAnalyzer.java:111)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:81)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:512)
> at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1317)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1457)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
> at 
> org.apache.hadoop.hi

[jira] [Updated] (HIVE-19911) Hive delete queries fail with Invalid table alias or column reference

2018-06-21 Thread Mykhailo Kysliuk (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mykhailo Kysliuk updated HIVE-19911:

Status: Patch Available  (was: Open)

> Hive delete queries fail with Invalid table alias or column reference
> -
>
> Key: HIVE-19911
> URL: https://issues.apache.org/jira/browse/HIVE-19911
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.3.3
>Reporter: Mykhailo Kysliuk
>Assignee: Mykhailo Kysliuk
>Priority: Major
> Attachments: HIVE-19911.1-branch-2.3.patch, 
> HIVE-19911.2-branch-2.3.patch
>
>
> Env:
> hadoop-2.7.0
> hive-2.3.3
> OS:
> centos-release-7-5.1804.el7.centos.x86_64
> Steps to reproduce (at hive cli):
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> DROP TABLE IF EXISTS detaillineitem_all;
> DROP TABLE IF EXISTS detaillineitem_all_delete_1526330755128;
> CREATE TABLE `detaillineitem_all`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3)
> ) PARTITIONED BY (
>   `branchnumber_p` varchar(3))
> CLUSTERED BY (
>   detailid)
> INTO 25 BUCKETS
> STORED AS ORC
> TBLPROPERTIES (
>   'orc.compress'='NONE',
>   'transactional'='true');
> CREATE TABLE `detaillineitem_all_delete_1526330755128`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3),
>   `branchnumber_p` varchar(3));
> DELETE from detaillineitem_all WHERE EXISTS (
> SELECT
> 1
> FROM
> detaillineitem_all_delete_1526330755128 AS t1
> WHERE
> (detaillineitem_all.detailid = t1.detailid)
>   AND
> (detaillineitem_all.branchnumber = CAST(t1.branchnumber AS STRING)));
> {code}
> Exception:
> {code}
> 2018-06-15T16:51:48,625 ERROR [f6bd86a7-04e5-4284-9031-3b9a0ccc80f3 main] 
> ql.Driver: FAILED: SemanticException Line 0:-1 Invalid table alias or column 
> reference 'sq_1': (possible column names are: mber)) sq_corr_1)) (tok_where 
> (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 0:-1 Invalid table 
> alias or column reference 'sq_1': (possible column names are: mber)) 
> sq_corr_1)) (tok_where (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11620)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11568)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11536)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11514)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapGroupByForSemijoin(SemanticAnalyzer.java:8416)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:8305)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3278)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:9592)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10549)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10427)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:11125)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:11138)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10807)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:73)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:462)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeDelete(UpdateDeleteSemanticAnalyzer.java:111)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:81)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:512)
> at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1317)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1457)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
> at 
> org.apache.hadoop.hi

[jira] [Updated] (HIVE-19911) Hive delete queries fail with Invalid table alias or column reference

2018-06-21 Thread Mykhailo Kysliuk (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mykhailo Kysliuk updated HIVE-19911:

Attachment: (was: HIVE-19911.2-branch-2.3.patch)

> Hive delete queries fail with Invalid table alias or column reference
> -
>
> Key: HIVE-19911
> URL: https://issues.apache.org/jira/browse/HIVE-19911
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.3.3
>Reporter: Mykhailo Kysliuk
>Assignee: Mykhailo Kysliuk
>Priority: Major
> Attachments: HIVE-19911.1-branch-2.3.patch, 
> HIVE-19911.2-branch-2.3.patch
>
>
> Env:
> hadoop-2.7.0
> hive-2.3.3
> OS:
> centos-release-7-5.1804.el7.centos.x86_64
> Steps to reproduce (at hive cli):
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> DROP TABLE IF EXISTS detaillineitem_all;
> DROP TABLE IF EXISTS detaillineitem_all_delete_1526330755128;
> CREATE TABLE `detaillineitem_all`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3)
> ) PARTITIONED BY (
>   `branchnumber_p` varchar(3))
> CLUSTERED BY (
>   detailid)
> INTO 25 BUCKETS
> STORED AS ORC
> TBLPROPERTIES (
>   'orc.compress'='NONE',
>   'transactional'='true');
> CREATE TABLE `detaillineitem_all_delete_1526330755128`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3),
>   `branchnumber_p` varchar(3));
> DELETE from detaillineitem_all WHERE EXISTS (
> SELECT
> 1
> FROM
> detaillineitem_all_delete_1526330755128 AS t1
> WHERE
> (detaillineitem_all.detailid = t1.detailid)
>   AND
> (detaillineitem_all.branchnumber = CAST(t1.branchnumber AS STRING)));
> {code}
> Exception:
> {code}
> 2018-06-15T16:51:48,625 ERROR [f6bd86a7-04e5-4284-9031-3b9a0ccc80f3 main] 
> ql.Driver: FAILED: SemanticException Line 0:-1 Invalid table alias or column 
> reference 'sq_1': (possible column names are: mber)) sq_corr_1)) (tok_where 
> (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 0:-1 Invalid table 
> alias or column reference 'sq_1': (possible column names are: mber)) 
> sq_corr_1)) (tok_where (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11620)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11568)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11536)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11514)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapGroupByForSemijoin(SemanticAnalyzer.java:8416)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:8305)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3278)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:9592)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10549)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10427)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:11125)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:11138)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10807)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:73)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:462)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeDelete(UpdateDeleteSemanticAnalyzer.java:111)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:81)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:512)
> at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1317)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1457)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
> at 
> org.

[jira] [Updated] (HIVE-19941) Row based Filters added via Hive Ranger policies are not pushed to druid

2018-06-21 Thread Nishant Bangarwa (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19941?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nishant Bangarwa updated HIVE-19941:

Attachment: HIVE-19941.1.patch

> Row based Filters added via Hive Ranger policies are not pushed to druid
> 
>
> Key: HIVE-19941
> URL: https://issues.apache.org/jira/browse/HIVE-19941
> Project: Hive
>  Issue Type: Bug
>Reporter: Nishant Bangarwa
>Assignee: Nishant Bangarwa
>Priority: Major
> Attachments: HIVE-19941.1.patch, HIVE-19941.patch
>
>
> Issue is that when applying table mask we add virtual columns, however 
> non-native tables do not have virtual columns, we need to skip adding virtual 
> columns when generating masking query. 
> Stack Trace - 
> {code} 
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 1:79 Invalid table 
> alias or column reference 'BLOCK__OFFSET__INSIDE__FILE'
> : (possible column names are: __time, yearmonth, year, month, dayofmonth, 
> dayofweek, weekofyear, hour, minute, second, payment_typ
> e, fare_amount, surcharge, mta_tax, tip_amount, tolls_amount, total_amount, 
> trip_time)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11830)
>  ~[hive-exec-2.1.0.2.6.
> 4.0-91.jar:2.1.0.2.6.4.0-91]
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11778)
>  ~[hive-exec-2.1.0.2.6.4.0
> -91.jar:2.1.0.2.6.4.0-91]
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.genSelectLogicalPlan(CalcitePlanner.java:3780)
>  ~[hi
> ve-exec-2.1.0.2.6.4.0-91.jar:2.1.0.2.6.4.0-91]
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.genLogicalPlan(CalcitePlanner.java:4117)
>  ~[hive-exe
> c-2.1.0.2.6.4.0-91.jar:2.1.0.2.6.4.0-91]
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.genLogicalPlan(CalcitePlanner.java:4016)
>  ~[hive-exe
> c-2.1.0.2.6.4.0-91.jar:2.1.0.2.6.4.0-91]
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.genLogicalPlan(CalcitePlanner.java:4060)
>  ~[hive-exe
> c-2.1.0.2.6.4.0-91.jar:2.1.0.2.6.4.0-91]
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.apply(CalcitePlanner.java:1340)
>  ~[hive-exec-2.1.0.2
> .6.4.0-91.jar:2.1.0.2.6.4.0-91]
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.apply(CalcitePlanner.java:1277)
>  ~[hive-exec-2.1.0.2
> .6.4.0-91.jar:2.1.0.2.6.4.0-91]
> at org.apache.calcite.tools.Frameworks$1.apply(Frameworks.java:113) 
> ~[calcite-core-1.10.0.2.6.4.0-91.jar:1.10.0.2.6.4.0-91
> ]
> at 
> org.apache.calcite.prepare.CalcitePrepareImpl.perform(CalcitePrepareImpl.java:997)
>  ~[calcite-core-1.10.0.2.6.4.0-91.jar
> :1.10.0.2.6.4.0-91]
> at 
> org.apache.calcite.tools.Frameworks.withPrepare(Frameworks.java:149) 
> ~[calcite-core-1.10.0.2.6.4.0-91.jar:1.10.0.2.6.4.
> 0-91]
> at 
> org.apache.calcite.tools.Frameworks.withPlanner(Frameworks.java:106) 
> ~[calcite-core-1.10.0.2.6.4.0-91.jar:1.10.0.2.6.4.
> 0-91]
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner.logicalPlan(CalcitePlanner.java:1082)
>  ~[hive-exec-2.1.0.2.6.4.0-91.jar:2
> .1.0.2.6.4.0-91]
> {code} 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (HIVE-19942) Hive Notification: All events for indexes should have table name

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519685#comment-16519685
 ] 

Hive QA commented on HIVE-19942:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928549/HIVE-19942.02-branch-2.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 16 failed/errored test(s), 10683 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestAccumuloCliDriver.testCliDriver[accumulo_queries]
 (batchId=227)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[avro_tableproperty_optimize]
 (batchId=22)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[explaindenpendencydiffengs]
 (batchId=38)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[llap_smb] 
(batchId=142)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_ppd_basic] 
(batchId=139)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[table_nonprintable]
 (batchId=140)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[join_acid_non_acid]
 (batchId=158)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[union_fast_stats]
 (batchId=153)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vectorized_parquet_types]
 (batchId=155)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[merge_negative_5]
 (batchId=88)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[explaindenpendencydiffengs]
 (batchId=115)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[vectorization_input_format_excludes]
 (batchId=117)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[vectorized_ptf] 
(batchId=125)
org.apache.hive.hcatalog.api.TestHCatClient.testTransportFailure (batchId=176)
org.apache.hive.jdbc.TestSSL.testMetastoreWithSSL (batchId=221)
org.apache.hive.service.server.TestHS2HttpServer.testContextRootUrlRewrite 
(batchId=192)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11975/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11975/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11975/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 16 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928549 - PreCommit-HIVE-Build

> Hive Notification: All events for indexes should have table name
> 
>
> Key: HIVE-19942
> URL: https://issues.apache.org/jira/browse/HIVE-19942
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Affects Versions: 2.3.2
>Reporter: Bharathkrishna Guruvayoor Murali
>Assignee: Bharathkrishna Guruvayoor Murali
>Priority: Major
> Attachments: HIVE-19942.02-branch-2.patch, HIVE-19942.1.patch
>
>
> All the events for indexes: Create Index, Alter Index, Drop Index have the 
> TBL_NAME as null. The TBL_NAME should be populated with the table on which 
> the index is created.
> This makes it easier to decide whether to process the event or not without 
> needing to parse the json message (which is a slower process).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (HIVE-19963) metadata_only_queries.q fails

2018-06-21 Thread Steve Yeom (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19963?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Yeom reassigned HIVE-19963:
-

Assignee: Steve Yeom

> metadata_only_queries.q fails
> -
>
> Key: HIVE-19963
> URL: https://issues.apache.org/jira/browse/HIVE-19963
> Project: Hive
>  Issue Type: Sub-task
>  Components: Transactions
>Affects Versions: 4.0.0
>Reporter: Steve Yeom
>Assignee: Steve Yeom
>Priority: Major
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19928:
--
Attachment: HIVE-19928.7.patch

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Attachments: HIVE-19928.1.patch, HIVE-19928.2.patch, 
> HIVE-19928.3.patch, HIVE-19928.4.patch, HIVE-19928.5.patch, 
> HIVE-19928.6.patch, HIVE-19928.7.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19951) Vectorization: Need to disable encoded LLAP I/O for ORC when there is data type conversion (Schema Evolution)

2018-06-21 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-19951:

Attachment: HIVE-19951.02.patch

> Vectorization: Need to disable encoded LLAP I/O for ORC when there is data 
> type conversion  (Schema Evolution)
> --
>
> Key: HIVE-19951
> URL: https://issues.apache.org/jira/browse/HIVE-19951
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: HIVE-19951.01.patch, HIVE-19951.02.patch
>
>
> Currently, reading encoded ORC data does not support data type conversion.  
> So, encoded reading and cache populating needs to be disabled.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19951) Vectorization: Need to disable encoded LLAP I/O for ORC when there is data type conversion (Schema Evolution)

2018-06-21 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-19951:

Status: Patch Available  (was: In Progress)

> Vectorization: Need to disable encoded LLAP I/O for ORC when there is data 
> type conversion  (Schema Evolution)
> --
>
> Key: HIVE-19951
> URL: https://issues.apache.org/jira/browse/HIVE-19951
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: HIVE-19951.01.patch, HIVE-19951.02.patch
>
>
> Currently, reading encoded ORC data does not support data type conversion.  
> So, encoded reading and cache populating needs to be disabled.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19951) Vectorization: Need to disable encoded LLAP I/O for ORC when there is data type conversion (Schema Evolution)

2018-06-21 Thread Matt McCline (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matt McCline updated HIVE-19951:

Status: In Progress  (was: Patch Available)

> Vectorization: Need to disable encoded LLAP I/O for ORC when there is data 
> type conversion  (Schema Evolution)
> --
>
> Key: HIVE-19951
> URL: https://issues.apache.org/jira/browse/HIVE-19951
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: HIVE-19951.01.patch
>
>
> Currently, reading encoded ORC data does not support data type conversion.  
> So, encoded reading and cache populating needs to be disabled.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-12192) Hive should carry out timestamp computations in UTC

2018-06-21 Thread Jesus Camacho Rodriguez (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-12192?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jesus Camacho Rodriguez updated HIVE-12192:
---
Attachment: HIVE-12192.18.patch

> Hive should carry out timestamp computations in UTC
> ---
>
> Key: HIVE-12192
> URL: https://issues.apache.org/jira/browse/HIVE-12192
> Project: Hive
>  Issue Type: Sub-task
>  Components: Hive
>Reporter: Ryan Blue
>Assignee: Jesus Camacho Rodriguez
>Priority: Blocker
>  Labels: timestamp
> Attachments: HIVE-12192.01.patch, HIVE-12192.02.patch, 
> HIVE-12192.03.patch, HIVE-12192.04.patch, HIVE-12192.05.patch, 
> HIVE-12192.06.patch, HIVE-12192.07.patch, HIVE-12192.08.patch, 
> HIVE-12192.09.patch, HIVE-12192.10.patch, HIVE-12192.11.patch, 
> HIVE-12192.12.patch, HIVE-12192.13.patch, HIVE-12192.14.patch, 
> HIVE-12192.15.patch, HIVE-12192.16.patch, HIVE-12192.17.patch, 
> HIVE-12192.18.patch, HIVE-12192.patch
>
>
> Hive currently uses the "local" time of a java.sql.Timestamp to represent the 
> SQL data type TIMESTAMP WITHOUT TIME ZONE. The purpose is to be able to use 
> {{Timestamp#getYear()}} and similar methods to implement SQL functions like 
> {{year}}.
> When the SQL session's time zone is a DST zone, such as America/Los_Angeles 
> that alternates between PST and PDT, there are times that cannot be 
> represented because the effective zone skips them.
> {code}
> hive> select TIMESTAMP '2015-03-08 02:10:00.101';
> 2015-03-08 03:10:00.101
> {code}
> Using UTC instead of the SQL session time zone as the underlying zone for a 
> java.sql.Timestamp avoids this bug, while still returning correct values for 
> {{getYear}} etc. Using UTC as the convenience representation (timestamp 
> without time zone has no real zone) would make timestamp calculations more 
> consistent and avoid similar problems in the future.
> Notably, this would break the {{unix_timestamp}} UDF that specifies the 
> result is with respect to ["the default timezone and default 
> locale"|https://cwiki.apache.org/confluence/display/Hive/LanguageManual+UDF#LanguageManualUDF-DateFunctions].
>  That function would need to be updated to use the 
> {{System.getProperty("user.timezone")}} zone.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19928) Load Data for managed tables should set the owner of loaded files to a configurable user

2018-06-21 Thread Deepak Jaiswal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Deepak Jaiswal updated HIVE-19928:
--
Attachment: HIVE-19928.6.patch

> Load Data for managed tables should set the owner of loaded files to a 
> configurable user
> 
>
> Key: HIVE-19928
> URL: https://issues.apache.org/jira/browse/HIVE-19928
> Project: Hive
>  Issue Type: Task
>Reporter: Deepak Jaiswal
>Assignee: Deepak Jaiswal
>Priority: Major
> Attachments: HIVE-19928.1.patch, HIVE-19928.2.patch, 
> HIVE-19928.3.patch, HIVE-19928.4.patch, HIVE-19928.5.patch, HIVE-19928.6.patch
>
>
> load data of managed tables should set the owner of the loaded files to a 
> configurable user. the default user should be hive.
> If the owner of existing file is not hive, then a rename/move operation 
> should be replaced by copy with the copied file having hive as owner.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (HIVE-19911) Hive delete queries fail with Invalid table alias or column reference

2018-06-21 Thread Mykhailo Kysliuk (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mykhailo Kysliuk updated HIVE-19911:

Attachment: HIVE-19911.2-branch-2.3.patch

> Hive delete queries fail with Invalid table alias or column reference
> -
>
> Key: HIVE-19911
> URL: https://issues.apache.org/jira/browse/HIVE-19911
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.3.3
>Reporter: Mykhailo Kysliuk
>Assignee: Mykhailo Kysliuk
>Priority: Major
> Attachments: HIVE-19911.1-branch-2.3.patch, 
> HIVE-19911.2-branch-2.3.patch
>
>
> Env:
> hadoop-2.7.0
> hive-2.3.3
> OS:
> centos-release-7-5.1804.el7.centos.x86_64
> Steps to reproduce (at hive cli):
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> DROP TABLE IF EXISTS detaillineitem_all;
> DROP TABLE IF EXISTS detaillineitem_all_delete_1526330755128;
> CREATE TABLE `detaillineitem_all`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3)
> ) PARTITIONED BY (
>   `branchnumber_p` varchar(3))
> CLUSTERED BY (
>   detailid)
> INTO 25 BUCKETS
> STORED AS ORC
> TBLPROPERTIES (
>   'orc.compress'='NONE',
>   'transactional'='true');
> CREATE TABLE `detaillineitem_all_delete_1526330755128`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3),
>   `branchnumber_p` varchar(3));
> DELETE from detaillineitem_all WHERE EXISTS (
> SELECT
> 1
> FROM
> detaillineitem_all_delete_1526330755128 AS t1
> WHERE
> (detaillineitem_all.detailid = t1.detailid)
>   AND
> (detaillineitem_all.branchnumber = CAST(t1.branchnumber AS STRING)));
> {code}
> Exception:
> {code}
> 2018-06-15T16:51:48,625 ERROR [f6bd86a7-04e5-4284-9031-3b9a0ccc80f3 main] 
> ql.Driver: FAILED: SemanticException Line 0:-1 Invalid table alias or column 
> reference 'sq_1': (possible column names are: mber)) sq_corr_1)) (tok_where 
> (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 0:-1 Invalid table 
> alias or column reference 'sq_1': (possible column names are: mber)) 
> sq_corr_1)) (tok_where (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11620)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11568)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11536)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11514)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapGroupByForSemijoin(SemanticAnalyzer.java:8416)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:8305)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3278)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:9592)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10549)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10427)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:11125)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:11138)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10807)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:73)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:462)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeDelete(UpdateDeleteSemanticAnalyzer.java:111)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:81)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:512)
> at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1317)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1457)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
> at 
> org.apache.hado

[jira] [Updated] (HIVE-19911) Hive delete queries fail with Invalid table alias or column reference

2018-06-21 Thread Mykhailo Kysliuk (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mykhailo Kysliuk updated HIVE-19911:

Status: Patch Available  (was: In Progress)

> Hive delete queries fail with Invalid table alias or column reference
> -
>
> Key: HIVE-19911
> URL: https://issues.apache.org/jira/browse/HIVE-19911
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.3.3
>Reporter: Mykhailo Kysliuk
>Assignee: Mykhailo Kysliuk
>Priority: Major
> Attachments: HIVE-19911.1-branch-2.3.patch, 
> HIVE-19911.2-branch-2.3.patch
>
>
> Env:
> hadoop-2.7.0
> hive-2.3.3
> OS:
> centos-release-7-5.1804.el7.centos.x86_64
> Steps to reproduce (at hive cli):
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> DROP TABLE IF EXISTS detaillineitem_all;
> DROP TABLE IF EXISTS detaillineitem_all_delete_1526330755128;
> CREATE TABLE `detaillineitem_all`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3)
> ) PARTITIONED BY (
>   `branchnumber_p` varchar(3))
> CLUSTERED BY (
>   detailid)
> INTO 25 BUCKETS
> STORED AS ORC
> TBLPROPERTIES (
>   'orc.compress'='NONE',
>   'transactional'='true');
> CREATE TABLE `detaillineitem_all_delete_1526330755128`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3),
>   `branchnumber_p` varchar(3));
> DELETE from detaillineitem_all WHERE EXISTS (
> SELECT
> 1
> FROM
> detaillineitem_all_delete_1526330755128 AS t1
> WHERE
> (detaillineitem_all.detailid = t1.detailid)
>   AND
> (detaillineitem_all.branchnumber = CAST(t1.branchnumber AS STRING)));
> {code}
> Exception:
> {code}
> 2018-06-15T16:51:48,625 ERROR [f6bd86a7-04e5-4284-9031-3b9a0ccc80f3 main] 
> ql.Driver: FAILED: SemanticException Line 0:-1 Invalid table alias or column 
> reference 'sq_1': (possible column names are: mber)) sq_corr_1)) (tok_where 
> (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 0:-1 Invalid table 
> alias or column reference 'sq_1': (possible column names are: mber)) 
> sq_corr_1)) (tok_where (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11620)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11568)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11536)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11514)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapGroupByForSemijoin(SemanticAnalyzer.java:8416)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:8305)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3278)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:9592)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10549)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10427)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:11125)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:11138)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10807)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:73)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:462)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeDelete(UpdateDeleteSemanticAnalyzer.java:111)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:81)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:512)
> at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1317)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1457)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
> at 
> org.apache.ha

[jira] [Updated] (HIVE-19911) Hive delete queries fail with Invalid table alias or column reference

2018-06-21 Thread Mykhailo Kysliuk (JIRA)


 [ 
https://issues.apache.org/jira/browse/HIVE-19911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mykhailo Kysliuk updated HIVE-19911:

Status: In Progress  (was: Patch Available)

> Hive delete queries fail with Invalid table alias or column reference
> -
>
> Key: HIVE-19911
> URL: https://issues.apache.org/jira/browse/HIVE-19911
> Project: Hive
>  Issue Type: Bug
>  Components: Transactions
>Affects Versions: 2.3.3
>Reporter: Mykhailo Kysliuk
>Assignee: Mykhailo Kysliuk
>Priority: Major
> Attachments: HIVE-19911.1-branch-2.3.patch
>
>
> Env:
> hadoop-2.7.0
> hive-2.3.3
> OS:
> centos-release-7-5.1804.el7.centos.x86_64
> Steps to reproduce (at hive cli):
> {code}
> set hive.support.concurrency=true;
> set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager;
> DROP TABLE IF EXISTS detaillineitem_all;
> DROP TABLE IF EXISTS detaillineitem_all_delete_1526330755128;
> CREATE TABLE `detaillineitem_all`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3)
> ) PARTITIONED BY (
>   `branchnumber_p` varchar(3))
> CLUSTERED BY (
>   detailid)
> INTO 25 BUCKETS
> STORED AS ORC
> TBLPROPERTIES (
>   'orc.compress'='NONE',
>   'transactional'='true');
> CREATE TABLE `detaillineitem_all_delete_1526330755128`(
>   `detailid` decimal(20,0),
>   `branchnumber` varchar(3),
>   `branchnumber_p` varchar(3));
> DELETE from detaillineitem_all WHERE EXISTS (
> SELECT
> 1
> FROM
> detaillineitem_all_delete_1526330755128 AS t1
> WHERE
> (detaillineitem_all.detailid = t1.detailid)
>   AND
> (detaillineitem_all.branchnumber = CAST(t1.branchnumber AS STRING)));
> {code}
> Exception:
> {code}
> 2018-06-15T16:51:48,625 ERROR [f6bd86a7-04e5-4284-9031-3b9a0ccc80f3 main] 
> ql.Driver: FAILED: SemanticException Line 0:-1 Invalid table alias or column 
> reference 'sq_1': (possible column names are: mber)) sq_corr_1)) (tok_where 
> (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> org.apache.hadoop.hive.ql.parse.SemanticException: Line 0:-1 Invalid table 
> alias or column reference 'sq_1': (possible column names are: mber)) 
> sq_corr_1)) (tok_where (= 1 1), (. (tok_table_or_col sq_1) sq_corr_1))
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genAllExprNodeDesc(SemanticAnalyzer.java:11620)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11568)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11536)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genExprNodeDesc(SemanticAnalyzer.java:11514)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genMapGroupByForSemijoin(SemanticAnalyzer.java:8416)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genJoinOperator(SemanticAnalyzer.java:8305)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3278)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:9592)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10549)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:10427)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:11125)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:11138)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:10807)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:73)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.reparseAndSuperAnalyze(UpdateDeleteSemanticAnalyzer.java:462)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeDelete(UpdateDeleteSemanticAnalyzer.java:111)
> at 
> org.apache.hadoop.hive.ql.parse.UpdateDeleteSemanticAnalyzer.analyzeInternal(UpdateDeleteSemanticAnalyzer.java:81)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:258)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:512)
> at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:1317)
> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1457)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLoc

[jira] [Commented] (HIVE-19951) Vectorization: Need to disable encoded LLAP I/O for ORC when there is data type conversion (Schema Evolution)

2018-06-21 Thread Hive QA (JIRA)


[ 
https://issues.apache.org/jira/browse/HIVE-19951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16519655#comment-16519655
 ] 

Hive QA commented on HIVE-19951:




Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12928531/HIVE-19951.01.patch

{color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 2 failed/errored test(s), 14541 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[orc_ppd_schema_evol_3a]
 (batchId=150)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[vector_llap_io_data_conversion]
 (batchId=164)
{noformat}

Test results: 
https://builds.apache.org/job/PreCommit-HIVE-Build/11974/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/11974/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-11974/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 2 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12928531 - PreCommit-HIVE-Build

> Vectorization: Need to disable encoded LLAP I/O for ORC when there is data 
> type conversion  (Schema Evolution)
> --
>
> Key: HIVE-19951
> URL: https://issues.apache.org/jira/browse/HIVE-19951
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Matt McCline
>Assignee: Matt McCline
>Priority: Critical
> Attachments: HIVE-19951.01.patch
>
>
> Currently, reading encoded ORC data does not support data type conversion.  
> So, encoded reading and cache populating needs to be disabled.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


  1   2   >