[jira] [Commented] (HIVE-28042) DigestMD5 token expired or does not exist error while opening a new connection to HMS

2024-01-29 Thread Butao Zhang (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812178#comment-17812178
 ] 

Butao Zhang commented on HIVE-28042:


 Just thinking aloud. :)

set 
hive.cluster.delegation.token.store.class=org.apache.hadoop.hive.metastore.security.DBTokenStore;

Did you try to use DBTokenStore to store the token info? Maybe it is better 
than the default token store MemoryTokenStore, maybe it can fix your issue?

> DigestMD5 token expired or does not exist error while opening a new 
> connection to HMS
> -
>
> Key: HIVE-28042
> URL: https://issues.apache.org/jira/browse/HIVE-28042
> Project: Hive
>  Issue Type: Bug
>Reporter: Vikram Ahuja
>Assignee: Vikram Ahuja
>Priority: Major
>
> Hello,
> In our deployment we are facing the following exception in the HMS logs when 
> a HMS connection is opened from the HS2 in cases where a session is open for 
> a long time leading to query failures:
> {code:java}
> 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
> transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
> failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
> password    
> at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
>     
> at 
> com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
>     
> at 
> org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
>     at 
> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java)    
> at 
> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
>     
> at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
>     
> at java.security.AccessController.doPrivileged(Native Method)    
> at javax.security.auth.Subject.doAs(Subject.javA)    
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
>     
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) 
>    
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java)   
>  
> at java.lang.Thread.run(Thread.java)Caused by: 
> org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
> does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
> realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
> sequenceNumber=3294063, masterKeyId=7601    
> at 
> org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
>     
> at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
>     ... 15 more {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28042) DigestMD5 token expired or does not exist error while opening a new connection to HMS

2024-01-29 Thread Vikram Ahuja (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Ahuja updated HIVE-28042:

Summary: DigestMD5 token expired or does not exist error while opening a 
new connection to HMS  (was: DigestMD5 error during opening connection to HMS)

> DigestMD5 token expired or does not exist error while opening a new 
> connection to HMS
> -
>
> Key: HIVE-28042
> URL: https://issues.apache.org/jira/browse/HIVE-28042
> Project: Hive
>  Issue Type: Bug
>Reporter: Vikram Ahuja
>Assignee: Vikram Ahuja
>Priority: Major
>
> Hello,
> In our deployment we are facing the following exception in the HMS logs when 
> a HMS connection is opened from the HS2 in cases where a session is open for 
> a long time leading to query failures:
> {code:java}
> 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
> transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
> failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
> password    
> at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
>     
> at 
> com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
>     
> at 
> org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
>     at 
> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java)    
> at 
> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
>     
> at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
>     
> at java.security.AccessController.doPrivileged(Native Method)    
> at javax.security.auth.Subject.doAs(Subject.javA)    
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
>     
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) 
>    
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java)   
>  
> at java.lang.Thread.run(Thread.java)Caused by: 
> org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
> does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
> realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
> sequenceNumber=3294063, masterKeyId=7601    
> at 
> org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
>     
> at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
>     ... 15 more {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (HIVE-28042) DigestMD5 error during opening connection to HMS

2024-01-29 Thread Vikram Ahuja (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812159#comment-17812159
 ] 

Vikram Ahuja commented on HIVE-28042:
-

Raised PR: [https://github.com/apache/hive/pull/5049] for the same

> DigestMD5 error during opening connection to HMS
> 
>
> Key: HIVE-28042
> URL: https://issues.apache.org/jira/browse/HIVE-28042
> Project: Hive
>  Issue Type: Bug
>Reporter: Vikram Ahuja
>Assignee: Vikram Ahuja
>Priority: Major
>
> Hello,
> In our deployment we are facing the following exception in the HMS logs when 
> a HMS connection is opened from the HS2 in cases where a session is open for 
> a long time leading to query failures:
> {code:java}
> 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
> transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
> failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
> password    
> at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
>     
> at 
> com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
>     
> at 
> org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
>     at 
> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java)    
> at 
> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
>     
> at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
>     
> at java.security.AccessController.doPrivileged(Native Method)    
> at javax.security.auth.Subject.doAs(Subject.javA)    
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
>     
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) 
>    
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java)   
>  
> at java.lang.Thread.run(Thread.java)Caused by: 
> org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
> does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
> realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
> sequenceNumber=3294063, masterKeyId=7601    
> at 
> org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
>     
> at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
>     
> at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
>     ... 15 more {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (HIVE-28002) Make hive.metastore.batch.retrieve.max property to be read from hiveconf when it is requested from HS2

2024-01-29 Thread Vikram Ahuja (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28002?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Ahuja resolved HIVE-28002.
-
Fix Version/s: 4.0.0-beta-1
   Resolution: Not A Problem

> Make hive.metastore.batch.retrieve.max property to be read from hiveconf when 
> it is requested from HS2
> --
>
> Key: HIVE-28002
> URL: https://issues.apache.org/jira/browse/HIVE-28002
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 4.0.0-beta-1
>Reporter: Vikram Ahuja
>Assignee: Vikram Ahuja
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0-beta-1
>
>
> The property hive.metastore.batch.retrieve.max is used by both the 
> client(Hs2) and the HMS when fetching objects(tables/partitions) in batches. 
> This property is present in both hiveConf as well as metastoreConf. For some 
> use cases in the client(Hs2), this value is referred from the metastoreConf. 
> To change these use cases to HiveConf.java to have a set demarkation between 
> HS2 and HMS use cases.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work stopped] (HIVE-28002) Make hive.metastore.batch.retrieve.max property to be read from hiveconf when it is requested from HS2

2024-01-29 Thread Vikram Ahuja (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28002?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-28002 stopped by Vikram Ahuja.
---
> Make hive.metastore.batch.retrieve.max property to be read from hiveconf when 
> it is requested from HS2
> --
>
> Key: HIVE-28002
> URL: https://issues.apache.org/jira/browse/HIVE-28002
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 4.0.0-beta-1
>Reporter: Vikram Ahuja
>Assignee: Vikram Ahuja
>Priority: Minor
>  Labels: pull-request-available
>
> The property hive.metastore.batch.retrieve.max is used by both the 
> client(Hs2) and the HMS when fetching objects(tables/partitions) in batches. 
> This property is present in both hiveConf as well as metastoreConf. For some 
> use cases in the client(Hs2), this value is referred from the metastoreConf. 
> To change these use cases to HiveConf.java to have a set demarkation between 
> HS2 and HMS use cases.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (HIVE-28042) DigestMD5 error during opening connection to HMS

2024-01-29 Thread Vikram Ahuja (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812151#comment-17812151
 ] 

Vikram Ahuja edited comment on HIVE-28042 at 1/30/24 5:45 AM:
--

*Another instance of this issue:*

 
{code:java}
2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
transport.TSaslTransport (TSaslTransport.java:open) 
- SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO 
error acquiring password
at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    
at 
com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
    
at 
org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
    
at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java)    
at 
org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
    
at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    
at java.security.AccessController.doPrivileged(Native Method)    
at javax.security.auth.Subject.doAs(Subject.javA)    
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) 
   
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
    
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
    
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java)   
 
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java)  
  
at java.lang.Thread.run(Thread.java)Caused by: 
org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
sequenceNumber=3294063, masterKeyId=7601    
at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    
at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
    at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    ... 15 more {code}
 

 

*Analysis of the issue:*

This particular issue is only happening when the HS2 tries to open a new Digest 
MD5 based Thrift TSaslClientTransport in cases where the session is open for a 
long time.

HS2 uses the same metaStoreClient object across all the connections that is 
embedded in Hive.java but in some cases we have observed that is recreating a 
new metaStoreClient with a fresh connection(TSaslClientTransport). Two use 
cases that I discovered which were leading to these issues were:
 # 
 ## MSCK repair
 ## RetryingMetaStoreClient in case of any HMS issues(applicable to any sql 
query which interacts with the HMS)

 

*Root cause of this issue:*

There is a background thread called ExpiredTokenRemover running in HMS (class: 
TokenStoreDelegationTokenSecretManager.java ). This expiry thread itself is 
removing the token from the tokenStore after the renewal time has passed and 
also removing it after expiry time, but is should only remove it post expiry 
time as the token can be renewed till then.

 

Will be raising a fix for the same by changing the code where token is deleted 
after renewal time itself has passed.


was (Author: vikramahuja_):
*Another instance of this issue:*

 
{code:java}
2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
transport.TSaslTransport (TSaslTransport.java:open) 
- SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO 
error acquiring password
at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    
at 
com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
    
at 
org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
    
at org.apache.thrift.transport.TSaslTransport

[jira] [Updated] (HIVE-28042) DigestMD5 error during opening connection to HMS

2024-01-29 Thread Vikram Ahuja (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Ahuja updated HIVE-28042:

Description: 
Hello,

In our deployment we are facing the following exception in the HMS logs when a 
HMS connection is opened from the HS2 in cases where a session is open for a 
long time leading to query failures:
{code:java}
2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
password    
at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    
at 
com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
    
at 
org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
    at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) 
   
at 
org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
    
at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    
at java.security.AccessController.doPrivileged(Native Method)    
at javax.security.auth.Subject.doAs(Subject.javA)    
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) 
   
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
    
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
    
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java)   
 
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java)  
  
at java.lang.Thread.run(Thread.java)Caused by: 
org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
sequenceNumber=3294063, masterKeyId=7601    
at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    
at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
    
at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    ... 15 more {code}
 

  was:
Hello,

In our deployment we are facing the following exception in the HMS logs when a 
HMS connection is opened from the HS2 in cases where a session is open for a 
long time leading to query failures:
{code:java}
2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
password    at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    at 
com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
    at 
org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
    at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) 
   at 
org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
    at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    at java.security.AccessController.doPrivileged(Native Method)    at 
javax.security.auth.Subject.doAs(Subject.javA)    at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) 
   at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
    at 

[jira] [Commented] (HIVE-28042) DigestMD5 error during opening connection to HMS

2024-01-29 Thread Vikram Ahuja (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812151#comment-17812151
 ] 

Vikram Ahuja commented on HIVE-28042:
-

*Another instance of this issue:*

 
{code:java}
2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
transport.TSaslTransport (TSaslTransport.java:open) 
- SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO 
error acquiring password
at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    
at 
com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
    
at 
org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
    
at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java)    
at 
org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
    
at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    
at java.security.AccessController.doPrivileged(Native Method)    
at javax.security.auth.Subject.doAs(Subject.javA)    
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) 
   
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
    
at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
    
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java)   
 
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java)  
  
at java.lang.Thread.run(Thread.java)Caused by: 
org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
sequenceNumber=3294063, masterKeyId=7601    
at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    
at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
    
at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
    at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    ... 15 more {code}
 

 

*Analysis of the issue:*

This particular issue is only happening when the HS2 tries to open a new Digest 
MD5 based Thrift TSaslClientTransport in cases where the session is one for a 
long time.

HS2 uses the same metaStoreClient object across all the connections that is 
embedded in Hive.java but in some cases we have observed that is recreating a 
new metaStoreClient with a fresh connection(TSaslClientTransport). Two use 
cases that I discovered which were leading to these issues were:
 # 
 ## MSCK repair
 ## RetryingMetaStoreClient in case of any HMS issues(applicable to any sql 
query which interacts with the HMS)

 

*Root cause of this issue:*

There is a background thread called ExpiredTokenRemover running in HMS (class:  
TokenStoreDelegationTokenSecretManager.java ). This expiry thread itself is 
removing the token from the tokenStore after the renewal time has passed and 
also removing it after expiry time, but is should only remove it post expiry 
time as the token can be renewed till then.

 

Will be raising a fix for the same by changing the code where token is deleted 
after renewal time itself has passed.

> DigestMD5 error during opening connection to HMS
> 
>
> Key: HIVE-28042
> URL: https://issues.apache.org/jira/browse/HIVE-28042
> Project: Hive
>  Issue Type: Bug
>Reporter: Vikram Ahuja
>Assignee: Vikram Ahuja
>Priority: Major
>
> Hello,
> In our deployment we are facing the following exception in the HMS logs when 
> a HMS connection is opened from the HS2 in cases where a session is open for 
> a long time leading to query failures:
> {code:java}
> 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
> transport.TSaslTransport (TSaslTransport.java:open) 

[jira] [Updated] (HIVE-24167) TPC-DS query 14 fails while generating plan for the filter

2024-01-29 Thread okumin (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-24167?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

okumin updated HIVE-24167:
--
Status: Patch Available  (was: Open)

I implemented one of the approaches in my mind.

[https://github.com/apache/hive/pull/5037]

 

Also, I wrote down what I've investigated to resolve this issue.

https://gist.github.com/okumin/b111fe0a911507bdf6a7204f49b9cb72

> TPC-DS query 14 fails while generating plan for the filter
> --
>
> Key: HIVE-24167
> URL: https://issues.apache.org/jira/browse/HIVE-24167
> Project: Hive
>  Issue Type: Sub-task
>  Components: CBO
>Reporter: Stamatis Zampetakis
>Assignee: okumin
>Priority: Major
>  Labels: hive-4.1.0-must, pull-request-available
>
> TPC-DS query 14 (cbo_query14.q and query4.q) fail with NPE on the metastore 
> with the partitioned TPC-DS 30TB dataset while generating the plan for the 
> filter.
> The problem can be reproduced using the PR in HIVE-23965.
> The current stacktrace shows that the NPE appears while trying to display the 
> debug message but even if this line didn't exist it would fail again later on.
> {noformat}
> java.lang.NullPointerException
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:10867)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11765)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11622)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11649)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11622)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11649)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11635)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlanForSubQueryPredicate(SemanticAnalyzer.java:3375)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3473)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:10819)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11765)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11622)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11625)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11625)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11649)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11622)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11649)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11635)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:12417)
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner.genOPTree(CalcitePlanner.java:718)
> at 
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:12519)
> at 
> org.apache.hadoop.hive.ql.parse.CalcitePlanner.analyzeInternal(CalcitePlanner.java:443)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:301)
> at 
> org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:171)
> at 
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:301)
> at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220)
> at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:173)
> at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:414)
> at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:363)
> at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:357)
> at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.compileAndRespond(ReExecDriver.java:129)
> at 
> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:231)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:258)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processCmd1(CliDriver.java:203)
> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:129)
> at 
> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:424)
> at

[jira] [Updated] (HIVE-27938) Iceberg: Fix java.lang.ClassCastException during vectorized reads on partition columns

2024-01-29 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27938?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-27938:
--
Labels: pull-request-available  (was: )

> Iceberg: Fix java.lang.ClassCastException during vectorized reads on 
> partition columns 
> ---
>
> Key: HIVE-27938
> URL: https://issues.apache.org/jira/browse/HIVE-27938
> Project: Hive
>  Issue Type: Bug
>Reporter: Simhadri Govindappa
>Assignee: Simhadri Govindappa
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> 1: jdbc:hive2://localhost:10001/> CREATE EXTERNAL TABLE ice3   (`col1` int, 
> `calday` date) PARTITIONED BY SPEC (calday)   stored by iceberg 
> tblproperties('format-version'='2'); 
> 1: jdbc:hive2://localhost:10001/>insert into ice3 values(1, '2020-11-20'); 
> 1: jdbc:hive2://localhost:10001/> select count(calday) from ice3;
> {code}
> Full stack trace: 
> {code:java}
> INFO  : Compiling 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): 
> select count(calday) from ice3INFO  : No Stats for default@ice3, Columns: 
> caldayINFO  : Semantic Analysis Completed (retrial = false)INFO  : Created 
> Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, type:bigint, 
> comment:null)], properties:null)INFO  : Completed compiling 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab); 
> Time taken: 0.196 secondsINFO  : Operation QUERY obtained 0 locksINFO  : 
> Executing 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): 
> select count(calday) from ice3INFO  : Query ID = 
> root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO  : Total jobs = 
> 1INFO  : Launching Job 1 out of 1INFO  : Starting task [Stage-1:MAPRED] in 
> serial modeINFO  : Subscribed to counters: [] for queryId: 
> root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO  : Session is 
> already openINFO  : Dag name: select count(calday) from ice3 (Stage-1)INFO  : 
> HS2 Host: [localhost], Query ID: 
> [root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab], Dag ID: 
> [dag_1701888162260_0001_2], DAG Session ID: 
> [application_1701888162260_0001]INFO  : Status: Running (Executing on YARN 
> cluster with App id application_1701888162260_0001)
> --
>         VERTICES      MODE        STATUS  TOTAL  COMPLETED  RUNNING  PENDING  
> FAILED  
> KILLED--Map
>  1            container       RUNNING      1          0        0        1     
>   4       0Reducer 2        container        INITED      1          0        
> 0        1       0       
> 0--VERTICES:
>  00/02  [>>--] 0%    ELAPSED TIME: 1.41 
> s--ERROR
>  : Status: FailedERROR : Vertex failed, vertexName=Map 1, 
> vertexId=vertex_1701888162260_0001_2_00, diagnostics=[Task failed, 
> taskId=task_1701888162260_0001_2_00_00, diagnostics=[TaskAttempt 0 
> failed, info=[Error: Error while running task ( failure ) : 
> attempt_1701888162260_0001_2_00_00_0:java.lang.RuntimeException: 
> org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: 
> java.lang.ClassCastException: java.time.LocalDate cannot be cast to 
> org.apache.hadoop.hive.common.type.Dateat 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:348)
>  at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:276)   
> at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:381)
>  at 
> org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:82)
>at 
> org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:69)
>at java.security.AccessController.doPrivileged(Native Method)   at 
> javax.security.auth.Subject.doAs(Subject.java:422)   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1878)
>  at 
> org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:69)
> at 
> org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:39)
> at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36)  
> at 
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:111)
>   at 
> com.google.common.util.concurrent.InterruptibleTask.run(Inte

[jira] [Updated] (HIVE-27938) Iceberg: Fix java.lang.ClassCastException during vectorized reads on partition columns

2024-01-29 Thread Simhadri Govindappa (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27938?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simhadri Govindappa updated HIVE-27938:
---
Summary: Iceberg: Fix java.lang.ClassCastException during vectorized reads 
on partition columns   (was: Iceberg: Date type Partitioned column throws 
java.lang.ClassCastException: java.time.LocalDate cannot be cast to 
org.apache.hadoop.hive.common.type.Date)

> Iceberg: Fix java.lang.ClassCastException during vectorized reads on 
> partition columns 
> ---
>
> Key: HIVE-27938
> URL: https://issues.apache.org/jira/browse/HIVE-27938
> Project: Hive
>  Issue Type: Bug
>Reporter: Simhadri Govindappa
>Assignee: Simhadri Govindappa
>Priority: Major
>
> {code:java}
> 1: jdbc:hive2://localhost:10001/> CREATE EXTERNAL TABLE ice3   (`col1` int, 
> `calday` date) PARTITIONED BY SPEC (calday)   stored by iceberg 
> tblproperties('format-version'='2'); 
> 1: jdbc:hive2://localhost:10001/>insert into ice3 values(1, '2020-11-20'); 
> 1: jdbc:hive2://localhost:10001/> select count(calday) from ice3;
> {code}
> Full stack trace: 
> {code:java}
> INFO  : Compiling 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): 
> select count(calday) from ice3INFO  : No Stats for default@ice3, Columns: 
> caldayINFO  : Semantic Analysis Completed (retrial = false)INFO  : Created 
> Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, type:bigint, 
> comment:null)], properties:null)INFO  : Completed compiling 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab); 
> Time taken: 0.196 secondsINFO  : Operation QUERY obtained 0 locksINFO  : 
> Executing 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): 
> select count(calday) from ice3INFO  : Query ID = 
> root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO  : Total jobs = 
> 1INFO  : Launching Job 1 out of 1INFO  : Starting task [Stage-1:MAPRED] in 
> serial modeINFO  : Subscribed to counters: [] for queryId: 
> root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO  : Session is 
> already openINFO  : Dag name: select count(calday) from ice3 (Stage-1)INFO  : 
> HS2 Host: [localhost], Query ID: 
> [root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab], Dag ID: 
> [dag_1701888162260_0001_2], DAG Session ID: 
> [application_1701888162260_0001]INFO  : Status: Running (Executing on YARN 
> cluster with App id application_1701888162260_0001)
> --
>         VERTICES      MODE        STATUS  TOTAL  COMPLETED  RUNNING  PENDING  
> FAILED  
> KILLED--Map
>  1            container       RUNNING      1          0        0        1     
>   4       0Reducer 2        container        INITED      1          0        
> 0        1       0       
> 0--VERTICES:
>  00/02  [>>--] 0%    ELAPSED TIME: 1.41 
> s--ERROR
>  : Status: FailedERROR : Vertex failed, vertexName=Map 1, 
> vertexId=vertex_1701888162260_0001_2_00, diagnostics=[Task failed, 
> taskId=task_1701888162260_0001_2_00_00, diagnostics=[TaskAttempt 0 
> failed, info=[Error: Error while running task ( failure ) : 
> attempt_1701888162260_0001_2_00_00_0:java.lang.RuntimeException: 
> org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: 
> java.lang.ClassCastException: java.time.LocalDate cannot be cast to 
> org.apache.hadoop.hive.common.type.Dateat 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:348)
>  at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:276)   
> at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:381)
>  at 
> org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:82)
>at 
> org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:69)
>at java.security.AccessController.doPrivileged(Native Method)   at 
> javax.security.auth.Subject.doAs(Subject.java:422)   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1878)
>  at 
> org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:69)
> at 
> org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:39)
> at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36)  
> at 
> com.google.common.util.concurrent

[jira] [Commented] (HIVE-27938) Iceberg: Date type Partitioned column throws java.lang.ClassCastException: java.time.LocalDate cannot be cast to org.apache.hadoop.hive.common.type.Date

2024-01-29 Thread Simhadri Govindappa (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-27938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812108#comment-17812108
 ] 

Simhadri Govindappa commented on HIVE-27938:


The error is also present for DATE and DECIMAL columns

> Iceberg: Date type Partitioned column throws java.lang.ClassCastException: 
> java.time.LocalDate cannot be cast to org.apache.hadoop.hive.common.type.Date
> 
>
> Key: HIVE-27938
> URL: https://issues.apache.org/jira/browse/HIVE-27938
> Project: Hive
>  Issue Type: Bug
>Reporter: Simhadri Govindappa
>Assignee: Simhadri Govindappa
>Priority: Major
>
> {code:java}
> 1: jdbc:hive2://localhost:10001/> CREATE EXTERNAL TABLE ice3   (`col1` int, 
> `calday` date) PARTITIONED BY SPEC (calday)   stored by iceberg 
> tblproperties('format-version'='2'); 
> 1: jdbc:hive2://localhost:10001/>insert into ice3 values(1, '2020-11-20'); 
> 1: jdbc:hive2://localhost:10001/> select count(calday) from ice3;
> {code}
> Full stack trace: 
> {code:java}
> INFO  : Compiling 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): 
> select count(calday) from ice3INFO  : No Stats for default@ice3, Columns: 
> caldayINFO  : Semantic Analysis Completed (retrial = false)INFO  : Created 
> Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, type:bigint, 
> comment:null)], properties:null)INFO  : Completed compiling 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab); 
> Time taken: 0.196 secondsINFO  : Operation QUERY obtained 0 locksINFO  : 
> Executing 
> command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): 
> select count(calday) from ice3INFO  : Query ID = 
> root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO  : Total jobs = 
> 1INFO  : Launching Job 1 out of 1INFO  : Starting task [Stage-1:MAPRED] in 
> serial modeINFO  : Subscribed to counters: [] for queryId: 
> root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO  : Session is 
> already openINFO  : Dag name: select count(calday) from ice3 (Stage-1)INFO  : 
> HS2 Host: [localhost], Query ID: 
> [root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab], Dag ID: 
> [dag_1701888162260_0001_2], DAG Session ID: 
> [application_1701888162260_0001]INFO  : Status: Running (Executing on YARN 
> cluster with App id application_1701888162260_0001)
> --
>         VERTICES      MODE        STATUS  TOTAL  COMPLETED  RUNNING  PENDING  
> FAILED  
> KILLED--Map
>  1            container       RUNNING      1          0        0        1     
>   4       0Reducer 2        container        INITED      1          0        
> 0        1       0       
> 0--VERTICES:
>  00/02  [>>--] 0%    ELAPSED TIME: 1.41 
> s--ERROR
>  : Status: FailedERROR : Vertex failed, vertexName=Map 1, 
> vertexId=vertex_1701888162260_0001_2_00, diagnostics=[Task failed, 
> taskId=task_1701888162260_0001_2_00_00, diagnostics=[TaskAttempt 0 
> failed, info=[Error: Error while running task ( failure ) : 
> attempt_1701888162260_0001_2_00_00_0:java.lang.RuntimeException: 
> org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: 
> java.lang.ClassCastException: java.time.LocalDate cannot be cast to 
> org.apache.hadoop.hive.common.type.Dateat 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:348)
>  at 
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:276)   
> at 
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:381)
>  at 
> org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:82)
>at 
> org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:69)
>at java.security.AccessController.doPrivileged(Native Method)   at 
> javax.security.auth.Subject.doAs(Subject.java:422)   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1878)
>  at 
> org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:69)
> at 
> org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:39)
> at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36)  
> at 
> com.google.common.util.concurrent.TrustedListenableFutu

[jira] [Assigned] (HIVE-4752) Add support for hs2 api to use thrift over http

2024-01-29 Thread Vaibhav Gumashta (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-4752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vaibhav Gumashta reassigned HIVE-4752:
--

Assignee: Vaibhav Gumashta

> Add support for hs2 api to use thrift over http
> ---
>
> Key: HIVE-4752
> URL: https://issues.apache.org/jira/browse/HIVE-4752
> Project: Hive
>  Issue Type: New Feature
>  Components: HiveServer2
>Affects Versions: 0.13.0
>Reporter: Thejas Nair
>Assignee: Vaibhav Gumashta
>Priority: Major
>
> Hiveserver2 acts as service on the cluster for external applications. One way 
> to implement access control to services on a hadoop cluster to have a gateway 
> server authorizes service requests before forwarding them to the server. The 
> [knox project | http://wiki.apache.org/incubator/knox] has taken this 
> approach to simplify cluster security management.
> Other services on hadoop cluster such as webhdfs and webhcat already use 
> HTTP. Having hiveserver2 also support thrift over http transport will enable 
> securing hiveserver2 as well using the same approach.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28015) Iceberg: Add identifier-field-ids support in Hive

2024-01-29 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-28015:
--
Labels: pull-request-available  (was: )

> Iceberg: Add identifier-field-ids support in Hive
> -
>
> Key: HIVE-28015
> URL: https://issues.apache.org/jira/browse/HIVE-28015
> Project: Hive
>  Issue Type: Improvement
>  Components: Iceberg integration
>Affects Versions: 4.0.0
>Reporter: Denys Kuzmenko
>Assignee: Butao Zhang
>Priority: Major
>  Labels: pull-request-available
>
> Some writer engines require primary keys on a table so that they can use them 
> for writing equality deletes (only the PK cols are written to the eq-delete 
> files).
> Hive currently doesn't reject setting PKs for Iceberg tables, however, it 
> just ignores them. This succeeds:
> {code:java}
> create table ice_pk (i int, j int, primary key(i)) stored by iceberg;
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28045) Alter table without cascade seems to change metadata on existing partitions (Avro, ORC)

2024-01-29 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-28045:

Description: 
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched, like below:
{code}
aaa bbb NULL
{code}

the same applies to Avro, if there is no external schema defined (which is 
table level one)

repro: 
https://github.com/abstractdog/hive/commit/7ef97479796#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5

  was:
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched, like below:
{code}
aaa bbb NULL
{code}

the same applies to Avro, if there is no external schema defined (which is 
table level one)

repro: 
https://github.com/abstractdog/hive/commit/e0de2c52de288cafbba3bcec3533bc78890e22f9#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5


> Alter table without cascade seems to change metadata on existing partitions 
> (Avro, ORC)
> ---
>
> Key: HIVE-28045
> URL: https://issues.apache.org/jira/browse/HIVE-28045
> Project: Hive
>  Issue Type: Bug
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>
> {code}
> create external table test_orc (a string, b string) partitioned by (pt 
> string) STORED AS ORC;
> insert into test_orc PARTITION(pt='') values ("","");
> alter table test_orc add columns (c string); -- no CASCADE
> insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
> "ccc");
> select * from test_orc;
> {code}
> returns
> {code}
> aaa   bbb ccc 
> {code}
> which is unexpected without cascading the alter table, so the existing 
> partition is supposed to be untouched, like below:
> {code}
> aaa   bbb NULL
> {code}
> the same applies to Avro, if there is no external schema defined (which is 
> table level one)
> repro: 
> https://github.com/abstractdog/hive/commit/7ef97479796#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28045) Alter table without cascade seems to change metadata on existing partitions (Avro, ORC)

2024-01-29 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-28045:

Description: 
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched, like below:
{code}
aaa bbb NULL
{code}

the same applies to Avro, if there is no external schema defined (which is 
table level one)

repro: 
https://github.com/abstractdog/hive/commit/e0de2c52de288cafbba3bcec3533bc78890e22f9#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5

  was:
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched, like below:
{code}
aaa bbb NULL
{code}

the same applies to Avro, if there is no external schema defined (which is 
table level one)


> Alter table without cascade seems to change metadata on existing partitions 
> (Avro, ORC)
> ---
>
> Key: HIVE-28045
> URL: https://issues.apache.org/jira/browse/HIVE-28045
> Project: Hive
>  Issue Type: Bug
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>
> {code}
> create external table test_orc (a string, b string) partitioned by (pt 
> string) STORED AS ORC;
> insert into test_orc PARTITION(pt='') values ("","");
> alter table test_orc add columns (c string); -- no CASCADE
> insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
> "ccc");
> select * from test_orc;
> {code}
> returns
> {code}
> aaa   bbb ccc 
> {code}
> which is unexpected without cascading the alter table, so the existing 
> partition is supposed to be untouched, like below:
> {code}
> aaa   bbb NULL
> {code}
> the same applies to Avro, if there is no external schema defined (which is 
> table level one)
> repro: 
> https://github.com/abstractdog/hive/commit/e0de2c52de288cafbba3bcec3533bc78890e22f9#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28045) Alter table without cascade seems to change metadata on existing partitions (Avro, ORC)

2024-01-29 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-28045:

Description: 
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched, like below:
{code}
aaa bbb NULL
{code}

the same applies to Avro, if there is no external schema defined (which is 
table level one)

  was:
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched, like below:
{code}
aaa bbb NULL
{code}


> Alter table without cascade seems to change metadata on existing partitions 
> (Avro, ORC)
> ---
>
> Key: HIVE-28045
> URL: https://issues.apache.org/jira/browse/HIVE-28045
> Project: Hive
>  Issue Type: Bug
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>
> {code}
> create external table test_orc (a string, b string) partitioned by (pt 
> string) STORED AS ORC;
> insert into test_orc PARTITION(pt='') values ("","");
> alter table test_orc add columns (c string); -- no CASCADE
> insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
> "ccc");
> select * from test_orc;
> {code}
> returns
> {code}
> aaa   bbb ccc 
> {code}
> which is unexpected without cascading the alter table, so the existing 
> partition is supposed to be untouched, like below:
> {code}
> aaa   bbb NULL
> {code}
> the same applies to Avro, if there is no external schema defined (which is 
> table level one)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28045) Alter table without cascade seems to change metadata on existing partitions (Avro, ORC)

2024-01-29 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-28045:

Summary: Alter table without cascade seems to change metadata on existing 
partitions (Avro, ORC)  (was: Alter table without cascade seems to metadata on 
existing partitions (Avro, ORC))

> Alter table without cascade seems to change metadata on existing partitions 
> (Avro, ORC)
> ---
>
> Key: HIVE-28045
> URL: https://issues.apache.org/jira/browse/HIVE-28045
> Project: Hive
>  Issue Type: Bug
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>
> {code}
> create external table test_orc (a string, b string) partitioned by (pt 
> string) STORED AS ORC;
> insert into test_orc PARTITION(pt='') values ("","");
> alter table test_orc add columns (c string); -- no CASCADE
> insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
> "ccc");
> select * from test_orc;
> {code}
> returns
> {code}
> aaa   bbb ccc 
> {code}
> which is unexpected without cascading the alter table, so the existing 
> partition is supposed to be untouched, like below:
> {code}
> aaa   bbb NULL
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28045) Alter table without cascade seems to metadata on existing partitions (Avro, ORC)

2024-01-29 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-28045:

Description: 
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched, like below:
{code}
aaa bbb NULL
{code}

  was:
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched:
{code}
aaa bbb NULL
{code}


> Alter table without cascade seems to metadata on existing partitions (Avro, 
> ORC)
> 
>
> Key: HIVE-28045
> URL: https://issues.apache.org/jira/browse/HIVE-28045
> Project: Hive
>  Issue Type: Bug
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>
> {code}
> create external table test_orc (a string, b string) partitioned by (pt 
> string) STORED AS ORC;
> insert into test_orc PARTITION(pt='') values ("","");
> alter table test_orc add columns (c string); -- no CASCADE
> insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
> "ccc");
> select * from test_orc;
> {code}
> returns
> {code}
> aaa   bbb ccc 
> {code}
> which is unexpected without cascading the alter table, so the existing 
> partition is supposed to be untouched, like below:
> {code}
> aaa   bbb NULL
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28045) Alter table without cascade seems to metadata on existing partitions (Avro, ORC)

2024-01-29 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor updated HIVE-28045:

Description: 
{code}
create external table test_orc (a string, b string) partitioned by (pt string) 
STORED AS ORC;
insert into test_orc PARTITION(pt='') values ("","");

alter table test_orc add columns (c string); -- no CASCADE

insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
"ccc");
select * from test_orc;
{code}

returns
{code}
aaa bbb ccc 
{code}
which is unexpected without cascading the alter table, so the existing 
partition is supposed to be untouched:
{code}
aaa bbb NULL
{code}

> Alter table without cascade seems to metadata on existing partitions (Avro, 
> ORC)
> 
>
> Key: HIVE-28045
> URL: https://issues.apache.org/jira/browse/HIVE-28045
> Project: Hive
>  Issue Type: Bug
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>
> {code}
> create external table test_orc (a string, b string) partitioned by (pt 
> string) STORED AS ORC;
> insert into test_orc PARTITION(pt='') values ("","");
> alter table test_orc add columns (c string); -- no CASCADE
> insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", 
> "ccc");
> select * from test_orc;
> {code}
> returns
> {code}
> aaa   bbb ccc 
> {code}
> which is unexpected without cascading the alter table, so the existing 
> partition is supposed to be untouched:
> {code}
> aaa   bbb NULL
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (HIVE-28045) Alter table without cascade seems to metadata on existing partitions (Avro, ORC)

2024-01-29 Thread Jira
László Bodor created HIVE-28045:
---

 Summary: Alter table without cascade seems to metadata on existing 
partitions (Avro, ORC)
 Key: HIVE-28045
 URL: https://issues.apache.org/jira/browse/HIVE-28045
 Project: Hive
  Issue Type: Bug
Reporter: László Bodor






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (HIVE-28045) Alter table without cascade seems to metadata on existing partitions (Avro, ORC)

2024-01-29 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Bodor reassigned HIVE-28045:
---

Assignee: László Bodor

> Alter table without cascade seems to metadata on existing partitions (Avro, 
> ORC)
> 
>
> Key: HIVE-28045
> URL: https://issues.apache.org/jira/browse/HIVE-28045
> Project: Hive
>  Issue Type: Bug
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work started] (HIVE-27972) Set 'tez' as default value in hive.execution.engine

2024-01-29 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-27972?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-27972 started by László Bodor.
---
> Set 'tez' as default value in hive.execution.engine
> ---
>
> Key: HIVE-27972
> URL: https://issues.apache.org/jira/browse/HIVE-27972
> Project: Hive
>  Issue Type: Sub-task
>Reporter: László Bodor
>Assignee: László Bodor
>Priority: Major
>  Labels: pull-request-available
>
> Maybe this is not the first ticket addressing this, please link if it's a 
> duplicate.
> We need to set this to 'tez' to reflect that we have deprecated 'mr':
> https://github.com/apache/hive/blob/bd16e0098916aa5fc2dede99492c6a240b51e677/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java#L4567
> I'm expecting lots of UT failures because of this, as we're still running 
> those on mr (which might be fine where the actual unit test is not closely 
> related to the execution engine), so we'll see what to do.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1

2024-01-29 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-28043:
--
Labels: pull-request-available  (was: )

> Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1
> ---
>
> Key: HIVE-28043
> URL: https://issues.apache.org/jira/browse/HIVE-28043
> Project: Hive
>  Issue Type: Task
>Reporter: Anmol Sundaram
>Priority: Major
>  Labels: pull-request-available
>
> The recent stable version of ZK is 3.9.1
> We should see if we can have a similar upgrade for Hive as well
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1

2024-01-29 Thread Anmol Sundaram (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anmol Sundaram updated HIVE-28043:
--
Description: 
The recent stable version of ZK is 3.9.1

We should see if we can have a similar upgrade for Hive as well

 

  was:
Hadoop upgraded their Zookeeper versions to 3.8.3. See 
[HADOOP-18613|https://issues.apache.org/jira/browse/HADOOP-18613]

We should see if we can have a similar upgrade for Hive as well

 


> Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1
> ---
>
> Key: HIVE-28043
> URL: https://issues.apache.org/jira/browse/HIVE-28043
> Project: Hive
>  Issue Type: Task
>Reporter: Anmol Sundaram
>Priority: Major
>
> The recent stable version of ZK is 3.9.1
> We should see if we can have a similar upgrade for Hive as well
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1

2024-01-29 Thread Anmol Sundaram (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anmol Sundaram updated HIVE-28043:
--
Summary: Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1  (was: 
Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3)

> Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1
> ---
>
> Key: HIVE-28043
> URL: https://issues.apache.org/jira/browse/HIVE-28043
> Project: Hive
>  Issue Type: Task
>Reporter: Anmol Sundaram
>Priority: Major
>
> Hadoop upgraded their Zookeeper versions to 3.8.3. See 
> [HADOOP-18613|https://issues.apache.org/jira/browse/HADOOP-18613]
> We should see if we can have a similar upgrade for Hive as well
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (HIVE-28044) Replication tests to run on Tez

2024-01-29 Thread Jira
László Bodor created HIVE-28044:
---

 Summary: Replication tests to run on Tez
 Key: HIVE-28044
 URL: https://issues.apache.org/jira/browse/HIVE-28044
 Project: Hive
  Issue Type: Sub-task
Reporter: László Bodor






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28042) DigestMD5 error during opening connection to HMS

2024-01-29 Thread Vikram Ahuja (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vikram Ahuja updated HIVE-28042:

Description: 
Hello,

In our deployment we are facing the following exception in the HMS logs when a 
HMS connection is opened from the HS2 in cases where a session is open for a 
long time leading to query failures:
{code:java}
2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
password    at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    at 
com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
    at 
org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
    at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) 
   at 
org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
    at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    at java.security.AccessController.doPrivileged(Native Method)    at 
javax.security.auth.Subject.doAs(Subject.javA)    at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) 
   at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
    at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
    at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java)  
  at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) 
   at java.lang.Thread.run(Thread.java)Caused by: 
org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
sequenceNumber=3294063, masterKeyId=7601    at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
    at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    ... 15 more {code}
 

  was:
Hello,

In our deployment we are facing the following exception in the HMS logs when a 
HMS connection is opened from the HS2 in some cases leading to query failures:
{code:java}
2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
password    at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    at 
com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
    at 
org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
    at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) 
   at 
org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
    at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    at java.security.AccessController.doPrivileged(Native Method)    at 
javax.security.auth.Subject.doAs(Subject.javA)    at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) 
   at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
    at 
org.apache.thrift.server.TThreadPoolServer$WorkerProc

[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3

2024-01-29 Thread Anmol Sundaram (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anmol Sundaram updated HIVE-28043:
--
Description: 
Hadoop upgraded their Zookeeper versions to 3.8.3. See 
[HADOOP-18613|https://issues.apache.org/jira/browse/HADOOP-18613]

We should see if we can have a similar upgrade for Hive as well

 

  was:
Hadoop upgraded their Zookeeper versions to 3.8.3

We should see if we can have a similar upgrade for Hive as well

 


> Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3
> ---
>
> Key: HIVE-28043
> URL: https://issues.apache.org/jira/browse/HIVE-28043
> Project: Hive
>  Issue Type: Task
>Reporter: Anmol Sundaram
>Priority: Major
>
> Hadoop upgraded their Zookeeper versions to 3.8.3. See 
> [HADOOP-18613|https://issues.apache.org/jira/browse/HADOOP-18613]
> We should see if we can have a similar upgrade for Hive as well
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3

2024-01-29 Thread Anmol Sundaram (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anmol Sundaram updated HIVE-28043:
--
Description: 
Hadoop upgraded their Zookeeper versions to 3.8.3

We should see if we can have a similar upgrade for Hive as well

 

> Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3
> ---
>
> Key: HIVE-28043
> URL: https://issues.apache.org/jira/browse/HIVE-28043
> Project: Hive
>  Issue Type: Task
>Reporter: Anmol Sundaram
>Priority: Major
>
> Hadoop upgraded their Zookeeper versions to 3.8.3
> We should see if we can have a similar upgrade for Hive as well
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (HIVE-27929) Run TPC-DS queries and validate results correctness

2024-01-29 Thread Simhadri Govindappa (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-27929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17811790#comment-17811790
 ] 

Simhadri Govindappa edited comment on HIVE-27929 at 1/29/24 9:36 AM:
-

I was able to run a 1tb tpcds run hive master, with the following versions: 
 # Hive - master (last commit from 9th of jan)
 # Hadoop- 3.3.6
 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from 
classpath)

 

With these versions,
 * *ORC External:* I was able to run all the tpcds queries successfully. 
 * *ORC manager:* Faced the same issue described above. 


was (Author: simhadri-g):
I was able to run a 1tb tpcds run hive master, with the following versions: 
 # Hive - master (last commit from 9th of jan)
 # Hadoop- 3.3.6
 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from 
classpath)

 

With these versions,
 * *ORC External:* I was able to run all the tpcds queries successfully. 
 * *ORC manager:* Faced the same issue described above. ( HIVE-28004 )

> Run TPC-DS queries and validate results correctness
> ---
>
> Key: HIVE-27929
> URL: https://issues.apache.org/jira/browse/HIVE-27929
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Denys Kuzmenko
>Assignee: Simhadri Govindappa
>Priority: Major
>
> release branch: *branch-4.0*
> https://github.com/apache/hive/tree/branch-4.0



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (HIVE-27929) Run TPC-DS queries and validate results correctness

2024-01-29 Thread Simhadri Govindappa (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-27929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17811790#comment-17811790
 ] 

Simhadri Govindappa edited comment on HIVE-27929 at 1/29/24 9:35 AM:
-

I was able to run a 1tb tpcds run hive master, with the following versions: 
 # Hive - master (last commit from 9th of jan)
 # Hadoop- 3.3.6
 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from 
classpath)

 

With these versions,
 * *ORC External:* I was able to run all the tpcds queries successfully. 
 * *ORC manager:* Faced the same issue described above. ( HIVE-28004 )


was (Author: simhadri-g):
I was able to run a 1tb tpcds run hive master, with the following versions: 
 # Hive - master (last commit from 9th of jan)
 # Hadoop- 3.3.6
 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from 
classpath)

 

With these versions,
 * *ORC External:* I was able to run all the tpcds queries successfully. 
 * *ORC manager:* Faced the same issue described above. ( HIVE-28004 )

 

 

 

> Run TPC-DS queries and validate results correctness
> ---
>
> Key: HIVE-27929
> URL: https://issues.apache.org/jira/browse/HIVE-27929
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Denys Kuzmenko
>Assignee: Simhadri Govindappa
>Priority: Major
>
> release branch: *branch-4.0*
> https://github.com/apache/hive/tree/branch-4.0



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (HIVE-27929) Run TPC-DS queries and validate results correctness

2024-01-29 Thread Simhadri Govindappa (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-27929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17811790#comment-17811790
 ] 

Simhadri Govindappa commented on HIVE-27929:


I was able to run a 1tb tpcds run hive master, with the following versions: 
 # Hive - master (last commit from 9th of jan)
 # Hadoop- 3.3.6
 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from 
classpath)

 

With these versions,
 * *ORC External:* I was able to run all the tpcds queries successfully. 
 * *ORC manager:* Faced the same issue described above. ( HIVE-28004 )

 

 

 

> Run TPC-DS queries and validate results correctness
> ---
>
> Key: HIVE-27929
> URL: https://issues.apache.org/jira/browse/HIVE-27929
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Denys Kuzmenko
>Assignee: Simhadri Govindappa
>Priority: Major
>
> release branch: *branch-4.0*
> https://github.com/apache/hive/tree/branch-4.0



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3

2024-01-29 Thread Anmol Sundaram (Jira)
Anmol Sundaram created HIVE-28043:
-

 Summary: Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3
 Key: HIVE-28043
 URL: https://issues.apache.org/jira/browse/HIVE-28043
 Project: Hive
  Issue Type: Task
Reporter: Anmol Sundaram






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28041) Upgrade jetty to 9.4.53 due to CVE-2023-44487

2024-01-29 Thread Anmol Sundaram (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28041?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anmol Sundaram updated HIVE-28041:
--
Description: 
At the time of raising this, Hadoop also uses Jetty 9.4.53

Ref - https://github.com/apache/hadoop/blob/trunk/hadoop-project/pom.xml#L40

> Upgrade jetty to 9.4.53 due to CVE-2023-44487
> -
>
> Key: HIVE-28041
> URL: https://issues.apache.org/jira/browse/HIVE-28041
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Anmol Sundaram
>Priority: Major
>  Labels: pull-request-available
>
> At the time of raising this, Hadoop also uses Jetty 9.4.53
> Ref - https://github.com/apache/hadoop/blob/trunk/hadoop-project/pom.xml#L40



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28040) Upgrade netty to 4.1.100 due to CVE-2023-44487

2024-01-29 Thread Anmol Sundaram (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anmol Sundaram updated HIVE-28040:
--
Description: 
At the time of raising this, Hadoop and Tez also use 4.1.100.Final in their 
trunk

Related - https://issues.apache.org/jira/browse/TEZ-4538

> Upgrade netty to 4.1.100 due to CVE-2023-44487
> --
>
> Key: HIVE-28040
> URL: https://issues.apache.org/jira/browse/HIVE-28040
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Anmol Sundaram
>Priority: Major
>  Labels: pull-request-available
>
> At the time of raising this, Hadoop and Tez also use 4.1.100.Final in their 
> trunk
> Related - https://issues.apache.org/jira/browse/TEZ-4538



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28040) Upgrade netty to 4.1.100 due to CVE-2023-44487

2024-01-29 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-28040:
--
Labels: pull-request-available  (was: )

> Upgrade netty to 4.1.100 due to CVE-2023-44487
> --
>
> Key: HIVE-28040
> URL: https://issues.apache.org/jira/browse/HIVE-28040
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Anmol Sundaram
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-28041) Upgrade jetty to 9.4.53 due to CVE-2023-44487

2024-01-29 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28041?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-28041:
--
Labels: pull-request-available  (was: )

> Upgrade jetty to 9.4.53 due to CVE-2023-44487
> -
>
> Key: HIVE-28041
> URL: https://issues.apache.org/jira/browse/HIVE-28041
> Project: Hive
>  Issue Type: Task
>  Components: Hive
>Reporter: Anmol Sundaram
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (HIVE-28042) DigestMD5 error during opening connection to HMS

2024-01-29 Thread Vikram Ahuja (Jira)
Vikram Ahuja created HIVE-28042:
---

 Summary: DigestMD5 error during opening connection to HMS
 Key: HIVE-28042
 URL: https://issues.apache.org/jira/browse/HIVE-28042
 Project: Hive
  Issue Type: Bug
Reporter: Vikram Ahuja
Assignee: Vikram Ahuja


Hello,

In our deployment we are facing the following exception in the HMS logs when a 
HMS connection is opened from the HS2 in some cases leading to query failures:
{code:java}
2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
password    at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    at 
com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
    at 
org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
    at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) 
   at 
org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
    at 
org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
    at java.security.AccessController.doPrivileged(Native Method)    at 
javax.security.auth.Subject.doAs(Subject.javA)    at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) 
   at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
    at 
org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
    at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java)  
  at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) 
   at java.lang.Thread.run(Thread.java)Caused by: 
org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
sequenceNumber=3294063, masterKeyId=7601    at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    at 
org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
    at 
org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
    at 
com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
    ... 15 more {code}
 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work started] (HIVE-28042) DigestMD5 error during opening connection to HMS

2024-01-29 Thread Vikram Ahuja (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-28042 started by Vikram Ahuja.
---
> DigestMD5 error during opening connection to HMS
> 
>
> Key: HIVE-28042
> URL: https://issues.apache.org/jira/browse/HIVE-28042
> Project: Hive
>  Issue Type: Bug
>Reporter: Vikram Ahuja
>Assignee: Vikram Ahuja
>Priority: Major
>
> Hello,
> In our deployment we are facing the following exception in the HMS logs when 
> a HMS connection is opened from the HS2 in some cases leading to query 
> failures:
> {code:java}
> 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: 
> transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation 
> failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring 
> password    at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
>     at 
> com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java)
>     at 
> org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java)
>     at 
> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java)    
> at 
> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java)
>     at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java)
>     at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
>     at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java)
>     at java.security.AccessController.doPrivileged(Native Method)    
> at javax.security.auth.Subject.doAs(Subject.javA)    at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java)
>     at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java)
>     at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java)
>     at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java)
>     at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java)   
>  at java.lang.Thread.run(Thread.java)Caused by: 
> org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or 
> does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, 
> realUser=*, issueDate=1705973286139, maxDate=1706578086139, 
> sequenceNumber=3294063, masterKeyId=7601    at 
> org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
>     at 
> org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java)
>     at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java)
>     at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java)
>     at 
> com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java)
>     ... 15 more {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (HIVE-28041) Upgrade jetty to 9.4.53 due to CVE-2023-44487

2024-01-29 Thread Anmol Sundaram (Jira)
Anmol Sundaram created HIVE-28041:
-

 Summary: Upgrade jetty to 9.4.53 due to CVE-2023-44487
 Key: HIVE-28041
 URL: https://issues.apache.org/jira/browse/HIVE-28041
 Project: Hive
  Issue Type: Task
  Components: Hive
Reporter: Anmol Sundaram






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (HIVE-28040) Upgrade netty to 4.1.100 due to CVE-2023-44487

2024-01-29 Thread Anmol Sundaram (Jira)
Anmol Sundaram created HIVE-28040:
-

 Summary: Upgrade netty to 4.1.100 due to CVE-2023-44487
 Key: HIVE-28040
 URL: https://issues.apache.org/jira/browse/HIVE-28040
 Project: Hive
  Issue Type: Task
  Components: Hive
Reporter: Anmol Sundaram






--
This message was sent by Atlassian Jira
(v8.20.10#820010)