[jira] [Commented] (HIVE-28042) DigestMD5 token expired or does not exist error while opening a new connection to HMS
[ https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812178#comment-17812178 ] Butao Zhang commented on HIVE-28042: Just thinking aloud. :) set hive.cluster.delegation.token.store.class=org.apache.hadoop.hive.metastore.security.DBTokenStore; Did you try to use DBTokenStore to store the token info? Maybe it is better than the default token store MemoryTokenStore, maybe it can fix your issue? > DigestMD5 token expired or does not exist error while opening a new > connection to HMS > - > > Key: HIVE-28042 > URL: https://issues.apache.org/jira/browse/HIVE-28042 > Project: Hive > Issue Type: Bug >Reporter: Vikram Ahuja >Assignee: Vikram Ahuja >Priority: Major > > Hello, > In our deployment we are facing the following exception in the HMS logs when > a HMS connection is opened from the HS2 in cases where a session is open for > a long time leading to query failures: > {code:java} > 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: > transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation > failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring > password > at > com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) > > at > com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) > > at > org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) > at > org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) > at > org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) > > at > org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) > > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.javA) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) > > at > org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) > > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) > > at java.lang.Thread.run(Thread.java)Caused by: > org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or > does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, > realUser=*, issueDate=1705973286139, maxDate=1706578086139, > sequenceNumber=3294063, masterKeyId=7601 > at > org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) > > at > org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) > > at > com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) > ... 15 more {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28042) DigestMD5 token expired or does not exist error while opening a new connection to HMS
[ https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vikram Ahuja updated HIVE-28042: Summary: DigestMD5 token expired or does not exist error while opening a new connection to HMS (was: DigestMD5 error during opening connection to HMS) > DigestMD5 token expired or does not exist error while opening a new > connection to HMS > - > > Key: HIVE-28042 > URL: https://issues.apache.org/jira/browse/HIVE-28042 > Project: Hive > Issue Type: Bug >Reporter: Vikram Ahuja >Assignee: Vikram Ahuja >Priority: Major > > Hello, > In our deployment we are facing the following exception in the HMS logs when > a HMS connection is opened from the HS2 in cases where a session is open for > a long time leading to query failures: > {code:java} > 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: > transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation > failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring > password > at > com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) > > at > com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) > > at > org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) > at > org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) > at > org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) > > at > org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) > > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.javA) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) > > at > org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) > > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) > > at java.lang.Thread.run(Thread.java)Caused by: > org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or > does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, > realUser=*, issueDate=1705973286139, maxDate=1706578086139, > sequenceNumber=3294063, masterKeyId=7601 > at > org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) > > at > org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) > > at > com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) > ... 15 more {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28042) DigestMD5 error during opening connection to HMS
[ https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812159#comment-17812159 ] Vikram Ahuja commented on HIVE-28042: - Raised PR: [https://github.com/apache/hive/pull/5049] for the same > DigestMD5 error during opening connection to HMS > > > Key: HIVE-28042 > URL: https://issues.apache.org/jira/browse/HIVE-28042 > Project: Hive > Issue Type: Bug >Reporter: Vikram Ahuja >Assignee: Vikram Ahuja >Priority: Major > > Hello, > In our deployment we are facing the following exception in the HMS logs when > a HMS connection is opened from the HS2 in cases where a session is open for > a long time leading to query failures: > {code:java} > 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: > transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation > failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring > password > at > com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) > > at > com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) > > at > org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) > at > org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) > at > org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) > > at > org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) > > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.javA) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) > > at > org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) > > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) > > at java.lang.Thread.run(Thread.java)Caused by: > org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or > does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, > realUser=*, issueDate=1705973286139, maxDate=1706578086139, > sequenceNumber=3294063, masterKeyId=7601 > at > org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) > > at > org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) > > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) > > at > com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) > ... 15 more {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-28002) Make hive.metastore.batch.retrieve.max property to be read from hiveconf when it is requested from HS2
[ https://issues.apache.org/jira/browse/HIVE-28002?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vikram Ahuja resolved HIVE-28002. - Fix Version/s: 4.0.0-beta-1 Resolution: Not A Problem > Make hive.metastore.batch.retrieve.max property to be read from hiveconf when > it is requested from HS2 > -- > > Key: HIVE-28002 > URL: https://issues.apache.org/jira/browse/HIVE-28002 > Project: Hive > Issue Type: Bug >Affects Versions: 4.0.0-beta-1 >Reporter: Vikram Ahuja >Assignee: Vikram Ahuja >Priority: Minor > Labels: pull-request-available > Fix For: 4.0.0-beta-1 > > > The property hive.metastore.batch.retrieve.max is used by both the > client(Hs2) and the HMS when fetching objects(tables/partitions) in batches. > This property is present in both hiveConf as well as metastoreConf. For some > use cases in the client(Hs2), this value is referred from the metastoreConf. > To change these use cases to HiveConf.java to have a set demarkation between > HS2 and HMS use cases. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Work stopped] (HIVE-28002) Make hive.metastore.batch.retrieve.max property to be read from hiveconf when it is requested from HS2
[ https://issues.apache.org/jira/browse/HIVE-28002?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Work on HIVE-28002 stopped by Vikram Ahuja. --- > Make hive.metastore.batch.retrieve.max property to be read from hiveconf when > it is requested from HS2 > -- > > Key: HIVE-28002 > URL: https://issues.apache.org/jira/browse/HIVE-28002 > Project: Hive > Issue Type: Bug >Affects Versions: 4.0.0-beta-1 >Reporter: Vikram Ahuja >Assignee: Vikram Ahuja >Priority: Minor > Labels: pull-request-available > > The property hive.metastore.batch.retrieve.max is used by both the > client(Hs2) and the HMS when fetching objects(tables/partitions) in batches. > This property is present in both hiveConf as well as metastoreConf. For some > use cases in the client(Hs2), this value is referred from the metastoreConf. > To change these use cases to HiveConf.java to have a set demarkation between > HS2 and HMS use cases. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Comment Edited] (HIVE-28042) DigestMD5 error during opening connection to HMS
[ https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812151#comment-17812151 ] Vikram Ahuja edited comment on HIVE-28042 at 1/30/24 5:45 AM: -- *Another instance of this issue:* {code:java} 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) at com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) at org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) at org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) at org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.javA) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) at java.lang.Thread.run(Thread.java)Caused by: org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, realUser=*, issueDate=1705973286139, maxDate=1706578086139, sequenceNumber=3294063, masterKeyId=7601 at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) ... 15 more {code} *Analysis of the issue:* This particular issue is only happening when the HS2 tries to open a new Digest MD5 based Thrift TSaslClientTransport in cases where the session is open for a long time. HS2 uses the same metaStoreClient object across all the connections that is embedded in Hive.java but in some cases we have observed that is recreating a new metaStoreClient with a fresh connection(TSaslClientTransport). Two use cases that I discovered which were leading to these issues were: # ## MSCK repair ## RetryingMetaStoreClient in case of any HMS issues(applicable to any sql query which interacts with the HMS) *Root cause of this issue:* There is a background thread called ExpiredTokenRemover running in HMS (class: TokenStoreDelegationTokenSecretManager.java ). This expiry thread itself is removing the token from the tokenStore after the renewal time has passed and also removing it after expiry time, but is should only remove it post expiry time as the token can be renewed till then. Will be raising a fix for the same by changing the code where token is deleted after renewal time itself has passed. was (Author: vikramahuja_): *Another instance of this issue:* {code:java} 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) at com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) at org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) at org.apache.thrift.transport.TSaslTransport
[jira] [Updated] (HIVE-28042) DigestMD5 error during opening connection to HMS
[ https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vikram Ahuja updated HIVE-28042: Description: Hello, In our deployment we are facing the following exception in the HMS logs when a HMS connection is opened from the HS2 in cases where a session is open for a long time leading to query failures: {code:java} 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) at com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) at org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) at org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) at org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.javA) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) at java.lang.Thread.run(Thread.java)Caused by: org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, realUser=*, issueDate=1705973286139, maxDate=1706578086139, sequenceNumber=3294063, masterKeyId=7601 at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) ... 15 more {code} was: Hello, In our deployment we are facing the following exception in the HMS logs when a HMS connection is opened from the HS2 in cases where a session is open for a long time leading to query failures: {code:java} 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) at com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) at org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) at org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) at org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.javA) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) at
[jira] [Commented] (HIVE-28042) DigestMD5 error during opening connection to HMS
[ https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812151#comment-17812151 ] Vikram Ahuja commented on HIVE-28042: - *Another instance of this issue:* {code:java} 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) at com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) at org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) at org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) at org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.javA) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) at java.lang.Thread.run(Thread.java)Caused by: org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, realUser=*, issueDate=1705973286139, maxDate=1706578086139, sequenceNumber=3294063, masterKeyId=7601 at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) ... 15 more {code} *Analysis of the issue:* This particular issue is only happening when the HS2 tries to open a new Digest MD5 based Thrift TSaslClientTransport in cases where the session is one for a long time. HS2 uses the same metaStoreClient object across all the connections that is embedded in Hive.java but in some cases we have observed that is recreating a new metaStoreClient with a fresh connection(TSaslClientTransport). Two use cases that I discovered which were leading to these issues were: # ## MSCK repair ## RetryingMetaStoreClient in case of any HMS issues(applicable to any sql query which interacts with the HMS) *Root cause of this issue:* There is a background thread called ExpiredTokenRemover running in HMS (class: TokenStoreDelegationTokenSecretManager.java ). This expiry thread itself is removing the token from the tokenStore after the renewal time has passed and also removing it after expiry time, but is should only remove it post expiry time as the token can be renewed till then. Will be raising a fix for the same by changing the code where token is deleted after renewal time itself has passed. > DigestMD5 error during opening connection to HMS > > > Key: HIVE-28042 > URL: https://issues.apache.org/jira/browse/HIVE-28042 > Project: Hive > Issue Type: Bug >Reporter: Vikram Ahuja >Assignee: Vikram Ahuja >Priority: Major > > Hello, > In our deployment we are facing the following exception in the HMS logs when > a HMS connection is opened from the HS2 in cases where a session is open for > a long time leading to query failures: > {code:java} > 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: > transport.TSaslTransport (TSaslTransport.java:open)
[jira] [Updated] (HIVE-24167) TPC-DS query 14 fails while generating plan for the filter
[ https://issues.apache.org/jira/browse/HIVE-24167?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] okumin updated HIVE-24167: -- Status: Patch Available (was: Open) I implemented one of the approaches in my mind. [https://github.com/apache/hive/pull/5037] Also, I wrote down what I've investigated to resolve this issue. https://gist.github.com/okumin/b111fe0a911507bdf6a7204f49b9cb72 > TPC-DS query 14 fails while generating plan for the filter > -- > > Key: HIVE-24167 > URL: https://issues.apache.org/jira/browse/HIVE-24167 > Project: Hive > Issue Type: Sub-task > Components: CBO >Reporter: Stamatis Zampetakis >Assignee: okumin >Priority: Major > Labels: hive-4.1.0-must, pull-request-available > > TPC-DS query 14 (cbo_query14.q and query4.q) fail with NPE on the metastore > with the partitioned TPC-DS 30TB dataset while generating the plan for the > filter. > The problem can be reproduced using the PR in HIVE-23965. > The current stacktrace shows that the NPE appears while trying to display the > debug message but even if this line didn't exist it would fail again later on. > {noformat} > java.lang.NullPointerException > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:10867) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11765) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11622) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11649) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11622) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11649) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11635) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlanForSubQueryPredicate(SemanticAnalyzer.java:3375) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genFilterPlan(SemanticAnalyzer.java:3473) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genBodyPlan(SemanticAnalyzer.java:10819) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11765) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11622) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11625) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11625) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11649) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11622) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11649) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genPlan(SemanticAnalyzer.java:11635) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.genOPTree(SemanticAnalyzer.java:12417) > at > org.apache.hadoop.hive.ql.parse.CalcitePlanner.genOPTree(CalcitePlanner.java:718) > at > org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:12519) > at > org.apache.hadoop.hive.ql.parse.CalcitePlanner.analyzeInternal(CalcitePlanner.java:443) > at > org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:301) > at > org.apache.hadoop.hive.ql.parse.ExplainSemanticAnalyzer.analyzeInternal(ExplainSemanticAnalyzer.java:171) > at > org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:301) > at org.apache.hadoop.hive.ql.Compiler.analyze(Compiler.java:220) > at org.apache.hadoop.hive.ql.Compiler.compile(Compiler.java:104) > at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:173) > at org.apache.hadoop.hive.ql.Driver.compileInternal(Driver.java:414) > at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:363) > at org.apache.hadoop.hive.ql.Driver.compileAndRespond(Driver.java:357) > at > org.apache.hadoop.hive.ql.reexec.ReExecDriver.compileAndRespond(ReExecDriver.java:129) > at > org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:231) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:258) > at > org.apache.hadoop.hive.cli.CliDriver.processCmd1(CliDriver.java:203) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:129) > at > org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:424) > at
[jira] [Updated] (HIVE-27938) Iceberg: Fix java.lang.ClassCastException during vectorized reads on partition columns
[ https://issues.apache.org/jira/browse/HIVE-27938?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-27938: -- Labels: pull-request-available (was: ) > Iceberg: Fix java.lang.ClassCastException during vectorized reads on > partition columns > --- > > Key: HIVE-27938 > URL: https://issues.apache.org/jira/browse/HIVE-27938 > Project: Hive > Issue Type: Bug >Reporter: Simhadri Govindappa >Assignee: Simhadri Govindappa >Priority: Major > Labels: pull-request-available > > {code:java} > 1: jdbc:hive2://localhost:10001/> CREATE EXTERNAL TABLE ice3 (`col1` int, > `calday` date) PARTITIONED BY SPEC (calday) stored by iceberg > tblproperties('format-version'='2'); > 1: jdbc:hive2://localhost:10001/>insert into ice3 values(1, '2020-11-20'); > 1: jdbc:hive2://localhost:10001/> select count(calday) from ice3; > {code} > Full stack trace: > {code:java} > INFO : Compiling > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): > select count(calday) from ice3INFO : No Stats for default@ice3, Columns: > caldayINFO : Semantic Analysis Completed (retrial = false)INFO : Created > Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, type:bigint, > comment:null)], properties:null)INFO : Completed compiling > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab); > Time taken: 0.196 secondsINFO : Operation QUERY obtained 0 locksINFO : > Executing > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): > select count(calday) from ice3INFO : Query ID = > root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO : Total jobs = > 1INFO : Launching Job 1 out of 1INFO : Starting task [Stage-1:MAPRED] in > serial modeINFO : Subscribed to counters: [] for queryId: > root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO : Session is > already openINFO : Dag name: select count(calday) from ice3 (Stage-1)INFO : > HS2 Host: [localhost], Query ID: > [root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab], Dag ID: > [dag_1701888162260_0001_2], DAG Session ID: > [application_1701888162260_0001]INFO : Status: Running (Executing on YARN > cluster with App id application_1701888162260_0001) > -- > VERTICES MODE STATUS TOTAL COMPLETED RUNNING PENDING > FAILED > KILLED--Map > 1 container RUNNING 1 0 0 1 > 4 0Reducer 2 container INITED 1 0 > 0 1 0 > 0--VERTICES: > 00/02 [>>--] 0% ELAPSED TIME: 1.41 > s--ERROR > : Status: FailedERROR : Vertex failed, vertexName=Map 1, > vertexId=vertex_1701888162260_0001_2_00, diagnostics=[Task failed, > taskId=task_1701888162260_0001_2_00_00, diagnostics=[TaskAttempt 0 > failed, info=[Error: Error while running task ( failure ) : > attempt_1701888162260_0001_2_00_00_0:java.lang.RuntimeException: > org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: > java.lang.ClassCastException: java.time.LocalDate cannot be cast to > org.apache.hadoop.hive.common.type.Dateat > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:348) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:276) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:381) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:82) >at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:69) >at java.security.AccessController.doPrivileged(Native Method) at > javax.security.auth.Subject.doAs(Subject.java:422) at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1878) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:69) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:39) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:111) > at > com.google.common.util.concurrent.InterruptibleTask.run(Inte
[jira] [Updated] (HIVE-27938) Iceberg: Fix java.lang.ClassCastException during vectorized reads on partition columns
[ https://issues.apache.org/jira/browse/HIVE-27938?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Simhadri Govindappa updated HIVE-27938: --- Summary: Iceberg: Fix java.lang.ClassCastException during vectorized reads on partition columns (was: Iceberg: Date type Partitioned column throws java.lang.ClassCastException: java.time.LocalDate cannot be cast to org.apache.hadoop.hive.common.type.Date) > Iceberg: Fix java.lang.ClassCastException during vectorized reads on > partition columns > --- > > Key: HIVE-27938 > URL: https://issues.apache.org/jira/browse/HIVE-27938 > Project: Hive > Issue Type: Bug >Reporter: Simhadri Govindappa >Assignee: Simhadri Govindappa >Priority: Major > > {code:java} > 1: jdbc:hive2://localhost:10001/> CREATE EXTERNAL TABLE ice3 (`col1` int, > `calday` date) PARTITIONED BY SPEC (calday) stored by iceberg > tblproperties('format-version'='2'); > 1: jdbc:hive2://localhost:10001/>insert into ice3 values(1, '2020-11-20'); > 1: jdbc:hive2://localhost:10001/> select count(calday) from ice3; > {code} > Full stack trace: > {code:java} > INFO : Compiling > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): > select count(calday) from ice3INFO : No Stats for default@ice3, Columns: > caldayINFO : Semantic Analysis Completed (retrial = false)INFO : Created > Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, type:bigint, > comment:null)], properties:null)INFO : Completed compiling > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab); > Time taken: 0.196 secondsINFO : Operation QUERY obtained 0 locksINFO : > Executing > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): > select count(calday) from ice3INFO : Query ID = > root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO : Total jobs = > 1INFO : Launching Job 1 out of 1INFO : Starting task [Stage-1:MAPRED] in > serial modeINFO : Subscribed to counters: [] for queryId: > root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO : Session is > already openINFO : Dag name: select count(calday) from ice3 (Stage-1)INFO : > HS2 Host: [localhost], Query ID: > [root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab], Dag ID: > [dag_1701888162260_0001_2], DAG Session ID: > [application_1701888162260_0001]INFO : Status: Running (Executing on YARN > cluster with App id application_1701888162260_0001) > -- > VERTICES MODE STATUS TOTAL COMPLETED RUNNING PENDING > FAILED > KILLED--Map > 1 container RUNNING 1 0 0 1 > 4 0Reducer 2 container INITED 1 0 > 0 1 0 > 0--VERTICES: > 00/02 [>>--] 0% ELAPSED TIME: 1.41 > s--ERROR > : Status: FailedERROR : Vertex failed, vertexName=Map 1, > vertexId=vertex_1701888162260_0001_2_00, diagnostics=[Task failed, > taskId=task_1701888162260_0001_2_00_00, diagnostics=[TaskAttempt 0 > failed, info=[Error: Error while running task ( failure ) : > attempt_1701888162260_0001_2_00_00_0:java.lang.RuntimeException: > org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: > java.lang.ClassCastException: java.time.LocalDate cannot be cast to > org.apache.hadoop.hive.common.type.Dateat > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:348) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:276) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:381) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:82) >at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:69) >at java.security.AccessController.doPrivileged(Native Method) at > javax.security.auth.Subject.doAs(Subject.java:422) at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1878) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:69) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:39) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at > com.google.common.util.concurrent
[jira] [Commented] (HIVE-27938) Iceberg: Date type Partitioned column throws java.lang.ClassCastException: java.time.LocalDate cannot be cast to org.apache.hadoop.hive.common.type.Date
[ https://issues.apache.org/jira/browse/HIVE-27938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17812108#comment-17812108 ] Simhadri Govindappa commented on HIVE-27938: The error is also present for DATE and DECIMAL columns > Iceberg: Date type Partitioned column throws java.lang.ClassCastException: > java.time.LocalDate cannot be cast to org.apache.hadoop.hive.common.type.Date > > > Key: HIVE-27938 > URL: https://issues.apache.org/jira/browse/HIVE-27938 > Project: Hive > Issue Type: Bug >Reporter: Simhadri Govindappa >Assignee: Simhadri Govindappa >Priority: Major > > {code:java} > 1: jdbc:hive2://localhost:10001/> CREATE EXTERNAL TABLE ice3 (`col1` int, > `calday` date) PARTITIONED BY SPEC (calday) stored by iceberg > tblproperties('format-version'='2'); > 1: jdbc:hive2://localhost:10001/>insert into ice3 values(1, '2020-11-20'); > 1: jdbc:hive2://localhost:10001/> select count(calday) from ice3; > {code} > Full stack trace: > {code:java} > INFO : Compiling > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): > select count(calday) from ice3INFO : No Stats for default@ice3, Columns: > caldayINFO : Semantic Analysis Completed (retrial = false)INFO : Created > Hive schema: Schema(fieldSchemas:[FieldSchema(name:_c0, type:bigint, > comment:null)], properties:null)INFO : Completed compiling > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab); > Time taken: 0.196 secondsINFO : Operation QUERY obtained 0 locksINFO : > Executing > command(queryId=root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab): > select count(calday) from ice3INFO : Query ID = > root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO : Total jobs = > 1INFO : Launching Job 1 out of 1INFO : Starting task [Stage-1:MAPRED] in > serial modeINFO : Subscribed to counters: [] for queryId: > root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feabINFO : Session is > already openINFO : Dag name: select count(calday) from ice3 (Stage-1)INFO : > HS2 Host: [localhost], Query ID: > [root_20231206184246_e8da1539-7537-45fe-af67-4c7ba219feab], Dag ID: > [dag_1701888162260_0001_2], DAG Session ID: > [application_1701888162260_0001]INFO : Status: Running (Executing on YARN > cluster with App id application_1701888162260_0001) > -- > VERTICES MODE STATUS TOTAL COMPLETED RUNNING PENDING > FAILED > KILLED--Map > 1 container RUNNING 1 0 0 1 > 4 0Reducer 2 container INITED 1 0 > 0 1 0 > 0--VERTICES: > 00/02 [>>--] 0% ELAPSED TIME: 1.41 > s--ERROR > : Status: FailedERROR : Vertex failed, vertexName=Map 1, > vertexId=vertex_1701888162260_0001_2_00, diagnostics=[Task failed, > taskId=task_1701888162260_0001_2_00_00, diagnostics=[TaskAttempt 0 > failed, info=[Error: Error while running task ( failure ) : > attempt_1701888162260_0001_2_00_00_0:java.lang.RuntimeException: > org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: > java.lang.ClassCastException: java.time.LocalDate cannot be cast to > org.apache.hadoop.hive.common.type.Dateat > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:348) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:276) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:381) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:82) >at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:69) >at java.security.AccessController.doPrivileged(Native Method) at > javax.security.auth.Subject.doAs(Subject.java:422) at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1878) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:69) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:39) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at > com.google.common.util.concurrent.TrustedListenableFutu
[jira] [Assigned] (HIVE-4752) Add support for hs2 api to use thrift over http
[ https://issues.apache.org/jira/browse/HIVE-4752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vaibhav Gumashta reassigned HIVE-4752: -- Assignee: Vaibhav Gumashta > Add support for hs2 api to use thrift over http > --- > > Key: HIVE-4752 > URL: https://issues.apache.org/jira/browse/HIVE-4752 > Project: Hive > Issue Type: New Feature > Components: HiveServer2 >Affects Versions: 0.13.0 >Reporter: Thejas Nair >Assignee: Vaibhav Gumashta >Priority: Major > > Hiveserver2 acts as service on the cluster for external applications. One way > to implement access control to services on a hadoop cluster to have a gateway > server authorizes service requests before forwarding them to the server. The > [knox project | http://wiki.apache.org/incubator/knox] has taken this > approach to simplify cluster security management. > Other services on hadoop cluster such as webhdfs and webhcat already use > HTTP. Having hiveserver2 also support thrift over http transport will enable > securing hiveserver2 as well using the same approach. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28015) Iceberg: Add identifier-field-ids support in Hive
[ https://issues.apache.org/jira/browse/HIVE-28015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28015: -- Labels: pull-request-available (was: ) > Iceberg: Add identifier-field-ids support in Hive > - > > Key: HIVE-28015 > URL: https://issues.apache.org/jira/browse/HIVE-28015 > Project: Hive > Issue Type: Improvement > Components: Iceberg integration >Affects Versions: 4.0.0 >Reporter: Denys Kuzmenko >Assignee: Butao Zhang >Priority: Major > Labels: pull-request-available > > Some writer engines require primary keys on a table so that they can use them > for writing equality deletes (only the PK cols are written to the eq-delete > files). > Hive currently doesn't reject setting PKs for Iceberg tables, however, it > just ignores them. This succeeds: > {code:java} > create table ice_pk (i int, j int, primary key(i)) stored by iceberg; > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28045) Alter table without cascade seems to change metadata on existing partitions (Avro, ORC)
[ https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] László Bodor updated HIVE-28045: Description: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched, like below: {code} aaa bbb NULL {code} the same applies to Avro, if there is no external schema defined (which is table level one) repro: https://github.com/abstractdog/hive/commit/7ef97479796#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5 was: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched, like below: {code} aaa bbb NULL {code} the same applies to Avro, if there is no external schema defined (which is table level one) repro: https://github.com/abstractdog/hive/commit/e0de2c52de288cafbba3bcec3533bc78890e22f9#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5 > Alter table without cascade seems to change metadata on existing partitions > (Avro, ORC) > --- > > Key: HIVE-28045 > URL: https://issues.apache.org/jira/browse/HIVE-28045 > Project: Hive > Issue Type: Bug >Reporter: László Bodor >Assignee: László Bodor >Priority: Major > > {code} > create external table test_orc (a string, b string) partitioned by (pt > string) STORED AS ORC; > insert into test_orc PARTITION(pt='') values ("",""); > alter table test_orc add columns (c string); -- no CASCADE > insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", > "ccc"); > select * from test_orc; > {code} > returns > {code} > aaa bbb ccc > {code} > which is unexpected without cascading the alter table, so the existing > partition is supposed to be untouched, like below: > {code} > aaa bbb NULL > {code} > the same applies to Avro, if there is no external schema defined (which is > table level one) > repro: > https://github.com/abstractdog/hive/commit/7ef97479796#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28045) Alter table without cascade seems to change metadata on existing partitions (Avro, ORC)
[ https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] László Bodor updated HIVE-28045: Description: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched, like below: {code} aaa bbb NULL {code} the same applies to Avro, if there is no external schema defined (which is table level one) repro: https://github.com/abstractdog/hive/commit/e0de2c52de288cafbba3bcec3533bc78890e22f9#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5 was: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched, like below: {code} aaa bbb NULL {code} the same applies to Avro, if there is no external schema defined (which is table level one) > Alter table without cascade seems to change metadata on existing partitions > (Avro, ORC) > --- > > Key: HIVE-28045 > URL: https://issues.apache.org/jira/browse/HIVE-28045 > Project: Hive > Issue Type: Bug >Reporter: László Bodor >Assignee: László Bodor >Priority: Major > > {code} > create external table test_orc (a string, b string) partitioned by (pt > string) STORED AS ORC; > insert into test_orc PARTITION(pt='') values ("",""); > alter table test_orc add columns (c string); -- no CASCADE > insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", > "ccc"); > select * from test_orc; > {code} > returns > {code} > aaa bbb ccc > {code} > which is unexpected without cascading the alter table, so the existing > partition is supposed to be untouched, like below: > {code} > aaa bbb NULL > {code} > the same applies to Avro, if there is no external schema defined (which is > table level one) > repro: > https://github.com/abstractdog/hive/commit/e0de2c52de288cafbba3bcec3533bc78890e22f9#diff-1ca7746de2d6586dc2d9c1edbe5c18ca35c37fd39382c030142c782a5326faa5 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28045) Alter table without cascade seems to change metadata on existing partitions (Avro, ORC)
[ https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] László Bodor updated HIVE-28045: Description: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched, like below: {code} aaa bbb NULL {code} the same applies to Avro, if there is no external schema defined (which is table level one) was: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched, like below: {code} aaa bbb NULL {code} > Alter table without cascade seems to change metadata on existing partitions > (Avro, ORC) > --- > > Key: HIVE-28045 > URL: https://issues.apache.org/jira/browse/HIVE-28045 > Project: Hive > Issue Type: Bug >Reporter: László Bodor >Assignee: László Bodor >Priority: Major > > {code} > create external table test_orc (a string, b string) partitioned by (pt > string) STORED AS ORC; > insert into test_orc PARTITION(pt='') values ("",""); > alter table test_orc add columns (c string); -- no CASCADE > insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", > "ccc"); > select * from test_orc; > {code} > returns > {code} > aaa bbb ccc > {code} > which is unexpected without cascading the alter table, so the existing > partition is supposed to be untouched, like below: > {code} > aaa bbb NULL > {code} > the same applies to Avro, if there is no external schema defined (which is > table level one) -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28045) Alter table without cascade seems to change metadata on existing partitions (Avro, ORC)
[ https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] László Bodor updated HIVE-28045: Summary: Alter table without cascade seems to change metadata on existing partitions (Avro, ORC) (was: Alter table without cascade seems to metadata on existing partitions (Avro, ORC)) > Alter table without cascade seems to change metadata on existing partitions > (Avro, ORC) > --- > > Key: HIVE-28045 > URL: https://issues.apache.org/jira/browse/HIVE-28045 > Project: Hive > Issue Type: Bug >Reporter: László Bodor >Assignee: László Bodor >Priority: Major > > {code} > create external table test_orc (a string, b string) partitioned by (pt > string) STORED AS ORC; > insert into test_orc PARTITION(pt='') values ("",""); > alter table test_orc add columns (c string); -- no CASCADE > insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", > "ccc"); > select * from test_orc; > {code} > returns > {code} > aaa bbb ccc > {code} > which is unexpected without cascading the alter table, so the existing > partition is supposed to be untouched, like below: > {code} > aaa bbb NULL > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28045) Alter table without cascade seems to metadata on existing partitions (Avro, ORC)
[ https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] László Bodor updated HIVE-28045: Description: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched, like below: {code} aaa bbb NULL {code} was: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched: {code} aaa bbb NULL {code} > Alter table without cascade seems to metadata on existing partitions (Avro, > ORC) > > > Key: HIVE-28045 > URL: https://issues.apache.org/jira/browse/HIVE-28045 > Project: Hive > Issue Type: Bug >Reporter: László Bodor >Assignee: László Bodor >Priority: Major > > {code} > create external table test_orc (a string, b string) partitioned by (pt > string) STORED AS ORC; > insert into test_orc PARTITION(pt='') values ("",""); > alter table test_orc add columns (c string); -- no CASCADE > insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", > "ccc"); > select * from test_orc; > {code} > returns > {code} > aaa bbb ccc > {code} > which is unexpected without cascading the alter table, so the existing > partition is supposed to be untouched, like below: > {code} > aaa bbb NULL > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28045) Alter table without cascade seems to metadata on existing partitions (Avro, ORC)
[ https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] László Bodor updated HIVE-28045: Description: {code} create external table test_orc (a string, b string) partitioned by (pt string) STORED AS ORC; insert into test_orc PARTITION(pt='') values ("",""); alter table test_orc add columns (c string); -- no CASCADE insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", "ccc"); select * from test_orc; {code} returns {code} aaa bbb ccc {code} which is unexpected without cascading the alter table, so the existing partition is supposed to be untouched: {code} aaa bbb NULL {code} > Alter table without cascade seems to metadata on existing partitions (Avro, > ORC) > > > Key: HIVE-28045 > URL: https://issues.apache.org/jira/browse/HIVE-28045 > Project: Hive > Issue Type: Bug >Reporter: László Bodor >Assignee: László Bodor >Priority: Major > > {code} > create external table test_orc (a string, b string) partitioned by (pt > string) STORED AS ORC; > insert into test_orc PARTITION(pt='') values ("",""); > alter table test_orc add columns (c string); -- no CASCADE > insert overwrite table test_orc PARTITION(pt='') values ("aaa", "bbb", > "ccc"); > select * from test_orc; > {code} > returns > {code} > aaa bbb ccc > {code} > which is unexpected without cascading the alter table, so the existing > partition is supposed to be untouched: > {code} > aaa bbb NULL > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28045) Alter table without cascade seems to metadata on existing partitions (Avro, ORC)
László Bodor created HIVE-28045: --- Summary: Alter table without cascade seems to metadata on existing partitions (Avro, ORC) Key: HIVE-28045 URL: https://issues.apache.org/jira/browse/HIVE-28045 Project: Hive Issue Type: Bug Reporter: László Bodor -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (HIVE-28045) Alter table without cascade seems to metadata on existing partitions (Avro, ORC)
[ https://issues.apache.org/jira/browse/HIVE-28045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] László Bodor reassigned HIVE-28045: --- Assignee: László Bodor > Alter table without cascade seems to metadata on existing partitions (Avro, > ORC) > > > Key: HIVE-28045 > URL: https://issues.apache.org/jira/browse/HIVE-28045 > Project: Hive > Issue Type: Bug >Reporter: László Bodor >Assignee: László Bodor >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Work started] (HIVE-27972) Set 'tez' as default value in hive.execution.engine
[ https://issues.apache.org/jira/browse/HIVE-27972?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Work on HIVE-27972 started by László Bodor. --- > Set 'tez' as default value in hive.execution.engine > --- > > Key: HIVE-27972 > URL: https://issues.apache.org/jira/browse/HIVE-27972 > Project: Hive > Issue Type: Sub-task >Reporter: László Bodor >Assignee: László Bodor >Priority: Major > Labels: pull-request-available > > Maybe this is not the first ticket addressing this, please link if it's a > duplicate. > We need to set this to 'tez' to reflect that we have deprecated 'mr': > https://github.com/apache/hive/blob/bd16e0098916aa5fc2dede99492c6a240b51e677/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java#L4567 > I'm expecting lots of UT failures because of this, as we're still running > those on mr (which might be fine where the actual unit test is not closely > related to the execution engine), so we'll see what to do. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1
[ https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28043: -- Labels: pull-request-available (was: ) > Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1 > --- > > Key: HIVE-28043 > URL: https://issues.apache.org/jira/browse/HIVE-28043 > Project: Hive > Issue Type: Task >Reporter: Anmol Sundaram >Priority: Major > Labels: pull-request-available > > The recent stable version of ZK is 3.9.1 > We should see if we can have a similar upgrade for Hive as well > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1
[ https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anmol Sundaram updated HIVE-28043: -- Description: The recent stable version of ZK is 3.9.1 We should see if we can have a similar upgrade for Hive as well was: Hadoop upgraded their Zookeeper versions to 3.8.3. See [HADOOP-18613|https://issues.apache.org/jira/browse/HADOOP-18613] We should see if we can have a similar upgrade for Hive as well > Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1 > --- > > Key: HIVE-28043 > URL: https://issues.apache.org/jira/browse/HIVE-28043 > Project: Hive > Issue Type: Task >Reporter: Anmol Sundaram >Priority: Major > > The recent stable version of ZK is 3.9.1 > We should see if we can have a similar upgrade for Hive as well > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1
[ https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anmol Sundaram updated HIVE-28043: -- Summary: Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1 (was: Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3) > Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.9.1 > --- > > Key: HIVE-28043 > URL: https://issues.apache.org/jira/browse/HIVE-28043 > Project: Hive > Issue Type: Task >Reporter: Anmol Sundaram >Priority: Major > > Hadoop upgraded their Zookeeper versions to 3.8.3. See > [HADOOP-18613|https://issues.apache.org/jira/browse/HADOOP-18613] > We should see if we can have a similar upgrade for Hive as well > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28044) Replication tests to run on Tez
László Bodor created HIVE-28044: --- Summary: Replication tests to run on Tez Key: HIVE-28044 URL: https://issues.apache.org/jira/browse/HIVE-28044 Project: Hive Issue Type: Sub-task Reporter: László Bodor -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28042) DigestMD5 error during opening connection to HMS
[ https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vikram Ahuja updated HIVE-28042: Description: Hello, In our deployment we are facing the following exception in the HMS logs when a HMS connection is opened from the HS2 in cases where a session is open for a long time leading to query failures: {code:java} 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) at com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) at org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) at org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) at org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.javA) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) at java.lang.Thread.run(Thread.java)Caused by: org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, realUser=*, issueDate=1705973286139, maxDate=1706578086139, sequenceNumber=3294063, masterKeyId=7601 at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) ... 15 more {code} was: Hello, In our deployment we are facing the following exception in the HMS logs when a HMS connection is opened from the HS2 in some cases leading to query failures: {code:java} 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) at com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) at org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) at org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) at org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.javA) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) at org.apache.thrift.server.TThreadPoolServer$WorkerProc
[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3
[ https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anmol Sundaram updated HIVE-28043: -- Description: Hadoop upgraded their Zookeeper versions to 3.8.3. See [HADOOP-18613|https://issues.apache.org/jira/browse/HADOOP-18613] We should see if we can have a similar upgrade for Hive as well was: Hadoop upgraded their Zookeeper versions to 3.8.3 We should see if we can have a similar upgrade for Hive as well > Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3 > --- > > Key: HIVE-28043 > URL: https://issues.apache.org/jira/browse/HIVE-28043 > Project: Hive > Issue Type: Task >Reporter: Anmol Sundaram >Priority: Major > > Hadoop upgraded their Zookeeper versions to 3.8.3. See > [HADOOP-18613|https://issues.apache.org/jira/browse/HADOOP-18613] > We should see if we can have a similar upgrade for Hive as well > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3
[ https://issues.apache.org/jira/browse/HIVE-28043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anmol Sundaram updated HIVE-28043: -- Description: Hadoop upgraded their Zookeeper versions to 3.8.3 We should see if we can have a similar upgrade for Hive as well > Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3 > --- > > Key: HIVE-28043 > URL: https://issues.apache.org/jira/browse/HIVE-28043 > Project: Hive > Issue Type: Task >Reporter: Anmol Sundaram >Priority: Major > > Hadoop upgraded their Zookeeper versions to 3.8.3 > We should see if we can have a similar upgrade for Hive as well > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Comment Edited] (HIVE-27929) Run TPC-DS queries and validate results correctness
[ https://issues.apache.org/jira/browse/HIVE-27929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17811790#comment-17811790 ] Simhadri Govindappa edited comment on HIVE-27929 at 1/29/24 9:36 AM: - I was able to run a 1tb tpcds run hive master, with the following versions: # Hive - master (last commit from 9th of jan) # Hadoop- 3.3.6 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from classpath) With these versions, * *ORC External:* I was able to run all the tpcds queries successfully. * *ORC manager:* Faced the same issue described above. was (Author: simhadri-g): I was able to run a 1tb tpcds run hive master, with the following versions: # Hive - master (last commit from 9th of jan) # Hadoop- 3.3.6 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from classpath) With these versions, * *ORC External:* I was able to run all the tpcds queries successfully. * *ORC manager:* Faced the same issue described above. ( HIVE-28004 ) > Run TPC-DS queries and validate results correctness > --- > > Key: HIVE-27929 > URL: https://issues.apache.org/jira/browse/HIVE-27929 > Project: Hive > Issue Type: Sub-task >Reporter: Denys Kuzmenko >Assignee: Simhadri Govindappa >Priority: Major > > release branch: *branch-4.0* > https://github.com/apache/hive/tree/branch-4.0 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Comment Edited] (HIVE-27929) Run TPC-DS queries and validate results correctness
[ https://issues.apache.org/jira/browse/HIVE-27929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17811790#comment-17811790 ] Simhadri Govindappa edited comment on HIVE-27929 at 1/29/24 9:35 AM: - I was able to run a 1tb tpcds run hive master, with the following versions: # Hive - master (last commit from 9th of jan) # Hadoop- 3.3.6 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from classpath) With these versions, * *ORC External:* I was able to run all the tpcds queries successfully. * *ORC manager:* Faced the same issue described above. ( HIVE-28004 ) was (Author: simhadri-g): I was able to run a 1tb tpcds run hive master, with the following versions: # Hive - master (last commit from 9th of jan) # Hadoop- 3.3.6 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from classpath) With these versions, * *ORC External:* I was able to run all the tpcds queries successfully. * *ORC manager:* Faced the same issue described above. ( HIVE-28004 ) > Run TPC-DS queries and validate results correctness > --- > > Key: HIVE-27929 > URL: https://issues.apache.org/jira/browse/HIVE-27929 > Project: Hive > Issue Type: Sub-task >Reporter: Denys Kuzmenko >Assignee: Simhadri Govindappa >Priority: Major > > release branch: *branch-4.0* > https://github.com/apache/hive/tree/branch-4.0 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-27929) Run TPC-DS queries and validate results correctness
[ https://issues.apache.org/jira/browse/HIVE-27929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17811790#comment-17811790 ] Simhadri Govindappa commented on HIVE-27929: I was able to run a 1tb tpcds run hive master, with the following versions: # Hive - master (last commit from 9th of jan) # Hadoop- 3.3.6 # Tez- 0.10.2 (with a patch to remove the conflicting hadoop-client jar from classpath) With these versions, * *ORC External:* I was able to run all the tpcds queries successfully. * *ORC manager:* Faced the same issue described above. ( HIVE-28004 ) > Run TPC-DS queries and validate results correctness > --- > > Key: HIVE-27929 > URL: https://issues.apache.org/jira/browse/HIVE-27929 > Project: Hive > Issue Type: Sub-task >Reporter: Denys Kuzmenko >Assignee: Simhadri Govindappa >Priority: Major > > release branch: *branch-4.0* > https://github.com/apache/hive/tree/branch-4.0 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28043) Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3
Anmol Sundaram created HIVE-28043: - Summary: Bump org.apache.zookeeper:zookeeper from 3.7.2 to 3.8.3 Key: HIVE-28043 URL: https://issues.apache.org/jira/browse/HIVE-28043 Project: Hive Issue Type: Task Reporter: Anmol Sundaram -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28041) Upgrade jetty to 9.4.53 due to CVE-2023-44487
[ https://issues.apache.org/jira/browse/HIVE-28041?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anmol Sundaram updated HIVE-28041: -- Description: At the time of raising this, Hadoop also uses Jetty 9.4.53 Ref - https://github.com/apache/hadoop/blob/trunk/hadoop-project/pom.xml#L40 > Upgrade jetty to 9.4.53 due to CVE-2023-44487 > - > > Key: HIVE-28041 > URL: https://issues.apache.org/jira/browse/HIVE-28041 > Project: Hive > Issue Type: Task > Components: Hive >Reporter: Anmol Sundaram >Priority: Major > Labels: pull-request-available > > At the time of raising this, Hadoop also uses Jetty 9.4.53 > Ref - https://github.com/apache/hadoop/blob/trunk/hadoop-project/pom.xml#L40 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28040) Upgrade netty to 4.1.100 due to CVE-2023-44487
[ https://issues.apache.org/jira/browse/HIVE-28040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Anmol Sundaram updated HIVE-28040: -- Description: At the time of raising this, Hadoop and Tez also use 4.1.100.Final in their trunk Related - https://issues.apache.org/jira/browse/TEZ-4538 > Upgrade netty to 4.1.100 due to CVE-2023-44487 > -- > > Key: HIVE-28040 > URL: https://issues.apache.org/jira/browse/HIVE-28040 > Project: Hive > Issue Type: Task > Components: Hive >Reporter: Anmol Sundaram >Priority: Major > Labels: pull-request-available > > At the time of raising this, Hadoop and Tez also use 4.1.100.Final in their > trunk > Related - https://issues.apache.org/jira/browse/TEZ-4538 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28040) Upgrade netty to 4.1.100 due to CVE-2023-44487
[ https://issues.apache.org/jira/browse/HIVE-28040?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28040: -- Labels: pull-request-available (was: ) > Upgrade netty to 4.1.100 due to CVE-2023-44487 > -- > > Key: HIVE-28040 > URL: https://issues.apache.org/jira/browse/HIVE-28040 > Project: Hive > Issue Type: Task > Components: Hive >Reporter: Anmol Sundaram >Priority: Major > Labels: pull-request-available > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28041) Upgrade jetty to 9.4.53 due to CVE-2023-44487
[ https://issues.apache.org/jira/browse/HIVE-28041?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28041: -- Labels: pull-request-available (was: ) > Upgrade jetty to 9.4.53 due to CVE-2023-44487 > - > > Key: HIVE-28041 > URL: https://issues.apache.org/jira/browse/HIVE-28041 > Project: Hive > Issue Type: Task > Components: Hive >Reporter: Anmol Sundaram >Priority: Major > Labels: pull-request-available > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28042) DigestMD5 error during opening connection to HMS
Vikram Ahuja created HIVE-28042: --- Summary: DigestMD5 error during opening connection to HMS Key: HIVE-28042 URL: https://issues.apache.org/jira/browse/HIVE-28042 Project: Hive Issue Type: Bug Reporter: Vikram Ahuja Assignee: Vikram Ahuja Hello, In our deployment we are facing the following exception in the HMS logs when a HMS connection is opened from the HS2 in some cases leading to query failures: {code:java} 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring password at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) at com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) at org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) at org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) at org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.javA) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) at java.lang.Thread.run(Thread.java)Caused by: org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, realUser=*, issueDate=1705973286139, maxDate=1706578086139, sequenceNumber=3294063, masterKeyId=7601 at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) at org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) at com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) ... 15 more {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Work started] (HIVE-28042) DigestMD5 error during opening connection to HMS
[ https://issues.apache.org/jira/browse/HIVE-28042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Work on HIVE-28042 started by Vikram Ahuja. --- > DigestMD5 error during opening connection to HMS > > > Key: HIVE-28042 > URL: https://issues.apache.org/jira/browse/HIVE-28042 > Project: Hive > Issue Type: Bug >Reporter: Vikram Ahuja >Assignee: Vikram Ahuja >Priority: Major > > Hello, > In our deployment we are facing the following exception in the HMS logs when > a HMS connection is opened from the HS2 in some cases leading to query > failures: > {code:java} > 2024-01-24T02:11:21,324 ERROR [TThreadPoolServer WorkerProcess-760394]: > transport.TSaslTransport (TSaslTransport.java:open) - SASL negotiation > failurejavax.security.sasl.SaslException: DIGEST-MD5: IO error acquiring > password at > com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) > at > com.sun.security.sasl.digest.DigestMD5Server.evaluateResponse(DigestMD5Server.java) > at > org.apache.thrift.transport.TSaslTransport$SaslParticipant.evaluateChallengeOrResponse(TSaslTransport.java) > at > org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java) > at > org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java) > at > org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java) > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.javA) at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java) > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java) > at > org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java) > at java.lang.Thread.run(Thread.java)Caused by: > org.apache.hadoop.security.token.SecretManager$InvalidToken: token expired or > does not exist: HIVE_DELEGATION_TOKEN owner=***, renewer=***, > realUser=*, issueDate=1705973286139, maxDate=1706578086139, > sequenceNumber=3294063, masterKeyId=7601 at > org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) > at > org.apache.hadoop.hive.metastore.security.TokenStoreDelegationTokenSecretManager.retrievePassword(TokenStoreDelegationTokenSecretManager.java) > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.getPassword(HadoopThriftAuthBridge.java) > at > org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$SaslDigestCallbackHandler.handle(HadoopThriftAuthBridge.java) > at > com.sun.security.sasl.digest.DigestMD5Server.validateClientResponse(DigestMD5Server.java) > ... 15 more {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28041) Upgrade jetty to 9.4.53 due to CVE-2023-44487
Anmol Sundaram created HIVE-28041: - Summary: Upgrade jetty to 9.4.53 due to CVE-2023-44487 Key: HIVE-28041 URL: https://issues.apache.org/jira/browse/HIVE-28041 Project: Hive Issue Type: Task Components: Hive Reporter: Anmol Sundaram -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28040) Upgrade netty to 4.1.100 due to CVE-2023-44487
Anmol Sundaram created HIVE-28040: - Summary: Upgrade netty to 4.1.100 due to CVE-2023-44487 Key: HIVE-28040 URL: https://issues.apache.org/jira/browse/HIVE-28040 Project: Hive Issue Type: Task Components: Hive Reporter: Anmol Sundaram -- This message was sent by Atlassian Jira (v8.20.10#820010)