[jira] [Commented] (HIVE-27113) Increasing default for hive.thrift.client.max.message.size to 2 GB

2023-03-08 Thread Riju Trivedi (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-27113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17698231#comment-17698231
 ] 

Riju Trivedi commented on HIVE-27113:
-

Thrift message size can go beyond 1g for wide tables with too many partitions 
(5k+), Hence increasing the default to max value 2 GB.

> Increasing default for hive.thrift.client.max.message.size to 2 GB
> --
>
> Key: HIVE-27113
> URL: https://issues.apache.org/jira/browse/HIVE-27113
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Riju Trivedi
>Assignee: Riju Trivedi
>Priority: Major
>
> HIVE_THRIFT_CLIENT_MAX_MESSAGE_SIZE("hive.thrift.client.max.message.size", 
> "1gb",
> new SizeValidator(-1L, true, (long) Integer.MAX_VALUE, true),
> "Thrift client configuration for max message size. 0 or -1 will use 
> the default defined in the Thrift " +
> "library. The upper limit is 2147483648 bytes (or 2gb).")
> Documentation on the help suggests setting 2147483648 while Integer Max is 
> 2147483647. So, it actually becomes -1 and gets set to thrift default limit 
> (100 MB)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-27113) Increasing default for hive.thrift.client.max.message.size to 2 GB

2023-03-08 Thread Riju Trivedi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27113?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Riju Trivedi updated HIVE-27113:

Summary: Increasing default for hive.thrift.client.max.message.size to 2 GB 
 (was: Documentation for hive.thrift.client.max.message.size config needs to be 
corrected)

> Increasing default for hive.thrift.client.max.message.size to 2 GB
> --
>
> Key: HIVE-27113
> URL: https://issues.apache.org/jira/browse/HIVE-27113
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Reporter: Riju Trivedi
>Assignee: Riju Trivedi
>Priority: Major
>
> HIVE_THRIFT_CLIENT_MAX_MESSAGE_SIZE("hive.thrift.client.max.message.size", 
> "1gb",
> new SizeValidator(-1L, true, (long) Integer.MAX_VALUE, true),
> "Thrift client configuration for max message size. 0 or -1 will use 
> the default defined in the Thrift " +
> "library. The upper limit is 2147483648 bytes (or 2gb).")
> Documentation on the help suggests setting 2147483648 while Integer Max is 
> 2147483647. So, it actually becomes -1 and gets set to thrift default limit 
> (100 MB)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27129) Enhanced support to Hive Client http support

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27129?focusedWorklogId=849982&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849982
 ]

ASF GitHub Bot logged work on HIVE-27129:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 05:46
Start Date: 09/Mar/23 05:46
Worklog Time Spent: 10m 
  Work Description: vihangk1 commented on code in PR #4104:
URL: https://github.com/apache/hive/pull/4104#discussion_r1130492724


##
standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java:
##
@@ -35,25 +35,15 @@
 import java.net.UnknownHostException;
 import java.nio.ByteBuffer;
 import java.security.PrivilegedExceptionAction;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.Iterator;
-import java.util.LinkedHashMap;
-import java.util.List;
-import java.util.Map;
+import java.util.*;

Review Comment:
   We generally don't use wild-card imports. Can you please revert this part 
and explicitly add the necessary imports (You may have to change your IDE 
settings so that it doesn't do this for automatically).



##
standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java:
##
@@ -621,31 +611,84 @@ private  T 
configureThriftMaxMessageSize(T transport) {
 return transport;
   }
 
+  private Map getAdditionalHeaders() {
+Map headers = new HashMap<>();
+String keyValuePairs = MetastoreConf.getVar(conf, 
ConfVars.METASTORE_CLIENT_ADDITIONAL_HEADERS);
+try {
+  List headerKeyValues = 
Splitter.on(',').trimResults().splitToList(keyValuePairs);
+  for (String header : headerKeyValues) {
+String[] parts = header.split("=");
+headers.put(parts[0].trim(), parts[1].trim());
+LOG.warn(parts[0].trim() + "=" + parts[1].trim());

Review Comment:
   Is this log needed? If not, let's remove it or change it to debug.



##
standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java:
##
@@ -621,31 +611,84 @@ private  T 
configureThriftMaxMessageSize(T transport) {
 return transport;
   }
 
+  private Map getAdditionalHeaders() {
+Map headers = new HashMap<>();
+String keyValuePairs = MetastoreConf.getVar(conf, 
ConfVars.METASTORE_CLIENT_ADDITIONAL_HEADERS);
+try {
+  List headerKeyValues = 
Splitter.on(',').trimResults().splitToList(keyValuePairs);
+  for (String header : headerKeyValues) {
+String[] parts = header.split("=");
+headers.put(parts[0].trim(), parts[1].trim());
+LOG.warn(parts[0].trim() + "=" + parts[1].trim());
+  }
+} catch (Exception ex) {
+  LOG.warn("Could not parse the headers provided in " + 
ConfVars.METASTORE_CLIENT_ADDITIONAL_HEADERS, ex);
+}
+return headers;
+  }
+
   /*
   Creates a THttpClient if HTTP mode is enabled. If Client auth mode is set to 
JWT,
   then the method fetches JWT from environment variable: HMS_JWT and sets in 
auth
   header in http request
*/
-  private THttpClient createHttpClient(URI store, boolean useSSL) throws 
MetaException,
-  TTransportException {
+  private THttpClient createHttpClient(URI store, boolean useSSL) throws 
MetaException, TTransportException {
 String path = MetaStoreUtils.getHttpPath(MetastoreConf.getVar(conf, 
ConfVars.THRIFT_HTTP_PATH));
-String httpUrl = (useSSL ? "https://"; : "http://";) + store.getHost() + ":" 
+ store.getPort() + path;
+String urlScheme;
+if (useSSL || Objects.equals(store.getScheme(), "https")) {
+  urlScheme = "https://";;
+} else {
+  urlScheme = "http://";;
+}
+String httpUrl = urlScheme + store.getHost() + ":" + store.getPort() + 
path;
+
+HttpClientBuilder httpClientBuilder = createHttpClientBuilder();
+THttpClient tHttpClient;
+try {
+  if (useSSL) {
+String trustStorePath = MetastoreConf.getVar(conf, 
ConfVars.SSL_TRUSTSTORE_PATH).trim();
+if (trustStorePath.isEmpty()) {
+  throw new IllegalArgumentException(ConfVars.SSL_TRUSTSTORE_PATH + " 
Not configured for SSL connection");
+}
+String trustStorePassword = MetastoreConf.getPassword(conf, 
MetastoreConf.ConfVars.SSL_TRUSTSTORE_PASSWORD);
+String trustStoreType = MetastoreConf.getVar(conf, 
ConfVars.SSL_TRUSTSTORE_TYPE).trim();
+String trustStoreAlgorithm = MetastoreConf.getVar(conf, 
ConfVars.SSL_TRUSTMANAGERFACTORY_ALGORITHM).trim();
+tHttpClient =
+SecurityUtils.getThriftHttpsClient(httpUrl, trustStorePath, 
trustStorePassword, trustStoreAlgorithm,
+trustStoreType, httpClientBuilder);
+  } else {
+tHttpClient = new THttpClient(httpUrl, httpClientBuilder.build());
+  }
+} catch (

[jira] [Resolved] (HIVE-27107) Fix improper metrics count for RESUME/RESET workflow

2023-03-08 Thread Shreenidhi (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shreenidhi resolved HIVE-27107.
---
Resolution: Fixed

> Fix improper metrics count for RESUME/RESET workflow
> 
>
> Key: HIVE-27107
> URL: https://issues.apache.org/jira/browse/HIVE-27107
> Project: Hive
>  Issue Type: Task
>Reporter: Shreenidhi
>Assignee: Shreenidhi
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> *Analysis :* 
> REPL DUMP during RESET checks all the tables listed in table_diff_complete
> and persists the table count in replication metrics. This is incorrect for 
> RESET, 
> because table_diff_complete has those tables which needs to be dropped too.
> *Fix :* 
> To fix this, one can take intersection of (tables in table_diff_complete, 
> actual tables on source) and capture the same into the replication_metrics



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27123) In PRE_OPTIMIZED_BOOTSTRAP operation, Load side stores dump_execution_id as -1

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27123?focusedWorklogId=849979&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849979
 ]

ASF GitHub Bot logged work on HIVE-27123:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 05:23
Start Date: 09/Mar/23 05:23
Worklog Time Spent: 10m 
  Work Description: pudidic merged PR #4100:
URL: https://github.com/apache/hive/pull/4100




Issue Time Tracking
---

Worklog Id: (was: 849979)
Time Spent: 40m  (was: 0.5h)

> In PRE_OPTIMIZED_BOOTSTRAP operation, Load side stores dump_execution_id as -1
> --
>
> Key: HIVE-27123
> URL: https://issues.apache.org/jira/browse/HIVE-27123
> Project: Hive
>  Issue Type: Bug
>Reporter: Harshal Patel
>Assignee: Harshal Patel
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> For PRE_OPTIMIZED_BOOSTRAP operation, dump side correctly stores execution 
> ids in replication_metrics but Load side stores dump_execution_id as -1 which 
> should be = execution_id of DUMP operation



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27107) Fix improper metrics count for RESUME/RESET workflow

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27107?focusedWorklogId=849977&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849977
 ]

ASF GitHub Bot logged work on HIVE-27107:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 05:22
Start Date: 09/Mar/23 05:22
Worklog Time Spent: 10m 
  Work Description: pudidic commented on PR #4084:
URL: https://github.com/apache/hive/pull/4084#issuecomment-1461304559

   LGTM +1.




Issue Time Tracking
---

Worklog Id: (was: 849977)
Time Spent: 0.5h  (was: 20m)

> Fix improper metrics count for RESUME/RESET workflow
> 
>
> Key: HIVE-27107
> URL: https://issues.apache.org/jira/browse/HIVE-27107
> Project: Hive
>  Issue Type: Task
>Reporter: Shreenidhi
>Assignee: Shreenidhi
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> *Analysis :* 
> REPL DUMP during RESET checks all the tables listed in table_diff_complete
> and persists the table count in replication metrics. This is incorrect for 
> RESET, 
> because table_diff_complete has those tables which needs to be dropped too.
> *Fix :* 
> To fix this, one can take intersection of (tables in table_diff_complete, 
> actual tables on source) and capture the same into the replication_metrics



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27107) Fix improper metrics count for RESUME/RESET workflow

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27107?focusedWorklogId=849978&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849978
 ]

ASF GitHub Bot logged work on HIVE-27107:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 05:22
Start Date: 09/Mar/23 05:22
Worklog Time Spent: 10m 
  Work Description: pudidic merged PR #4084:
URL: https://github.com/apache/hive/pull/4084




Issue Time Tracking
---

Worklog Id: (was: 849978)
Time Spent: 40m  (was: 0.5h)

> Fix improper metrics count for RESUME/RESET workflow
> 
>
> Key: HIVE-27107
> URL: https://issues.apache.org/jira/browse/HIVE-27107
> Project: Hive
>  Issue Type: Task
>Reporter: Shreenidhi
>Assignee: Shreenidhi
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> *Analysis :* 
> REPL DUMP during RESET checks all the tables listed in table_diff_complete
> and persists the table count in replication metrics. This is incorrect for 
> RESET, 
> because table_diff_complete has those tables which needs to be dropped too.
> *Fix :* 
> To fix this, one can take intersection of (tables in table_diff_complete, 
> actual tables on source) and capture the same into the replication_metrics



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27123) In PRE_OPTIMIZED_BOOTSTRAP operation, Load side stores dump_execution_id as -1

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27123?focusedWorklogId=849976&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849976
 ]

ASF GitHub Bot logged work on HIVE-27123:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 05:21
Start Date: 09/Mar/23 05:21
Worklog Time Spent: 10m 
  Work Description: pudidic commented on PR #4100:
URL: https://github.com/apache/hive/pull/4100#issuecomment-1461303691

   +1. LGTM.




Issue Time Tracking
---

Worklog Id: (was: 849976)
Time Spent: 0.5h  (was: 20m)

> In PRE_OPTIMIZED_BOOTSTRAP operation, Load side stores dump_execution_id as -1
> --
>
> Key: HIVE-27123
> URL: https://issues.apache.org/jira/browse/HIVE-27123
> Project: Hive
>  Issue Type: Bug
>Reporter: Harshal Patel
>Assignee: Harshal Patel
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> For PRE_OPTIMIZED_BOOSTRAP operation, dump side correctly stores execution 
> ids in replication_metrics but Load side stores dump_execution_id as -1 which 
> should be = execution_id of DUMP operation



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (HIVE-26919) Fix-test-case-TestReplicationOptimisedBootstrap.testReverseFailoverBeforeOptimizedBootstrap

2023-03-08 Thread Amit Saonerkar (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-26919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Amit Saonerkar resolved HIVE-26919.
---
Resolution: Fixed

> Fix-test-case-TestReplicationOptimisedBootstrap.testReverseFailoverBeforeOptimizedBootstrap
> ---
>
> Key: HIVE-26919
> URL: https://issues.apache.org/jira/browse/HIVE-26919
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Amit Saonerkar
>Assignee: Amit Saonerkar
>Priority: Minor
>
> This Jira is related to test case failure corresponding to Jira CDPD-48053



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (HIVE-26919) Fix-test-case-TestReplicationOptimisedBootstrap.testReverseFailoverBeforeOptimizedBootstrap

2023-03-08 Thread Amit Saonerkar (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-26919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17698179#comment-17698179
 ] 

Amit Saonerkar commented on HIVE-26919:
---

https://github.com/apache/hive/commit/0c059c6e9bba35f353cd3212c5c01e86eba51125

> Fix-test-case-TestReplicationOptimisedBootstrap.testReverseFailoverBeforeOptimizedBootstrap
> ---
>
> Key: HIVE-26919
> URL: https://issues.apache.org/jira/browse/HIVE-26919
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Reporter: Amit Saonerkar
>Assignee: Amit Saonerkar
>Priority: Minor
>
> This Jira is related to test case failure corresponding to Jira CDPD-48053



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (HIVE-26920) Add-new-view-in-sys-db-to-capture-failover-and-failback-metrics

2023-03-08 Thread Amit Saonerkar (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-26920?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Amit Saonerkar resolved HIVE-26920.
---
Resolution: Fixed

> Add-new-view-in-sys-db-to-capture-failover-and-failback-metrics
> ---
>
> Key: HIVE-26920
> URL: https://issues.apache.org/jira/browse/HIVE-26920
> Project: Hive
>  Issue Type: New Feature
>  Components: HiveServer2
>Reporter: Amit Saonerkar
>Assignee: Amit Saonerkar
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> Corresponding to CDPD-46702



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (HIVE-26921) Add failover_type, failover_endpoint to replication metrics metadata

2023-03-08 Thread Amit Saonerkar (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-26921?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Amit Saonerkar resolved HIVE-26921.
---
Resolution: Fixed

> Add failover_type, failover_endpoint to replication metrics metadata
> 
>
> Key: HIVE-26921
> URL: https://issues.apache.org/jira/browse/HIVE-26921
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Amit Saonerkar
>Assignee: Amit Saonerkar
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> Corresponding to CDPD-46494



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27129) Enhanced support to Hive Client http support

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27129?focusedWorklogId=849975&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849975
 ]

ASF GitHub Bot logged work on HIVE-27129:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 05:08
Start Date: 09/Mar/23 05:08
Worklog Time Spent: 10m 
  Work Description: junlinzeng-db opened a new pull request, #4104:
URL: https://github.com/apache/hive/pull/4104

   
   
   ### What changes were proposed in this pull request?
   
   
   
   ### Why are the changes needed?
   
   
   
   ### Does this PR introduce _any_ user-facing change?
   
   
   This PR targets to improve the usability of the http hive metastore client, 
by adding support for custom headers and also, supporting default trust store 
for https.
   
   1. Add a new hive conf to allow users to inject custom headers
   2. Do not force the headers to be http when the useSSL is turned off, so 
that if the users are able to use https if they want, w/o a custom truststore.
   
   
   ### How was this patch tested?
   
   
   ut for the http headers
   




Issue Time Tracking
---

Worklog Id: (was: 849975)
Remaining Estimate: 0h
Time Spent: 10m

> Enhanced support to Hive Client http support
> 
>
> Key: HIVE-27129
> URL: https://issues.apache.org/jira/browse/HIVE-27129
> Project: Hive
>  Issue Type: Improvement
>Reporter: Junlin Zeng
>Assignee: Junlin Zeng
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently we support using http in the hive metastore connection. However, we 
> do not support custom headers and also default trust store. This ticket 
> tracks the work to improve the http journey.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-27129) Enhanced support to Hive Client http support

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27129?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-27129:
--
Labels: pull-request-available  (was: )

> Enhanced support to Hive Client http support
> 
>
> Key: HIVE-27129
> URL: https://issues.apache.org/jira/browse/HIVE-27129
> Project: Hive
>  Issue Type: Improvement
>Reporter: Junlin Zeng
>Assignee: Junlin Zeng
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently we support using http in the hive metastore connection. However, we 
> do not support custom headers and also default trust store. This ticket 
> tracks the work to improve the http journey.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27087) Fix TestMiniSparkOnYarnCliDriver test failures on branch-3

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27087?focusedWorklogId=849966&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849966
 ]

ASF GitHub Bot logged work on HIVE-27087:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 03:57
Start Date: 09/Mar/23 03:57
Worklog Time Spent: 10m 
  Work Description: vihangk1 commented on PR #4067:
URL: https://github.com/apache/hive/pull/4067#issuecomment-1461235075

   @zabetak I am running out of ideas here. Any idea why the replication factor 
seen in the q.out is 1 instead of 3. I cannot reproduce it on my local. All the 
q.out file diff failures work for me on my local.




Issue Time Tracking
---

Worklog Id: (was: 849966)
Time Spent: 2h 40m  (was: 2.5h)

> Fix TestMiniSparkOnYarnCliDriver test failures on branch-3
> --
>
> Key: HIVE-27087
> URL: https://issues.apache.org/jira/browse/HIVE-27087
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2h 40m
>  Remaining Estimate: 0h
>
> TestMiniSparkOnYarnCliDriver are failing with the error below
> [ERROR] 2023-02-16 14:13:08.991 [Driver] SparkContext - Error initializing 
> SparkContext.
> java.lang.RuntimeException: java.lang.NoSuchFieldException: 
> DEFAULT_TINY_CACHE_SIZE
> at 
> org.apache.spark.network.util.NettyUtils.getPrivateStaticField(NettyUtils.java:131)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.util.NettyUtils.createPooledByteBufAllocator(NettyUtils.java:118)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.server.TransportServer.init(TransportServer.java:94) 
> ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.server.TransportServer.(TransportServer.java:73)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.TransportContext.createServer(TransportContext.java:114)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.rpc.netty.NettyRpcEnv.startServer(NettyRpcEnv.scala:119) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory$$anonfun$4.apply(NettyRpcEnv.scala:465)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory$$anonfun$4.apply(NettyRpcEnv.scala:464)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:2271)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160) 
> ~[scala-library-2.11.8.jar:?]
> at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:2263) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory.create(NettyRpcEnv.scala:469) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.rpc.RpcEnv$.create(RpcEnv.scala:57) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkEnv$.create(SparkEnv.scala:249) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:175) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:256) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkContext.(SparkContext.scala:423) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.api.java.JavaSparkContext.(JavaSparkContext.scala:58) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.hive.spark.client.RemoteDriver.(RemoteDriver.java:161) 
> [hive-exec-3.2.0-SNAPSHOT.jar:3.2.0-SNAPSHOT]
> at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:536) 
> [hive-exec-3.2.0-SNAPSHOT.jar:3.2.0-SNAPSHOT]
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_322]
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> ~[?:1.8.0_322]
> The root cause of the problem is that we upgrade the netty library from 
> 4.1.17.Final to 4.1.69.Final. The upgraded library does not have 
> `DEFAULT_TINY_CACHE_SIZE` field 
> [here|https://github.com/netty/netty/blob/netty-4.1.51.Final/buffer/src/main/java/io/netty/buffer/PooledByteBufAllocator.java#L46]
>  which was removed in 4.1.52.Final



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27087) Fix TestMiniSparkOnYarnCliDriver test failures on branch-3

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27087?focusedWorklogId=849965&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849965
 ]

ASF GitHub Bot logged work on HIVE-27087:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 03:51
Start Date: 09/Mar/23 03:51
Worklog Time Spent: 10m 
  Work Description: amanraj2520 commented on PR #4067:
URL: https://github.com/apache/hive/pull/4067#issuecomment-1461230107

   I cannot repro in my local as well. Checking still.




Issue Time Tracking
---

Worklog Id: (was: 849965)
Time Spent: 2.5h  (was: 2h 20m)

> Fix TestMiniSparkOnYarnCliDriver test failures on branch-3
> --
>
> Key: HIVE-27087
> URL: https://issues.apache.org/jira/browse/HIVE-27087
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2.5h
>  Remaining Estimate: 0h
>
> TestMiniSparkOnYarnCliDriver are failing with the error below
> [ERROR] 2023-02-16 14:13:08.991 [Driver] SparkContext - Error initializing 
> SparkContext.
> java.lang.RuntimeException: java.lang.NoSuchFieldException: 
> DEFAULT_TINY_CACHE_SIZE
> at 
> org.apache.spark.network.util.NettyUtils.getPrivateStaticField(NettyUtils.java:131)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.util.NettyUtils.createPooledByteBufAllocator(NettyUtils.java:118)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.server.TransportServer.init(TransportServer.java:94) 
> ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.server.TransportServer.(TransportServer.java:73)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.TransportContext.createServer(TransportContext.java:114)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.rpc.netty.NettyRpcEnv.startServer(NettyRpcEnv.scala:119) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory$$anonfun$4.apply(NettyRpcEnv.scala:465)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory$$anonfun$4.apply(NettyRpcEnv.scala:464)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:2271)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160) 
> ~[scala-library-2.11.8.jar:?]
> at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:2263) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory.create(NettyRpcEnv.scala:469) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.rpc.RpcEnv$.create(RpcEnv.scala:57) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkEnv$.create(SparkEnv.scala:249) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:175) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:256) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkContext.(SparkContext.scala:423) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.api.java.JavaSparkContext.(JavaSparkContext.scala:58) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.hive.spark.client.RemoteDriver.(RemoteDriver.java:161) 
> [hive-exec-3.2.0-SNAPSHOT.jar:3.2.0-SNAPSHOT]
> at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:536) 
> [hive-exec-3.2.0-SNAPSHOT.jar:3.2.0-SNAPSHOT]
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_322]
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> ~[?:1.8.0_322]
> The root cause of the problem is that we upgrade the netty library from 
> 4.1.17.Final to 4.1.69.Final. The upgraded library does not have 
> `DEFAULT_TINY_CACHE_SIZE` field 
> [here|https://github.com/netty/netty/blob/netty-4.1.51.Final/buffer/src/main/java/io/netty/buffer/PooledByteBufAllocator.java#L46]
>  which was removed in 4.1.52.Final



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (HIVE-26659) TPC-DS query 16, 69, 94 return wrong results.

2023-03-08 Thread Steve Carlin (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-26659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17698134#comment-17698134
 ] 

Steve Carlin commented on HIVE-26659:
-

Filed HIVE-27124 to temporarily turn off the anti-join feature by default until 
this is fixed. 

> TPC-DS query 16, 69, 94 return wrong results.
> -
>
> Key: HIVE-26659
> URL: https://issues.apache.org/jira/browse/HIVE-26659
> Project: Hive
>  Issue Type: Sub-task
>Affects Versions: 4.0.0-alpha-2
>Reporter: Sungwoo Park
>Priority: Major
>
> TPC-DS query 16, 69, 94 return wrong results when hive.auto.convert.anti.join 
> is set to true.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work started] (HIVE-27129) Enhanced support to Hive Client http support

2023-03-08 Thread Junlin Zeng (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27129?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HIVE-27129 started by Junlin Zeng.
--
> Enhanced support to Hive Client http support
> 
>
> Key: HIVE-27129
> URL: https://issues.apache.org/jira/browse/HIVE-27129
> Project: Hive
>  Issue Type: Improvement
>Reporter: Junlin Zeng
>Assignee: Junlin Zeng
>Priority: Major
>
> Currently we support using http in the hive metastore connection. However, we 
> do not support custom headers and also default trust store. This ticket 
> tracks the work to improve the http journey.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (HIVE-27129) Enhanced support to Hive Client http support

2023-03-08 Thread Junlin Zeng (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27129?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junlin Zeng reassigned HIVE-27129:
--


> Enhanced support to Hive Client http support
> 
>
> Key: HIVE-27129
> URL: https://issues.apache.org/jira/browse/HIVE-27129
> Project: Hive
>  Issue Type: Improvement
>Reporter: Junlin Zeng
>Assignee: Junlin Zeng
>Priority: Major
>
> Currently we support using http in the hive metastore connection. However, we 
> do not support custom headers and also default trust store. This ticket 
> tracks the work to improve the http journey.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27089) Update permissions in orc_merge*.q.out files

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27089?focusedWorklogId=849933&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849933
 ]

ASF GitHub Bot logged work on HIVE-27089:
-

Author: ASF GitHub Bot
Created on: 09/Mar/23 00:22
Start Date: 09/Mar/23 00:22
Worklog Time Spent: 10m 
  Work Description: yigress commented on PR #4068:
URL: https://github.com/apache/hive/pull/4068#issuecomment-1461074375

   @zabetak thank you for taking time detailed analysis!




Issue Time Tracking
---

Worklog Id: (was: 849933)
Time Spent: 2h 50m  (was: 2h 40m)

> Update permissions in orc_merge*.q.out files
> 
>
> Key: HIVE-27089
> URL: https://issues.apache.org/jira/browse/HIVE-27089
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Aman Raj
>Assignee: Aman Raj
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.2.0
>
>  Time Spent: 2h 50m
>  Remaining Estimate: 0h
>
> The orc_merge1,2,3,4,10 .q tests fails with the following errors : 
> < -rw-rw-rw- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> HDFS PATH ###
> ---
> > -rw-r--r-- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> > HDFS PATH ###



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (HIVE-27128) Exception "Can't finish byte read from uncompressed stream DATA position" when querying ORC table

2023-03-08 Thread Dmitriy Fingerman (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27128?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmitriy Fingerman reassigned HIVE-27128:



> Exception "Can't finish byte read from uncompressed stream DATA position" 
> when querying ORC table
> -
>
> Key: HIVE-27128
> URL: https://issues.apache.org/jira/browse/HIVE-27128
> Project: Hive
>  Issue Type: Bug
>Reporter: Dmitriy Fingerman
>Assignee: Dmitriy Fingerman
>Priority: Critical
>
> Exception happening when querying an ORC table:
> {code:java}
> Caused by: java.io.EOFException: Can't finish byte read from uncompressed 
> stream DATA position: 393216 length: 393216 range: 23 offset: 376832 
> position: 16384 limit: 16384
>   at 
> org.apache.orc.impl.TreeReaderFactory$BytesColumnVectorUtil.commonReadByteArrays(TreeReaderFactory.java:1550)
>   at 
> org.apache.orc.impl.TreeReaderFactory$BytesColumnVectorUtil.readOrcByteArrays(TreeReaderFactory.java:1566)
>   at 
> org.apache.orc.impl.TreeReaderFactory$StringDirectTreeReader.nextVector(TreeReaderFactory.java:1662)
>   at 
> org.apache.orc.impl.TreeReaderFactory$StringTreeReader.nextVector(TreeReaderFactory.java:1508)
>   at 
> org.apache.hadoop.hive.ql.io.orc.encoded.EncodedTreeReaderFactory$StringStreamReader.nextVector(EncodedTreeReaderFactory.java:305)
>   at 
> org.apache.hadoop.hive.llap.io.decode.OrcEncodedDataConsumer.decodeBatch(OrcEncodedDataConsumer.java:196)
>   at 
> org.apache.hadoop.hive.llap.io.decode.OrcEncodedDataConsumer.decodeBatch(OrcEncodedDataConsumer.java:66)
>   at 
> org.apache.hadoop.hive.llap.io.decode.EncodedDataConsumer.consumeData(EncodedDataConsumer.java:122)
>   at 
> org.apache.hadoop.hive.llap.io.decode.EncodedDataConsumer.consumeData(EncodedDataConsumer.java:42)
>   at 
> org.apache.hadoop.hive.ql.io.orc.encoded.EncodedReaderImpl.readEncodedColumns(EncodedReaderImpl.java:608)
>   at 
> org.apache.hadoop.hive.llap.io.encoded.OrcEncodedDataReader.performDataRead(OrcEncodedDataReader.java:434)
>   at 
> org.apache.hadoop.hive.llap.io.encoded.OrcEncodedDataReader$4.run(OrcEncodedDataReader.java:282)
>   at 
> org.apache.hadoop.hive.llap.io.encoded.OrcEncodedDataReader$4.run(OrcEncodedDataReader.java:279)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1878)
>   at 
> org.apache.hadoop.hive.llap.io.encoded.OrcEncodedDataReader.callInternal(OrcEncodedDataReader.java:279)
>   at 
> org.apache.hadoop.hive.llap.io.encoded.OrcEncodedDataReader.callInternal(OrcEncodedDataReader.java:118)
>   at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36)
>   at 
> org.apache.hadoop.hive.llap.io.decode.EncodedDataConsumer$CpuRecordingCallable.call(EncodedDataConsumer.java:88)
>   at 
> org.apache.hadoop.hive.llap.io.decode.EncodedDataConsumer$CpuRecordingCallable.call(EncodedDataConsumer.java:73)
>  {code}
> I created a q-test that reproduces this issue:
> [https://github.com/difin/hive/commits/orc_read_err_qtest]
> This issue happens in Hive starting from the commit that upgraded ORC version 
> in Hive to ORC 1.6.7.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27124) Temporarily disable default usage of anti-join

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27124?focusedWorklogId=849914&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849914
 ]

ASF GitHub Bot logged work on HIVE-27124:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 21:00
Start Date: 08/Mar/23 21:00
Worklog Time Spent: 10m 
  Work Description: sonarcloud[bot] commented on PR #4101:
URL: https://github.com/apache/hive/pull/4101#issuecomment-1460869110

   Kudos, SonarCloud Quality Gate passed!    [![Quality Gate 
passed](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/QualityGateBadge/passed-16px.png
 'Quality Gate 
passed')](https://sonarcloud.io/dashboard?id=apache_hive&pullRequest=4101)
   
   
[![Bug](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/bug-16px.png
 
'Bug')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=BUG)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=BUG)
 [0 
Bugs](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=BUG)
  
   
[![Vulnerability](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/vulnerability-16px.png
 
'Vulnerability')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=VULNERABILITY)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=VULNERABILITY)
 [0 
Vulnerabilities](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=VULNERABILITY)
  
   [![Security 
Hotspot](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/security_hotspot-16px.png
 'Security 
Hotspot')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4101&resolved=false&types=SECURITY_HOTSPOT)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4101&resolved=false&types=SECURITY_HOTSPOT)
 [0 Security 
Hotspots](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4101&resolved=false&types=SECURITY_HOTSPOT)
  
   [![Code 
Smell](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/code_smell-16px.png
 'Code 
Smell')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=CODE_SMELL)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=CODE_SMELL)
 [0 Code 
Smells](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4101&resolved=false&types=CODE_SMELL)
   
   [![No Coverage 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/CoverageChart/NoCoverageInfo-16px.png
 'No Coverage 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4101&metric=coverage&view=list)
 No Coverage information  
   [![No Duplication 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/Duplications/NoDuplicationInfo-16px.png
 'No Duplication 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4101&metric=duplicated_lines_density&view=list)
 No Duplication information
   
   




Issue Time Tracking
---

Worklog Id: (was: 849914)
Time Spent: 0.5h  (was: 20m)

> Temporarily disable default usage of anti-join
> --
>
> Key: HIVE-27124
> URL: https://issues.apache.org/jira/browse/HIVE-27124
> Project: Hive
>  Issue Type: Improvement
>  Components: Hive
>Reporter: Steve Carlin
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Temporarily disabling the default usage for antijoin due to wrong results as 
> documented by HIVE-26659



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27087) Fix TestMiniSparkOnYarnCliDriver test failures on branch-3

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27087?focusedWorklogId=849902&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849902
 ]

ASF GitHub Bot logged work on HIVE-27087:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 19:47
Start Date: 08/Mar/23 19:47
Worklog Time Spent: 10m 
  Work Description: vihangk1 commented on PR #4067:
URL: https://github.com/apache/hive/pull/4067#issuecomment-1460758753

   The diffs that we see are due to the a difference in replication factor of 
the files on the hdfs. I wonder if the mini HDFS for the  tests is slow to 
replicate the blocks for these files. Added a sleep to confirm if that helps 
because I cannot reproduce it in my local even after running all the tests in 
the split.




Issue Time Tracking
---

Worklog Id: (was: 849902)
Time Spent: 2h 20m  (was: 2h 10m)

> Fix TestMiniSparkOnYarnCliDriver test failures on branch-3
> --
>
> Key: HIVE-27087
> URL: https://issues.apache.org/jira/browse/HIVE-27087
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Vihang Karajgaonkar
>Assignee: Vihang Karajgaonkar
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> TestMiniSparkOnYarnCliDriver are failing with the error below
> [ERROR] 2023-02-16 14:13:08.991 [Driver] SparkContext - Error initializing 
> SparkContext.
> java.lang.RuntimeException: java.lang.NoSuchFieldException: 
> DEFAULT_TINY_CACHE_SIZE
> at 
> org.apache.spark.network.util.NettyUtils.getPrivateStaticField(NettyUtils.java:131)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.util.NettyUtils.createPooledByteBufAllocator(NettyUtils.java:118)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.server.TransportServer.init(TransportServer.java:94) 
> ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.server.TransportServer.(TransportServer.java:73)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.network.TransportContext.createServer(TransportContext.java:114)
>  ~[spark-network-common_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.rpc.netty.NettyRpcEnv.startServer(NettyRpcEnv.scala:119) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory$$anonfun$4.apply(NettyRpcEnv.scala:465)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory$$anonfun$4.apply(NettyRpcEnv.scala:464)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:2271)
>  ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160) 
> ~[scala-library-2.11.8.jar:?]
> at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:2263) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.rpc.netty.NettyRpcEnvFactory.create(NettyRpcEnv.scala:469) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.rpc.RpcEnv$.create(RpcEnv.scala:57) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkEnv$.create(SparkEnv.scala:249) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:175) 
> ~[spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:256) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.spark.SparkContext.(SparkContext.scala:423) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at 
> org.apache.spark.api.java.JavaSparkContext.(JavaSparkContext.scala:58) 
> [spark-core_2.11-2.3.0.jar:2.3.0]
> at org.apache.hive.spark.client.RemoteDriver.(RemoteDriver.java:161) 
> [hive-exec-3.2.0-SNAPSHOT.jar:3.2.0-SNAPSHOT]
> at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:536) 
> [hive-exec-3.2.0-SNAPSHOT.jar:3.2.0-SNAPSHOT]
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_322]
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> ~[?:1.8.0_322]
> The root cause of the problem is that we upgrade the netty library from 
> 4.1.17.Final to 4.1.69.Final. The upgraded library does not have 
> `DEFAULT_TINY_CACHE_SIZE` field 
> [here|https://github.com/netty/netty/blob/netty-4.1.51.Final/buffer/src/main/java/io/netty/buffer/PooledByteBufAllocator.java#L46]
>  which was removed in 4.1.52.Final



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (HIVE-26735) Ability to sort the data during rebalancing compaction

2023-03-08 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-26735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Végh resolved HIVE-26735.

Resolution: Fixed

Merged to master, [~kkasa] , [~sbadhya] , [~dkuzmenko] thanks for the review!

> Ability to sort the data during rebalancing compaction
> --
>
> Key: HIVE-26735
> URL: https://issues.apache.org/jira/browse/HIVE-26735
> Project: Hive
>  Issue Type: Sub-task
>Reporter: László Végh
>Assignee: László Végh
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 6h 10m
>  Remaining Estimate: 0h
>
> Ability to sort the data during rebalancing compaction



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-26735) Ability to sort the data during rebalancing compaction

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-26735?focusedWorklogId=849893&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849893
 ]

ASF GitHub Bot logged work on HIVE-26735:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 18:14
Start Date: 08/Mar/23 18:14
Worklog Time Spent: 10m 
  Work Description: veghlaci05 merged PR #4077:
URL: https://github.com/apache/hive/pull/4077




Issue Time Tracking
---

Worklog Id: (was: 849893)
Time Spent: 6h 10m  (was: 6h)

> Ability to sort the data during rebalancing compaction
> --
>
> Key: HIVE-26735
> URL: https://issues.apache.org/jira/browse/HIVE-26735
> Project: Hive
>  Issue Type: Sub-task
>Reporter: László Végh
>Assignee: László Végh
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 6h 10m
>  Remaining Estimate: 0h
>
> Ability to sort the data during rebalancing compaction



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27013) Provide an option to enable iceberg manifest caching for all catalogs

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27013?focusedWorklogId=849892&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849892
 ]

ASF GitHub Bot logged work on HIVE-27013:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 18:13
Start Date: 08/Mar/23 18:13
Worklog Time Spent: 10m 
  Work Description: sonarcloud[bot] commented on PR #4088:
URL: https://github.com/apache/hive/pull/4088#issuecomment-1460631926

   Kudos, SonarCloud Quality Gate passed!    [![Quality Gate 
passed](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/QualityGateBadge/passed-16px.png
 'Quality Gate 
passed')](https://sonarcloud.io/dashboard?id=apache_hive&pullRequest=4088)
   
   
[![Bug](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/bug-16px.png
 
'Bug')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=BUG)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=BUG)
 [0 
Bugs](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=BUG)
  
   
[![Vulnerability](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/vulnerability-16px.png
 
'Vulnerability')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=VULNERABILITY)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=VULNERABILITY)
 [0 
Vulnerabilities](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=VULNERABILITY)
  
   [![Security 
Hotspot](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/security_hotspot-16px.png
 'Security 
Hotspot')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4088&resolved=false&types=SECURITY_HOTSPOT)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4088&resolved=false&types=SECURITY_HOTSPOT)
 [0 Security 
Hotspots](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4088&resolved=false&types=SECURITY_HOTSPOT)
  
   [![Code 
Smell](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/code_smell-16px.png
 'Code 
Smell')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=CODE_SMELL)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=CODE_SMELL)
 [0 Code 
Smells](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4088&resolved=false&types=CODE_SMELL)
   
   [![No Coverage 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/CoverageChart/NoCoverageInfo-16px.png
 'No Coverage 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4088&metric=coverage&view=list)
 No Coverage information  
   [![No Duplication 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/Duplications/NoDuplicationInfo-16px.png
 'No Duplication 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4088&metric=duplicated_lines_density&view=list)
 No Duplication information
   
   




Issue Time Tracking
---

Worklog Id: (was: 849892)
Time Spent: 3h  (was: 2h 50m)

> Provide an option to enable iceberg manifest caching for all catalogs
> -
>
> Key: HIVE-27013
> URL: https://issues.apache.org/jira/browse/HIVE-27013
> Project: Hive
>  Issue Type: Improvement
>  Components: Iceberg integration
>Reporter: Rajesh Balamohan
>Assignee: Ayush Saxena
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 3h
>  Remaining Estimate: 0h
>
> {color:#22}I tried the following thinking that it would work with iceberg 
> manifest caching; but it didn't.{color}
> {noformat}
> alter table store_sales set 
> tblproperties('io.manifest.cache-enabled'='true');{noformat}
> {color:#22}Creating this ticket as a placeholder to fix the same.{color}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (HIVE-27126) Display Yarn resources availability in real time along with Tez job progress

2023-03-08 Thread Jira


[ 
https://issues.apache.org/jira/browse/HIVE-27126?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17698031#comment-17698031
 ] 

László Bodor commented on HIVE-27126:
-

+1 on improving real-time data in beeline progress, let me link HIVE-25666 here

> Display Yarn resources availability in real time along with Tez job progress 
> -
>
> Key: HIVE-27126
> URL: https://issues.apache.org/jira/browse/HIVE-27126
> Project: Hive
>  Issue Type: Improvement
>  Components: HiveServer2
>Reporter: Taraka Rama Rao Lethavadla
>Priority: Major
>
> Right we don't know if the queue used in Hive query execution is having 
> enough resources or not. We are already displaying tez job details like 
> number of task per each vertex and how are they progressing. 
> If the resources available are not good enough to execute the query in 
> parallel or query that use to take shorter time is taking time, then we can 
> use this new detail to understand that queue is busy and resources are not 
> enough
>  
>  
>  
> {noformat}
> --
> VERTICES      MODE        STATUS  TOTAL  COMPLETED  RUNNING  PENDING  
> FAILED  KILLED
> --
> Map 1            container       RUNNING    230          0        10      230 
>     0       1
> Reducer 2        container        INITED      1          0        0        1  
>      0       0
> --
> VERTICES: 00/02  [>>--] 0%    ELAPSED TIME: 58.33 s
> --
> {noformat}
> So it would be great to display current resource availability in that 
> specific queue along with above info in real time would help understand why a 
> query can be slow
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27020) Implement a separate handler to handle aborted transaction cleanup

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27020?focusedWorklogId=849846&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849846
 ]

ASF GitHub Bot logged work on HIVE-27020:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 16:36
Start Date: 08/Mar/23 16:36
Worklog Time Spent: 10m 
  Work Description: sonarcloud[bot] commented on PR #4091:
URL: https://github.com/apache/hive/pull/4091#issuecomment-1460470703

   Kudos, SonarCloud Quality Gate passed!    [![Quality Gate 
passed](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/QualityGateBadge/passed-16px.png
 'Quality Gate 
passed')](https://sonarcloud.io/dashboard?id=apache_hive&pullRequest=4091)
   
   
[![Bug](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/bug-16px.png
 
'Bug')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=BUG)
 
[![E](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/E-16px.png
 
'E')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=BUG)
 [2 
Bugs](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=BUG)
  
   
[![Vulnerability](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/vulnerability-16px.png
 
'Vulnerability')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=VULNERABILITY)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=VULNERABILITY)
 [0 
Vulnerabilities](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=VULNERABILITY)
  
   [![Security 
Hotspot](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/security_hotspot-16px.png
 'Security 
Hotspot')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4091&resolved=false&types=SECURITY_HOTSPOT)
 
[![E](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/E-16px.png
 
'E')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4091&resolved=false&types=SECURITY_HOTSPOT)
 [1 Security 
Hotspot](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4091&resolved=false&types=SECURITY_HOTSPOT)
  
   [![Code 
Smell](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/code_smell-16px.png
 'Code 
Smell')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=CODE_SMELL)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=CODE_SMELL)
 [14 Code 
Smells](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4091&resolved=false&types=CODE_SMELL)
   
   [![No Coverage 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/CoverageChart/NoCoverageInfo-16px.png
 'No Coverage 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4091&metric=coverage&view=list)
 No Coverage information  
   [![No Duplication 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/Duplications/NoDuplicationInfo-16px.png
 'No Duplication 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4091&metric=duplicated_lines_density&view=list)
 No Duplication information
   
   




Issue Time Tracking
---

Worklog Id: (was: 849846)
Time Spent: 50m  (was: 40m)

> Implement a separate handler to handle aborted transaction cleanup
> --
>
> Key: HIVE-27020
> URL: https://issues.apache.org/jira/browse/HIVE-27020
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Sourabh Badhya
>Assignee: Sourabh Badhya
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> As described in the parent task, once the cleaner is separated into different 
> entities, implement a separate handler which can create requests for aborted 
> transactions cleanup. This would move the aborted transaction cleanup 
> exclusively to the cleaner.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-26735) Ability to sort the data during rebalancing compaction

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-26735?focusedWorklogId=849830&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849830
 ]

ASF GitHub Bot logged work on HIVE-26735:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 15:02
Start Date: 08/Mar/23 15:02
Worklog Time Spent: 10m 
  Work Description: sonarcloud[bot] commented on PR #4077:
URL: https://github.com/apache/hive/pull/4077#issuecomment-1460293044

   Kudos, SonarCloud Quality Gate passed!    [![Quality Gate 
passed](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/QualityGateBadge/passed-16px.png
 'Quality Gate 
passed')](https://sonarcloud.io/dashboard?id=apache_hive&pullRequest=4077)
   
   
[![Bug](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/bug-16px.png
 
'Bug')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=BUG)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=BUG)
 [0 
Bugs](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=BUG)
  
   
[![Vulnerability](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/vulnerability-16px.png
 
'Vulnerability')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=VULNERABILITY)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=VULNERABILITY)
 [0 
Vulnerabilities](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=VULNERABILITY)
  
   [![Security 
Hotspot](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/security_hotspot-16px.png
 'Security 
Hotspot')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4077&resolved=false&types=SECURITY_HOTSPOT)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4077&resolved=false&types=SECURITY_HOTSPOT)
 [0 Security 
Hotspots](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4077&resolved=false&types=SECURITY_HOTSPOT)
  
   [![Code 
Smell](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/code_smell-16px.png
 'Code 
Smell')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=CODE_SMELL)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=CODE_SMELL)
 [15 Code 
Smells](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4077&resolved=false&types=CODE_SMELL)
   
   [![No Coverage 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/CoverageChart/NoCoverageInfo-16px.png
 'No Coverage 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4077&metric=coverage&view=list)
 No Coverage information  
   [![No Duplication 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/Duplications/NoDuplicationInfo-16px.png
 'No Duplication 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4077&metric=duplicated_lines_density&view=list)
 No Duplication information
   
   




Issue Time Tracking
---

Worklog Id: (was: 849830)
Time Spent: 6h  (was: 5h 50m)

> Ability to sort the data during rebalancing compaction
> --
>
> Key: HIVE-26735
> URL: https://issues.apache.org/jira/browse/HIVE-26735
> Project: Hive
>  Issue Type: Sub-task
>Reporter: László Végh
>Assignee: László Végh
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 6h
>  Remaining Estimate: 0h
>
> Ability to sort the data during rebalancing compaction



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (HIVE-27089) Update permissions in orc_merge*.q.out files

2023-03-08 Thread Stamatis Zampetakis (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stamatis Zampetakis resolved HIVE-27089.

Fix Version/s: 3.2.0
   Resolution: Fixed

Fixed in 
https://github.com/apache/hive/commit/c0ec46292b18242f5a65df14141910d8197023d6. 
Thanks [~amanraj2520] for the PR!

> Update permissions in orc_merge*.q.out files
> 
>
> Key: HIVE-27089
> URL: https://issues.apache.org/jira/browse/HIVE-27089
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Aman Raj
>Assignee: Aman Raj
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.2.0
>
>  Time Spent: 2h 40m
>  Remaining Estimate: 0h
>
> The orc_merge1,2,3,4,10 .q tests fails with the following errors : 
> < -rw-rw-rw- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> HDFS PATH ###
> ---
> > -rw-r--r-- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> > HDFS PATH ###



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27089) Fix orc_merge*.q test failures

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27089?focusedWorklogId=849794&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849794
 ]

ASF GitHub Bot logged work on HIVE-27089:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 12:13
Start Date: 08/Mar/23 12:13
Worklog Time Spent: 10m 
  Work Description: zabetak closed pull request #4068: HIVE-27089 : Test 
fixes for orc_merge*.q
URL: https://github.com/apache/hive/pull/4068




Issue Time Tracking
---

Worklog Id: (was: 849794)
Time Spent: 2h 40m  (was: 2.5h)

> Fix orc_merge*.q test failures
> --
>
> Key: HIVE-27089
> URL: https://issues.apache.org/jira/browse/HIVE-27089
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Aman Raj
>Assignee: Aman Raj
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2h 40m
>  Remaining Estimate: 0h
>
> The orc_merge1,2,3,4,10 .q tests fails with the following errors : 
> < -rw-rw-rw- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> HDFS PATH ###
> ---
> > -rw-r--r-- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> > HDFS PATH ###



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-27089) Update permissions in orc_merge*.q.out files

2023-03-08 Thread Stamatis Zampetakis (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stamatis Zampetakis updated HIVE-27089:
---
Summary: Update permissions in orc_merge*.q.out files  (was: Fix 
orc_merge*.q test failures)

> Update permissions in orc_merge*.q.out files
> 
>
> Key: HIVE-27089
> URL: https://issues.apache.org/jira/browse/HIVE-27089
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Aman Raj
>Assignee: Aman Raj
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2h 40m
>  Remaining Estimate: 0h
>
> The orc_merge1,2,3,4,10 .q tests fails with the following errors : 
> < -rw-rw-rw- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> HDFS PATH ###
> ---
> > -rw-r--r-- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> > HDFS PATH ###



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27096) Fix test dbtxnmgr_showlocks.q

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27096?focusedWorklogId=849793&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849793
 ]

ASF GitHub Bot logged work on HIVE-27096:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 12:12
Start Date: 08/Mar/23 12:12
Worklog Time Spent: 10m 
  Work Description: zabetak closed pull request #4075: HIVE-27096 : Fix 
test dbtxnmgr_showlocks.q
URL: https://github.com/apache/hive/pull/4075




Issue Time Tracking
---

Worklog Id: (was: 849793)
Time Spent: 2h 20m  (was: 2h 10m)

> Fix test dbtxnmgr_showlocks.q
> -
>
> Key: HIVE-27096
> URL: https://issues.apache.org/jira/browse/HIVE-27096
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Aman Raj
>Assignee: Aman Raj
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> The test fails due to the following Assertion ERROR :
> Client Execution succeeded but contained differences (error code = 1) after 
> executing dbtxnmgr_showlocks.q 
> 73c73
> < 5 default partitioned_acid_table p=abc MINOR initiated --- --- --- --- 
> ---
> > 1 default partitioned_acid_table p=abc MINOR initiated --- --- --- ---
>  
> It expects a Compaction ID 1 but gets 5.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (HIVE-27096) Update dbtxnmgr_showlocks.q.out file

2023-03-08 Thread Stamatis Zampetakis (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27096?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stamatis Zampetakis resolved HIVE-27096.

Fix Version/s: 3.2.0
   Resolution: Fixed

Fixed in 
https://github.com/apache/hive/commit/f1db3f25a7d77f7fb8f4949d94dc2fe43b604bb2. 
Thanks for the PR [~amanraj2520]!

> Update dbtxnmgr_showlocks.q.out file
> 
>
> Key: HIVE-27096
> URL: https://issues.apache.org/jira/browse/HIVE-27096
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Aman Raj
>Assignee: Aman Raj
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.2.0
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> The test fails due to the following Assertion ERROR :
> Client Execution succeeded but contained differences (error code = 1) after 
> executing dbtxnmgr_showlocks.q 
> 73c73
> < 5 default partitioned_acid_table p=abc MINOR initiated --- --- --- --- 
> ---
> > 1 default partitioned_acid_table p=abc MINOR initiated --- --- --- ---
>  
> It expects a Compaction ID 1 but gets 5.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-27096) Update dbtxnmgr_showlocks.q.out file

2023-03-08 Thread Stamatis Zampetakis (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27096?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stamatis Zampetakis updated HIVE-27096:
---
Summary: Update dbtxnmgr_showlocks.q.out file  (was: Fix test 
dbtxnmgr_showlocks.q)

> Update dbtxnmgr_showlocks.q.out file
> 
>
> Key: HIVE-27096
> URL: https://issues.apache.org/jira/browse/HIVE-27096
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Aman Raj
>Assignee: Aman Raj
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> The test fails due to the following Assertion ERROR :
> Client Execution succeeded but contained differences (error code = 1) after 
> executing dbtxnmgr_showlocks.q 
> 73c73
> < 5 default partitioned_acid_table p=abc MINOR initiated --- --- --- --- 
> ---
> > 1 default partitioned_acid_table p=abc MINOR initiated --- --- --- ---
>  
> It expects a Compaction ID 1 but gets 5.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27100) Remove unused data/files from repo

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27100?focusedWorklogId=849789&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849789
 ]

ASF GitHub Bot logged work on HIVE-27100:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 12:01
Start Date: 08/Mar/23 12:01
Worklog Time Spent: 10m 
  Work Description: sonarcloud[bot] commented on PR #4081:
URL: https://github.com/apache/hive/pull/4081#issuecomment-1460052936

   Kudos, SonarCloud Quality Gate passed!    [![Quality Gate 
passed](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/QualityGateBadge/passed-16px.png
 'Quality Gate 
passed')](https://sonarcloud.io/dashboard?id=apache_hive&pullRequest=4081)
   
   
[![Bug](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/bug-16px.png
 
'Bug')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=BUG)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=BUG)
 [0 
Bugs](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=BUG)
  
   
[![Vulnerability](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/vulnerability-16px.png
 
'Vulnerability')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=VULNERABILITY)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=VULNERABILITY)
 [0 
Vulnerabilities](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=VULNERABILITY)
  
   [![Security 
Hotspot](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/security_hotspot-16px.png
 'Security 
Hotspot')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4081&resolved=false&types=SECURITY_HOTSPOT)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4081&resolved=false&types=SECURITY_HOTSPOT)
 [0 Security 
Hotspots](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4081&resolved=false&types=SECURITY_HOTSPOT)
  
   [![Code 
Smell](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/code_smell-16px.png
 'Code 
Smell')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=CODE_SMELL)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=CODE_SMELL)
 [3 Code 
Smells](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4081&resolved=false&types=CODE_SMELL)
   
   [![No Coverage 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/CoverageChart/NoCoverageInfo-16px.png
 'No Coverage 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4081&metric=coverage&view=list)
 No Coverage information  
   [![No Duplication 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/Duplications/NoDuplicationInfo-16px.png
 'No Duplication 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4081&metric=duplicated_lines_density&view=list)
 No Duplication information
   
   




Issue Time Tracking
---

Worklog Id: (was: 849789)
Time Spent: 40m  (was: 0.5h)

> Remove unused data/files from repo
> --
>
> Key: HIVE-27100
> URL: https://issues.apache.org/jira/browse/HIVE-27100
> Project: Hive
>  Issue Type: Task
>Reporter: Stamatis Zampetakis
>Assignee: Stamatis Zampetakis
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Some files under [https://github.com/apache/hive/tree/master/data/files] are 
> not referenced anywhere else in the repo and can be removed.
> Removing them makes it easier to see what is actually tested. Other minor 
> benefits:
>  * faster checkout times;
>  * smaller source/binary releases.
> The script that was used to find which files are not referenced can be found 
> below:
> {code:bash}
> for f in `ls data/files`; do
>   echo -n "$f "; 
>   grep -a -R "$f" --exclude-dir=".git" --exclude-dir=target 
> --exclude=\*.q.out --exclude=\*.class --exclude=\*.jar | wc -l | grep " 0$";
> done
> {code}
> +Output+
> {noformat}
> cbo_t4.txt 

[jira] [Work logged] (HIVE-27089) Fix orc_merge*.q test failures

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27089?focusedWorklogId=849784&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849784
 ]

ASF GitHub Bot logged work on HIVE-27089:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 11:49
Start Date: 08/Mar/23 11:49
Worklog Time Spent: 10m 
  Work Description: zabetak commented on PR #4068:
URL: https://github.com/apache/hive/pull/4068#issuecomment-1460040365

   @yigress I am not sure if it is LLAP related. If you check the 
`orc_merge1.q` file there are three `dfs -ls` operations always following an 
`INSERT OVERWRITE TABLE` query. I assume that the `OVERWRITE` query rewrites 
the directory every time. 
   
   Before HIVE-27016:
   ```
   INSERT OVERWRITE TABLE
   dfs -ls
   -rw-r--r--
   INSERT OVERWRITE TABLE
   dfs -ls
   -rw-rw-rw-
   INSERT OVERWRITE TABLE
   dfs -ls
   -rw-rw-rw-
   ```
   
   After HIVE-27016:
   ```
   INSERT OVERWRITE TABLE
   dfs -ls
   -rw-rw-rw-
   INSERT OVERWRITE TABLE
   dfs -ls
   -rw-rw-rw-
   INSERT OVERWRITE TABLE
   dfs -ls
   -rw-rw-rw-
   ```
   Worth mentioning that even before HIVE-27016 we have directories created 
with `-rw-rw-rw-` just not the first time.
   
   There must be a code-path that operates with different permissions and this 
is definitely a bug. If someone can pinpoint the exact reason of why this 
happens and raise a JIRA that would be great.
   
   However, given that permission differences existed even before HIVE-27016, I 
will not consider HIVE-27016 as a regression so I am OK to simply update the 
.q.out files for the moment.




Issue Time Tracking
---

Worklog Id: (was: 849784)
Time Spent: 2.5h  (was: 2h 20m)

> Fix orc_merge*.q test failures
> --
>
> Key: HIVE-27089
> URL: https://issues.apache.org/jira/browse/HIVE-27089
> Project: Hive
>  Issue Type: Sub-task
>Reporter: Aman Raj
>Assignee: Aman Raj
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2.5h
>  Remaining Estimate: 0h
>
> The orc_merge1,2,3,4,10 .q tests fails with the following errors : 
> < -rw-rw-rw- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> HDFS PATH ###
> ---
> > -rw-r--r-- 3 ### USER ### ### GROUP ### 349 ### HDFS DATE ### hdfs://### 
> > HDFS PATH ###



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27127) Upgrade jackson version to 2.13.5

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27127?focusedWorklogId=849779&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849779
 ]

ASF GitHub Bot logged work on HIVE-27127:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 11:11
Start Date: 08/Mar/23 11:11
Worklog Time Spent: 10m 
  Work Description: sonarcloud[bot] commented on PR #4102:
URL: https://github.com/apache/hive/pull/4102#issuecomment-1460001416

   Kudos, SonarCloud Quality Gate passed!    [![Quality Gate 
passed](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/QualityGateBadge/passed-16px.png
 'Quality Gate 
passed')](https://sonarcloud.io/dashboard?id=apache_hive&pullRequest=4102)
   
   
[![Bug](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/bug-16px.png
 
'Bug')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=BUG)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=BUG)
 [0 
Bugs](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=BUG)
  
   
[![Vulnerability](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/vulnerability-16px.png
 
'Vulnerability')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=VULNERABILITY)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=VULNERABILITY)
 [0 
Vulnerabilities](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=VULNERABILITY)
  
   [![Security 
Hotspot](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/security_hotspot-16px.png
 'Security 
Hotspot')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4102&resolved=false&types=SECURITY_HOTSPOT)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4102&resolved=false&types=SECURITY_HOTSPOT)
 [0 Security 
Hotspots](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4102&resolved=false&types=SECURITY_HOTSPOT)
  
   [![Code 
Smell](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/code_smell-16px.png
 'Code 
Smell')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=CODE_SMELL)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=CODE_SMELL)
 [3 Code 
Smells](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4102&resolved=false&types=CODE_SMELL)
   
   [![No Coverage 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/CoverageChart/NoCoverageInfo-16px.png
 'No Coverage 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4102&metric=coverage&view=list)
 No Coverage information  
   [![No Duplication 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/Duplications/NoDuplicationInfo-16px.png
 'No Duplication 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4102&metric=duplicated_lines_density&view=list)
 No Duplication information
   
   




Issue Time Tracking
---

Worklog Id: (was: 849779)
Time Spent: 20m  (was: 10m)

> Upgrade jackson version to 2.13.5
> -
>
> Key: HIVE-27127
> URL: https://issues.apache.org/jira/browse/HIVE-27127
> Project: Hive
>  Issue Type: Improvement
>Reporter: Sand Shreeya
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (HIVE-27034) Backport HIVE-22648 HIVE-24408 HIVE-26917: Upgrade parquet to 1.12.3 in branch-3

2023-03-08 Thread Sand Shreeya (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27034?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sand Shreeya reassigned HIVE-27034:
---

Assignee: (was: Sand Shreeya)

> Backport HIVE-22648 HIVE-24408 HIVE-26917: Upgrade parquet to 1.12.3 in 
> branch-3
> 
>
> Key: HIVE-27034
> URL: https://issues.apache.org/jira/browse/HIVE-27034
> Project: Hive
>  Issue Type: Task
>Reporter: Sand Shreeya
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-26900) Error message not representing the correct line number with a syntax error in a HQL File

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-26900?focusedWorklogId=849772&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849772
 ]

ASF GitHub Bot logged work on HIVE-26900:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 10:04
Start Date: 08/Mar/23 10:04
Worklog Time Spent: 10m 
  Work Description: shreeyasand closed pull request #4097: HIVE-26900: 
Error message not representing the correct line number wi…
URL: https://github.com/apache/hive/pull/4097




Issue Time Tracking
---

Worklog Id: (was: 849772)
Time Spent: 40m  (was: 0.5h)

> Error message not representing the correct line number with a syntax error in 
> a HQL File
> 
>
> Key: HIVE-26900
> URL: https://issues.apache.org/jira/browse/HIVE-26900
> Project: Hive
>  Issue Type: Bug
>Affects Versions: 3.1.2, 4.0.0-alpha-1, 4.0.0-alpha-2
>Reporter: Vikram Ahuja
>Priority: Minor
>  Labels: pull-request-available
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> When a wrong syntax is added in a HQL file, the error thrown by beeline while 
> running the HQL file is having the wrong line number.  The line number and 
> even the position is incorrect. Seems like parser is not considering spaces 
> and new lines and always throwing the error on line number 1 irrespective of 
> what line the error is on in the HQL file
>  
> For instance, consider the following test.hql file:
>  # --comment
>  # --comment
>  # SET hive.server2.logging.operation.enabled=true;
>  # SET hive.server2.logging.operation.level=VERBOSE;
>  # show tables;
>  #  
>  #  
>  #       CREATE TABLEE DUMMY;
>  
> when we call !run  test.hql in beeline or trigger ./beeline -u 
> jdbc:hive2://localhost:1 -f test.hql, The issue thrown by beeline is
> >>> CREATE TABLEE DUMMY;
> Error: Error while compiling statement: FAILED: ParseException line 1:7 
> cannot recongize input near 'CREATE' 'TABLEE' 'DUMMY' in ddl statement 
> (state=42000,code=4)
> The parser seems to be taking all the lines from 1 and is ignoring spaces in 
> the line.
> The error line in the parse exception is shown as 1:7 but it should have been 
> 8:13.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-27127) Upgrade jackson version to 2.13.5

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27127?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HIVE-27127:
--
Labels: pull-request-available  (was: )

> Upgrade jackson version to 2.13.5
> -
>
> Key: HIVE-27127
> URL: https://issues.apache.org/jira/browse/HIVE-27127
> Project: Hive
>  Issue Type: Improvement
>Reporter: Sand Shreeya
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27127) Upgrade jackson version to 2.13.5

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27127?focusedWorklogId=849771&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849771
 ]

ASF GitHub Bot logged work on HIVE-27127:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 10:03
Start Date: 08/Mar/23 10:03
Worklog Time Spent: 10m 
  Work Description: shreeyasand opened a new pull request, #4102:
URL: https://github.com/apache/hive/pull/4102

   
   
   ### What changes were proposed in this pull request?
   
   Upgrade jackson version to 2.13.5
   
   
   ### Why are the changes needed?
   
   To resolve CVE
   
   ### Does this PR introduce _any_ user-facing change?
   
   No
   
   ### How was this patch tested?
   
   Existing tests




Issue Time Tracking
---

Worklog Id: (was: 849771)
Remaining Estimate: 0h
Time Spent: 10m

> Upgrade jackson version to 2.13.5
> -
>
> Key: HIVE-27127
> URL: https://issues.apache.org/jira/browse/HIVE-27127
> Project: Hive
>  Issue Type: Improvement
>Reporter: Sand Shreeya
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (HIVE-27122) Use Caffeine for caching metadata objects in Compactor threads

2023-03-08 Thread Sourabh Badhya (Jira)


[ 
https://issues.apache.org/jira/browse/HIVE-27122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17697789#comment-17697789
 ] 

Sourabh Badhya commented on HIVE-27122:
---

Thanks [~akshatm] , [~kokila19] , [~rkirtir] , [~veghlaci05] for the reviews.

> Use Caffeine for caching metadata objects in Compactor threads
> --
>
> Key: HIVE-27122
> URL: https://issues.apache.org/jira/browse/HIVE-27122
> Project: Hive
>  Issue Type: Improvement
>Reporter: Sourabh Badhya
>Assignee: Sourabh Badhya
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>  Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> Currently, compactor threads make use of Guava package to cache metadata 
> objects like database/table objects. We should consider using Caffeine 
> package since it provides more control on the cache. It is also observed that 
> cache created from Caffeine package is more performant than cache created 
> from Guava package.
> Some benchmarks comparing Caffeine package vs Guava package - 
> [https://github.com/ben-manes/caffeine/wiki/Benchmarks]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (HIVE-27122) Use Caffeine for caching metadata objects in Compactor threads

2023-03-08 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HIVE-27122?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

László Végh updated HIVE-27122:
---
Fix Version/s: 4.0.0
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

Merged to master, thanks [~sbadhya] for the patch!

> Use Caffeine for caching metadata objects in Compactor threads
> --
>
> Key: HIVE-27122
> URL: https://issues.apache.org/jira/browse/HIVE-27122
> Project: Hive
>  Issue Type: Improvement
>Reporter: Sourabh Badhya
>Assignee: Sourabh Badhya
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>  Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> Currently, compactor threads make use of Guava package to cache metadata 
> objects like database/table objects. We should consider using Caffeine 
> package since it provides more control on the cache. It is also observed that 
> cache created from Caffeine package is more performant than cache created 
> from Guava package.
> Some benchmarks comparing Caffeine package vs Guava package - 
> [https://github.com/ben-manes/caffeine/wiki/Benchmarks]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27122) Use Caffeine for caching metadata objects in Compactor threads

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27122?focusedWorklogId=849761&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849761
 ]

ASF GitHub Bot logged work on HIVE-27122:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 08:57
Start Date: 08/Mar/23 08:57
Worklog Time Spent: 10m 
  Work Description: veghlaci05 merged PR #4098:
URL: https://github.com/apache/hive/pull/4098




Issue Time Tracking
---

Worklog Id: (was: 849761)
Time Spent: 3h 50m  (was: 3h 40m)

> Use Caffeine for caching metadata objects in Compactor threads
> --
>
> Key: HIVE-27122
> URL: https://issues.apache.org/jira/browse/HIVE-27122
> Project: Hive
>  Issue Type: Improvement
>Reporter: Sourabh Badhya
>Assignee: Sourabh Badhya
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> Currently, compactor threads make use of Guava package to cache metadata 
> objects like database/table objects. We should consider using Caffeine 
> package since it provides more control on the cache. It is also observed that 
> cache created from Caffeine package is more performant than cache created 
> from Guava package.
> Some benchmarks comparing Caffeine package vs Guava package - 
> [https://github.com/ben-manes/caffeine/wiki/Benchmarks]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27101) Support incremental materialized view rebuild when Iceberg source tables have insert operation only.

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27101?focusedWorklogId=849760&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849760
 ]

ASF GitHub Bot logged work on HIVE-27101:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 08:47
Start Date: 08/Mar/23 08:47
Worklog Time Spent: 10m 
  Work Description: sonarcloud[bot] commented on PR #4079:
URL: https://github.com/apache/hive/pull/4079#issuecomment-1459761551

   Kudos, SonarCloud Quality Gate passed!    [![Quality Gate 
passed](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/QualityGateBadge/passed-16px.png
 'Quality Gate 
passed')](https://sonarcloud.io/dashboard?id=apache_hive&pullRequest=4079)
   
   
[![Bug](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/bug-16px.png
 
'Bug')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=BUG)
 
[![B](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/B-16px.png
 
'B')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=BUG)
 [5 
Bugs](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=BUG)
  
   
[![Vulnerability](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/vulnerability-16px.png
 
'Vulnerability')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=VULNERABILITY)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=VULNERABILITY)
 [0 
Vulnerabilities](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=VULNERABILITY)
  
   [![Security 
Hotspot](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/security_hotspot-16px.png
 'Security 
Hotspot')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4079&resolved=false&types=SECURITY_HOTSPOT)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4079&resolved=false&types=SECURITY_HOTSPOT)
 [0 Security 
Hotspots](https://sonarcloud.io/project/security_hotspots?id=apache_hive&pullRequest=4079&resolved=false&types=SECURITY_HOTSPOT)
  
   [![Code 
Smell](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/common/code_smell-16px.png
 'Code 
Smell')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=CODE_SMELL)
 
[![A](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/RatingBadge/A-16px.png
 
'A')](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=CODE_SMELL)
 [9 Code 
Smells](https://sonarcloud.io/project/issues?id=apache_hive&pullRequest=4079&resolved=false&types=CODE_SMELL)
   
   [![No Coverage 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/CoverageChart/NoCoverageInfo-16px.png
 'No Coverage 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4079&metric=coverage&view=list)
 No Coverage information  
   [![No Duplication 
information](https://sonarsource.github.io/sonarcloud-github-static-resources/v2/checks/Duplications/NoDuplicationInfo-16px.png
 'No Duplication 
information')](https://sonarcloud.io/component_measures?id=apache_hive&pullRequest=4079&metric=duplicated_lines_density&view=list)
 No Duplication information
   
   




Issue Time Tracking
---

Worklog Id: (was: 849760)
Time Spent: 7h  (was: 6h 50m)

> Support incremental materialized view rebuild when Iceberg source tables have 
> insert operation only.
> 
>
> Key: HIVE-27101
> URL: https://issues.apache.org/jira/browse/HIVE-27101
> Project: Hive
>  Issue Type: Improvement
>  Components: Iceberg integration, Materialized views
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 7h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27117) Fix compaction related flaky tests

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27117?focusedWorklogId=849756&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849756
 ]

ASF GitHub Bot logged work on HIVE-27117:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 08:30
Start Date: 08/Mar/23 08:30
Worklog Time Spent: 10m 
  Work Description: veghlaci05 merged PR #4096:
URL: https://github.com/apache/hive/pull/4096




Issue Time Tracking
---

Worklog Id: (was: 849756)
Time Spent: 2h 40m  (was: 2.5h)

> Fix compaction related flaky tests
> --
>
> Key: HIVE-27117
> URL: https://issues.apache.org/jira/browse/HIVE-27117
> Project: Hive
>  Issue Type: Task
>Reporter: László Végh
>Assignee: László Végh
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2h 40m
>  Remaining Estimate: 0h
>
> The following tests turned out to be flaky recently:
>  * 
> org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testDropTableAndCompactionConcurrent
>  * 
> org.apache.hadoop.hive.ql.txn.compactor.TestCompactionMetrics.testInitiatorFailuresCountedCorrectly
>  * 
> org.apache.hadoop.hive.ql.txn.compactor.TestCrudCompactorOnTez.testMajorCompactionNotPartitionedWithoutBuckets
>  * 
> org.apache.hadoop.hive.ql.txn.compactor.TestCrudCompactorOnTez.testCompactionWithCreateTableProps



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27117) Fix compaction related flaky tests

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27117?focusedWorklogId=849754&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849754
 ]

ASF GitHub Bot logged work on HIVE-27117:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 08:14
Start Date: 08/Mar/23 08:14
Worklog Time Spent: 10m 
  Work Description: abstractdog commented on code in PR #4096:
URL: https://github.com/apache/hive/pull/4096#discussion_r1129093478


##
ql/src/test/org/apache/hadoop/hive/ql/TestTxnCommands2.java:
##
@@ -1434,6 +1434,8 @@ public void 
testTruncatePartitionAndCompactionConcurrent() throws Exception {
 execDDLOpAndCompactionConcurrently("TRUNCATE_PARTITION", true);
   }
   private void execDDLOpAndCompactionConcurrently(String opType, boolean 
isPartioned) throws Exception {
+hiveConf.setBoolVar(HiveConf.ConfVars.HIVE_COMPACTOR_GATHER_STATS, false);

Review Comment:
   please comment on why is this needed and then it can be merged +1





Issue Time Tracking
---

Worklog Id: (was: 849754)
Time Spent: 2.5h  (was: 2h 20m)

> Fix compaction related flaky tests
> --
>
> Key: HIVE-27117
> URL: https://issues.apache.org/jira/browse/HIVE-27117
> Project: Hive
>  Issue Type: Task
>Reporter: László Végh
>Assignee: László Végh
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 2.5h
>  Remaining Estimate: 0h
>
> The following tests turned out to be flaky recently:
>  * 
> org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testDropTableAndCompactionConcurrent
>  * 
> org.apache.hadoop.hive.ql.txn.compactor.TestCompactionMetrics.testInitiatorFailuresCountedCorrectly
>  * 
> org.apache.hadoop.hive.ql.txn.compactor.TestCrudCompactorOnTez.testMajorCompactionNotPartitionedWithoutBuckets
>  * 
> org.apache.hadoop.hive.ql.txn.compactor.TestCrudCompactorOnTez.testCompactionWithCreateTableProps



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Work logged] (HIVE-27101) Support incremental materialized view rebuild when Iceberg source tables have insert operation only.

2023-03-08 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HIVE-27101?focusedWorklogId=849753&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-849753
 ]

ASF GitHub Bot logged work on HIVE-27101:
-

Author: ASF GitHub Bot
Created on: 08/Mar/23 08:00
Start Date: 08/Mar/23 08:00
Worklog Time Spent: 10m 
  Work Description: kasakrisz commented on code in PR #4079:
URL: https://github.com/apache/hive/pull/4079#discussion_r1129081896


##
iceberg/iceberg-handler/src/main/java/org/apache/iceberg/mr/mapreduce/IcebergInputFormat.java:
##
@@ -116,6 +116,13 @@ private static TableScan createTableScan(Table table, 
Configuration conf) {
   scan = scan.useSnapshot(snapshotId);
 }
 
+if (snapshotId == -1) {
+  long snapshotIntervalFrom = 
conf.getLong(InputFormatConfig.SNAPSHOT_INTERVAL_FROM, -1);
+  if (snapshotIntervalFrom != -1) {
+scan = scan.appendsBetween(snapshotIntervalFrom, 
table.currentSnapshot().snapshotId());

Review Comment:
   Changed to use the new api but I also had to customize 
org.apache.iceberg.SerializableTable because it does not support it.





Issue Time Tracking
---

Worklog Id: (was: 849753)
Time Spent: 6h 50m  (was: 6h 40m)

> Support incremental materialized view rebuild when Iceberg source tables have 
> insert operation only.
> 
>
> Key: HIVE-27101
> URL: https://issues.apache.org/jira/browse/HIVE-27101
> Project: Hive
>  Issue Type: Improvement
>  Components: Iceberg integration, Materialized views
>Reporter: Krisztian Kasa
>Assignee: Krisztian Kasa
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 6h 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)