+1 (non-binding)

* Verified all checksums.
* Verified all signatures.
* Built from source.
    * mvn clean install -Piceberg -DskipTests
* Tests passed.
    * mvn --fail-never clean verify -Piceberg -Pitests
-Dmaven.test.jvm.args='-Xmx2048m -DJETTY_AVAILABLE_PROCESSORS=4'

I figured out why my test runs were failing in HTTP server initialization.
Jetty enforces thread leasing to warn or abort if there aren't enough
threads available [1]. During startup, it attempts to lease a thread per
NIO selector [2]. By default, the number of NIO selectors to use is
determined based on available CPUs [3]. This is mostly a passthrough to
Runtime.availableProcessors() [4]. In my case, running on a machine with 16
CPUs, this ended up creating more than 4 selectors, therefore requiring
more than 4 threads and violating the lease check. I was able to work
around this by passing the JETTY_AVAILABLE_PROCESSORS system property to
constrain the number of CPUs available to Jetty.

If we are intentionally constraining the pool to 4 threads during itests,
then would it also make sense to limit JETTY_AVAILABLE_PROCESSORS in
maven.test.jvm.args of the root pom.xml, so that others don't run into this
problem later? If so, I'll send a pull request.

[1]
https://github.com/eclipse/jetty.project/blob/jetty-9.4.40.v20210413/jetty-util/src/main/java/org/eclipse/jetty/util/thread/ThreadPoolBudget.java#L165
[2]
https://github.com/eclipse/jetty.project/blob/jetty-9.4.40.v20210413/jetty-io/src/main/java/org/eclipse/jetty/io/SelectorManager.java#L255
[3]
https://github.com/eclipse/jetty.project/blob/jetty-9.4.40.v20210413/jetty-io/src/main/java/org/eclipse/jetty/io/SelectorManager.java#L79
[4]
https://github.com/eclipse/jetty.project/blob/jetty-9.4.40.v20210413/jetty-util/src/main/java/org/eclipse/jetty/util/ProcessorUtils.java#L45

Chris Nauroth


On Thu, Oct 27, 2022 at 1:18 PM Alessandro Solimando <
alessandro.solima...@gmail.com> wrote:

> You are right Ayush, I got sidetracked by the release notes (* [HIVE-19217]
> - Upgrade to Hadoop 3.1.0) and I did not check the versions in the pom
> file, apologies for the false alarm but better safe than sorry.
>
> With the right versions in place (Hadoop 3.3.1 and Tez 10.0.2), tests
> including select, join, groupby, orderby, explain (ast, cbo, cbo cost,
> vectorization) are working correctly, against data in ORC and parquet
> file format.
>
> No problem for me either when running TestBeelinePasswordOption locally.
>
> So my vote turns into a +1 (non-binding).
>
> Thanks a lot Denys for pushing the release process forward, sorry again you
> all for the oversight!
>
> Best regards,
> Alessandro
>
> On Thu, 27 Oct 2022 at 20:03, Ayush Saxena <ayush...@gmail.com> wrote:
>
> > Hi Alessandro,
> > From this:
> >
> > > $ sw hadoop 3.1.0
> > > $ sw tez 0.10.0 (tried also 0.10.1)
> >
> >
> > I guess you are using the wrong versions, The Hadoop version to be used
> > should be 3.3.1[1] and the Tez version should be 0.10.2[2]
> >
> > The error also seems to be coming from Hadoop code
> >
> > > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> > > java.lang.NoSuchMethodError:
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> >
> >
> > The compareTo method in Hadoop was changed in HADOOP-16196, which isn't
> > there in Hadoop-3.1.0, it is there post 3.2.1 [3]
> >
> > Another stuff, TestBeelinePasswordOptions passes for me inside the source
> > directory.
> >
> > [*INFO*] -------------------------------------------------------
> >
> > [*INFO*]  T E S T S
> >
> > [*INFO*] -------------------------------------------------------
> >
> > [*INFO*] Running org.apache.hive.beeline.*TestBeelinePasswordOption*
> >
> > [*INFO*] *Tests run: 10*, Failures: 0, Errors: 0, Skipped: 0, Time
> elapsed:
> > 18.264 s - in org.apache.hive.beeline.*TestBeelinePasswordOption*
> >
> > -Ayush
> >
> > [1]
> >
> https://github.com/apache/hive/blob/release-4.0.0-alpha-2-rc0/pom.xml#L136
> > [2]
> >
> https://github.com/apache/hive/blob/release-4.0.0-alpha-2-rc0/pom.xml#L197
> > [3] https://issues.apache.org/jira/browse/HADOOP-16196
> >
> > On Thu, 27 Oct 2022 at 23:15, Alessandro Solimando <
> > alessandro.solima...@gmail.com> wrote:
> >
> > > Hi everyone,
> > >
> > > unfortunately my vote is -1 (although non-binding) due to a classpath
> > error
> > > which prevents queries involving Tez to complete (all the details at
> the
> > > end of the email, apologies for the lengthy text but I wanted to
> provide
> > > all the context).
> > >
> > > - verified gpg signature: OK
> > >
> > > $ wget https://www.apache.org/dist/hive/KEYS
> > >
> > > $ gpg --import KEYS
> > >
> > > ...
> > >
> > > $ gpg --verify apache-hive-4.0.0-alpha-2-bin.tar.gz.asc
> > > apache-hive-4.0.0-alpha-2-bin.tar.gz
> > >
> > > gpg: Signature made Thu 27 Oct 15:11:48 2022 CEST
> > >
> > > gpg:                using RSA key
> > 50606DE1BDBD5CF862A595A907C5682DAFC73125
> > >
> > > gpg:                issuer "dkuzme...@apache.org"
> > >
> > > gpg: Good signature from "Denys Kuzmenko (CODE SIGNING KEY) <
> > > dkuzme...@apache.org>" [unknown]
> > >
> > > gpg: WARNING: The key's User ID is not certified with a trusted
> > signature!
> > >
> > > gpg:          There is no indication that the signature belongs to the
> > > owner.
> > >
> > > Primary key fingerprint: 5060 6DE1 BDBD 5CF8 62A5  95A9 07C5 682D AFC7
> > 3125
> > >
> > > $ gpg --verify apache-hive-4.0.0-alpha-2-src.tar.gz.asc
> > > apache-hive-4.0.0-alpha-2-src.tar.gz
> > >
> > > gpg: Signature made Thu 27 Oct 15:12:08 2022 CEST
> > >
> > > gpg:                using RSA key
> > 50606DE1BDBD5CF862A595A907C5682DAFC73125
> > >
> > > gpg:                issuer "dkuzme...@apache.org"
> > >
> > > gpg: Good signature from "Denys Kuzmenko (CODE SIGNING KEY) <
> > > dkuzme...@apache.org>" [unknown]
> > >
> > > gpg: WARNING: The key's User ID is not certified with a trusted
> > signature!
> > >
> > > gpg:          There is no indication that the signature belongs to the
> > > owner.
> > >
> > > Primary key fingerprint: 5060 6DE1 BDBD 5CF8 62A5  95A9 07C5 682D AFC7
> > 3125
> > >
> > > (AFAIK, this warning is OK)
> > >
> > > - verified package checksum: OK
> > >
> > > $ diff <(cat apache-hive-4.0.0-alpha-2-src.tar.gz.sha256) <(shasum -a
> 256
> > > apache-hive-4.0.0-alpha-2-src.tar.gz)
> > >
> > > $ diff <(cat apache-hive-4.0.0-alpha-2-bin.tar.gz.sha256) <(shasum -a
> 256
> > > apache-hive-4.0.0-alpha-2-bin.tar.gz)
> > >
> > > - verified maven build (no tests): OK
> > >
> > > $ mvn clean install -DskipTests
> > >
> > > ...
> > >
> > > [INFO]
> > >
> ------------------------------------------------------------------------
> > >
> > > [INFO] BUILD SUCCESS
> > >
> > > [INFO]
> > >
> ------------------------------------------------------------------------
> > >
> > > [INFO] Total time:  04:31 min
> > >
> > > - checked release notes: OK
> > >
> > > - checked few modules in Nexus: OK
> > >
> > > - environment used:
> > >
> > > $ sw_vers
> > >
> > > ProductName: macOS
> > >
> > > ProductVersion: 11.6.8
> > >
> > > BuildVersion: 20G730
> > >
> > > $ mvn --version
> > >
> > > Apache Maven 3.8.1 (05c21c65bdfed0f71a2f2ada8b84da59348c4c5d)
> > >
> > > Maven home: .../.sdkman/candidates/maven/current
> > >
> > > Java version: 1.8.0_292, vendor: AdoptOpenJDK, runtime:
> > > .../.sdkman/candidates/java/8.0.292.hs-adpt/jre
> > >
> > > Default locale: en_IE, platform encoding: UTF-8
> > >
> > > OS name: "mac os x", version: "10.16", arch: "x86_64", family: "mac"
> > >
> > > $ java -version
> > >
> > > openjdk version "1.8.0_292"
> > >
> > > OpenJDK Runtime Environment (AdoptOpenJDK)(build 1.8.0_292-b10)
> > >
> > > OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.292-b10, mixed mode)
> > >
> > >
> > > Testing in hive-dev-box (https://github.com/kgyrtkirk/hive-dev-box):
> KO
> > >
> > > This is the setup I have used:
> > >
> > > $ sw hadoop 3.1.0
> > >
> > > $ sw tez 0.10.0 (tried also 0.10.1)
> > >
> > > $ sw hive
> > >
> > >
> >
> https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/apache-hive-4.0.0-alpha-2-bin.tar.gz
> > >
> > > In what follows the test data and query I have tried, with associated
> > > stacktrace for the error. It seems a classpath issue, probably there
> are
> > > multiple versions of the class ending up in the CP and the classloader
> > > happened to load the “wrong one”.
> > >
> > > CREATE TABLE test_stats_a (a int, b int) STORED AS ORC;
> > >
> > >
> > > > INSERT INTO test_stats_a (a, b) VALUES (0, 2);
> > > > INSERT INTO test_stats_a (a, b) VALUES (1, 2);
> > > > INSERT INTO test_stats_a (a, b) VALUES (2, 2);
> > > > INSERT INTO test_stats_a (a, b) VALUES (3, 2);
> > > > INSERT INTO test_stats_a (a, b) VALUES (4, 2);
> > > > INSERT INTO test_stats_a (a, b) VALUES (5, 2);
> > > > INSERT INTO test_stats_a (a, b) VALUES (6, 2);
> > > > INSERT INTO test_stats_a (a, b) VALUES (7, 2);
> > > > INSERT INTO test_stats_a (a, b) VALUES (8, 3);
> > > > INSERT INTO test_stats_a (a, b) VALUES (9, 4);
> > > > INSERT INTO test_stats_a (a, b) VALUES (10, 5);
> > > > INSERT INTO test_stats_a (a, b) VALUES (11, 6);
> > > > INSERT INTO test_stats_a (a, b) VALUES (12, 7);
> > > > INSERT INTO test_stats_a (a, b) VALUES (13, NULL);
> > > > INSERT INTO test_stats_a (a, b) VALUES (14, NULL);
> > >
> > >
> > > > CREATE TABLE test_stats_b (a int, b int) STORED AS ORC;
> > >
> > >
> > > > INSERT INTO test_stats_b (a, b) VALUES (0, 2);
> > > > INSERT INTO test_stats_b (a, b) VALUES (1, 2);
> > > > INSERT INTO test_stats_b (a, b) VALUES (2, 2);
> > > > INSERT INTO test_stats_b (a, b) VALUES (3, 2);
> > > > INSERT INTO test_stats_b (a, b) VALUES (4, 2);
> > > > INSERT INTO test_stats_b (a, b) VALUES (5, 2);
> > > > INSERT INTO test_stats_b (a, b) VALUES (6, 2);
> > > > INSERT INTO test_stats_b (a, b) VALUES (7, 2);
> > > > INSERT INTO test_stats_b (a, b) VALUES (8, 3);
> > > > INSERT INTO test_stats_b (a, b) VALUES (9, 4);
> > > > INSERT INTO test_stats_b (a, b) VALUES (10, 5);
> > > > INSERT INTO test_stats_b (a, b) VALUES (11, 6);
> > > > INSERT INTO test_stats_b (a, b) VALUES (12, 7);
> > > > INSERT INTO test_stats_b (a, b) VALUES (13, NULL);
> > > > INSERT INTO test_stats_b (a, b) VALUES (14, NULL);
> > >
> > >
> > >
> > > CREATE TABLE test_stats_c (a string, b int) STORED AS PARQUET;
> > >
> > >
> > > > INSERT INTO test_stats_c (a, b) VALUES ("a", 2);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("b", 2);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("c", 2);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("d", 2);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("e", 2);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("f", 2);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("g", 2);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("h", 2);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("i", 3);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("j", 4);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("k", 5);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("l", 6);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("m", 7);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("n", NULL);
> > > > INSERT INTO test_stats_c (a, b) VALUES ("o", NULL);
> > >
> > >
> > > SELECT * FROM test_stats_a t1 JOIN test_stats_b t2 ON (t1.a = t2.a)
> WHERE
> > > > t1.b < 3 AND t2.b > 1;
> > >
> > >
> > > INFO  : Completed compiling
> > > >
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b);
> > > > Time taken:
> > > > 4.171 seconds
> > > > INFO  : Operation QUERY obtained 0 locks
> > > > INFO  : Executing
> > > >
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b):
> > > > SELECT * FROM test_sta
> > > > ts_a t1 JOIN test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b
> > > 1
> > > > INFO  : Query ID =
> > > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b
> > > > INFO  : Total jobs = 1
> > > > INFO  : Launching Job 1 out of 1
> > > > INFO  : Starting task [Stage-1:MAPRED] in serial mode
> > > > DEBUG : Task getting executed using mapred tag :
> > > > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b,userid=de
> > > > v
> > > > INFO  : Subscribed to counters: [] for queryId:
> > > > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b
> > > > INFO  : Tez session hasn't been created yet. Opening session
> > > > DEBUG : No local resources to process (other than hive-exec)
> > > > INFO  : Dag name: SELECT * FROM test_st...... < 3 AND t2.b > 1
> > (Stage-1)
> > > > DEBUG : DagInfo: {"context":"Hive","description":"SELECT * FROM
> > > > test_stats_a t1 JOIN test_stats_b t2 ON (t1.a = t2
> > > > .a) WHERE t1.b < 3 AND t2.b > 1"}
> > > > DEBUG : Setting Tez DAG access for
> > > > queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b with
> > > > viewAclStr
> > > > ing=dev, modifyStr=dev
> > > > INFO  : Setting tez.task.scale.memory.reserve-fraction to
> > > > 0.30000001192092896
> > > > INFO  : HS2 Host: [alpha2], Query ID:
> > > > [dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b], Dag ID:
> > > > [dag_1666888075798_0001_1], DAG Session ID:
> > > [application_1666888075798_0001]
> > > > INFO  : Status: Running (Executing on YARN cluster with App id
> > > > application_1666888075798_0001)
> > >
> > >
> > > > ERROR : Status: Failed
> > > > ERROR : Vertex failed, vertexName=Map 2,
> > > > vertexId=vertex_1666888075798_0001_1_01, diagnostics=[Vertex
> > > > vertex_1666888075798_0001_1_01 [Map 2] killed/failed due
> > > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed,
> > > > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> > > java.lang.NoSuchMethodError:
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > >
> > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > >         at
> java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > >
> > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > >
> > >         at java.security.AccessController.doPrivileged(Native Method)
> > > >         at javax.security.auth.Subject.doAs(Subject.java:422)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > >
> > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >
> > >         at java.lang.Thread.run(Thread.java:748)
> > > > ]
> > > > ERROR : Vertex failed, vertexName=Map 1,
> > > > vertexId=vertex_1666888075798_0001_1_00, diagnostics=[Vertex
> > > > vertex_1666888075798_0001_1_00 [Map 1] killed/failed due
> > > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed,
> > > > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> > > java.lang.NoSuchMethodError:
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > >
> > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > >         at
> java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > >
> > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > >
> > >         at java.security.AccessController.doPrivileged(Native Method)
> > > >         at javax.security.auth.Subject.doAs(Subject.java:422)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > >
> > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > >         at java.lang.Thread.run(Thread.java:748)
> > > > ]
> > > > ERROR : DAG did not succeed due to VERTEX_FAILURE. failedVertices:2
> > > > killedVertices:0
> > > > ERROR : FAILED: Execution Error, return code 2 from
> > > > org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed,
> > vertexName=Map
> > > > 2, vertexId=vertex_1666888075798_0001_1_01, diagnostics=[Vertex
> > > > vertex_1666888075798_0001_1_01 [Map 2] killed/failed due
> > > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed,
> > > > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> > > java.lang.NoSuchMethodError:
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > >
> > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > >         at
> java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > >
> > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > >         at java.security.AccessController.doPrivileged(Native Method)
> > > >         at javax.security.auth.Subject.doAs(Subject.java:422)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > >
> > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > >         at java.lang.Thread.run(Thread.java:748)
> > > > ]Vertex failed, vertexName=Map 1,
> > > vertexId=vertex_1666888075798_0001_1_00,
> > > > diagnostics=[Vertex vertex_1666888075798_0001_1_00 [Map 1]
> > killed/failed
> > > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed,
> > > > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> > > java.lang.NoSuchMethodError:
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > >         at
> java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > >         at java.security.AccessController.doPrivileged(Native Method)
> > > >         at javax.security.auth.Subject.doAs(Subject.java:422)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > >         at java.lang.Thread.run(Thread.java:748)
> > > > ]DAG did not succeed due to VERTEX_FAILURE. failedVertices:2
> > > > killedVertices:0
> > > > DEBUG : Shutting down query SELECT * FROM test_stats_a t1 JOIN
> > > > test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b > 1
> > > > INFO  : Completed executing
> > > >
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b);
> > > > Time taken: 6.983 seconds
> > > > DEBUG : Shutting down query SELECT * FROM test_stats_a t1 JOIN
> > > > test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b > 1
> > > > Error: Error while compiling statement: FAILED: Execution Error,
> return
> > > > code 2 from org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex
> failed,
> > > > vertexName=Map 2, vertexId=vertex_1666888075798_0001_1_01,
> > > > diagnostics=[Vertex vertex_1666888075798_0001_1_01 [Map 2]
> > killed/failed
> > > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed,
> > > > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> > > java.lang.NoSuchMethodError:
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > >         at
> java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > >         at java.security.AccessController.doPrivileged(Native Method)
> > > >         at javax.security.auth.Subject.doAs(Subject.java:422)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > >         at java.lang.Thread.run(Thread.java:748)
> > > > ]Vertex failed, vertexName=Map 1,
> > > vertexId=vertex_1666888075798_0001_1_00,
> > > > diagnostics=[Vertex vertex_1666888075798_0001_1_00 [Map 1]
> > killed/failed
> > > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed,
> > > > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> > > java.lang.NoSuchMethodError:
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > >         at
> java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > >         at java.security.AccessController.doPrivileged(Native Method)
> > > >         at javax.security.auth.Subject.doAs(Subject.java:422)
> > > >         at
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > >         at
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > >         at
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > >         at java.lang.Thread.run(Thread.java:748)
> > > > ]DAG did not succeed due to VERTEX_FAILURE. failedVertices:2
> > > > killedVertices:0 (state=08S01,code=2)
> > >
> > >         at
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > >
> > >
> > > Best regards,
> > > Alessandro
> > >
> > > On Thu, 27 Oct 2022 at 19:01, Ayush Saxena <ayush...@gmail.com> wrote:
> > >
> > > > Chris,
> > > > The KEYS file is at:
> > > > https://downloads.apache.org/hive/KEYS
> > > >
> > > > -Ayush
> > > >
> > > > On Thu, 27 Oct 2022 at 21:58, Chris Nauroth <cnaur...@apache.org>
> > wrote:
> > > >
> > > > > Could someone please point me toward the right KEYS file to import
> so
> > > > that
> > > > > I can verify signatures? Thanks!
> > > > >
> > > > > I'm seeing numerous test failures due to "Insufficient configured
> > > > threads"
> > > > > while trying to start the HTTP server. One example is
> > > > > TestBeelinePasswordOption. Is anyone else seeing this? I noticed
> that
> > > > > HIVE-24484 set hive.server2.webui.max.threads to 4 in
> > > > > /data/conf/hive-site.xml. (The default in HiveConf.java is 50.)
> > > > >
> > > > > [INFO] Running org.apache.hive.beeline.TestBeelinePasswordOption
> > > > > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time
> > elapsed:
> > > > > 11.742 s <<< FAILURE! - in
> > > > > org.apache.hive.beeline.TestBeelinePasswordOption
> > > > > [ERROR] org.apache.hive.beeline.TestBeelinePasswordOption  Time
> > > elapsed:
> > > > > 11.733 s  <<< ERROR!
> > > > > org.apache.hive.service.ServiceException:
> > > > java.lang.IllegalStateException:
> > > > > Insufficient configured threads: required=4 < max=4 for
> > > > >
> > > > >
> > > >
> > >
> >
> QueuedThreadPool[hiveserver2-web]@628bd77e{STARTED,4<=4<=4,i=4,r=-1,q=0}[ReservedThreadExecutor@cfacf0
> > > > > {s=0/1,p=0}]
> > > > > at
> > > org.apache.hive.service.server.HiveServer2.start(HiveServer2.java:733)
> > > > > at org.apache.hive.jdbc.miniHS2.MiniHS2.start(MiniHS2.java:395)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hive.beeline.TestBeelinePasswordOption.preTests(TestBeelinePasswordOption.java:60)
> > > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> > > > > at java.lang.reflect.Method.invoke(Method.java:498)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
> > > > > at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
> > > > > at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138)
> > > > > at
> > > >
> > org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465)
> > > > > at
> > > > >
> > >
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451)
> > > > > Caused by: java.lang.IllegalStateException: Insufficient configured
> > > > > threads: required=4 < max=4 for
> > > > >
> > > > >
> > > >
> > >
> >
> QueuedThreadPool[hiveserver2-web]@628bd77e{STARTED,4<=4<=4,i=4,r=-1,q=0}[ReservedThreadExecutor@cfacf0
> > > > > {s=0/1,p=0}]
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.thread.ThreadPoolBudget.check(ThreadPoolBudget.java:165)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.thread.ThreadPoolBudget.leaseTo(ThreadPoolBudget.java:141)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.thread.ThreadPoolBudget.leaseFrom(ThreadPoolBudget.java:191)
> > > > > at org.eclipse.jetty.io
> > > > .SelectorManager.doStart(SelectorManager.java:255)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.component.ContainerLifeCycle.start(ContainerLifeCycle.java:169)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.component.ContainerLifeCycle.doStart(ContainerLifeCycle.java:110)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.server.AbstractConnector.doStart(AbstractConnector.java:321)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.server.AbstractNetworkConnector.doStart(AbstractNetworkConnector.java:81)
> > > > > at
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.server.ServerConnector.doStart(ServerConnector.java:234)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73)
> > > > > at org.eclipse.jetty.server.Server.doStart(Server.java:401)
> > > > > at
> > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73)
> > > > > at org.apache.hive.http.HttpServer.start(HttpServer.java:335)
> > > > > at
> > > org.apache.hive.service.server.HiveServer2.start(HiveServer2.java:729)
> > > > > ... 21 more
> > > > >
> > > > > Chris Nauroth
> > > > >
> > > > >
> > > > > On Thu, Oct 27, 2022 at 7:48 AM Ádám Szita <sz...@apache.org>
> wrote:
> > > > >
> > > > > > Hi,
> > > > > >
> > > > > > Thanks for rebuilding this RC, Denys.
> > > > > >
> > > > > > Alessandro: IMHO since there was no vote cast yet and we're
> talking
> > > > about
> > > > > > a build option change only, I guess it just doesn't worth
> > rebuilding
> > > > the
> > > > > > whole stuff from scratch to create a new RC.
> > > > > >
> > > > > > I give +1 (binding) to this RC, I verified the checksum, binary
> > > > content,
> > > > > > source content, built Hive from source and also tried out the
> > > artifacts
> > > > > in
> > > > > > a mini cluster environment. Built an HMS DB with the schema
> scripts
> > > > > > provided, did table creation, insert, delete, rollback (Iceberg).
> > > > > >
> > > > > > Thanks again, Denys for taking this up.
> > > > > >
> > > > > > On 2022/10/27 13:29:36 Alessandro Solimando wrote:
> > > > > > > Hi Denys,
> > > > > > > in other Apache communities I generally see that votes are
> > > cancelled
> > > > > and
> > > > > > a
> > > > > > > new RC is prepared when there are changes or blocking issues
> like
> > > in
> > > > > this
> > > > > > > case, not sure how things are done in Hive though.
> > > > > > >
> > > > > > > Best regards,
> > > > > > > Alessandro
> > > > > > >
> > > > > > > On Thu, 27 Oct 2022 at 15:22, Denys Kuzmenko <
> > > dkuzme...@cloudera.com
> > > > > > .invalid>
> > > > > > > wrote:
> > > > > > >
> > > > > > > > Hi Adam,
> > > > > > > >
> > > > > > > > Thanks for pointing that out! Upstream release guide is
> > outdated.
> > > > > Once
> > > > > > I
> > > > > > > > receive the edit rights, I'll amend the instructions.
> > > > > > > > Updated the release artifacts and checksums:
> > > > > > > >
> > > > > > > > Apache Hive 4.0.0-alpha-2 Release Candidate 0 is available
> > > > > > > > here:
> > > > https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/
> > > > > > > >
> > > > > > > >
> > > > > > > > The checksums are these:
> > > > > > > > -
> > > b4dbaac5530694f631af13677ffe5443addc148bd94176b27a109a6da67f5e0f
> > > > > > > > apache-hive-4.0.0-alpha-2-bin.tar.gz
> > > > > > > > -
> > > 8c4639915e9bf649f4a55cd9adb9d266aa15d8fa48ddfadb28ebead2c0aee4d0
> > > > > > > > apache-hive-4.0.0-alpha-2-src.tar.gz
> > > > > > > >
> > > > > > > > Maven artifacts are available
> > > > > > > > here:
> > > > > > > >
> > > > >
> > https://repository.apache.org/content/repositories/orgapachehive-1117/
> > > > > > > >
> > > > > > > > The tag release-4.0.0-alpha-2-rc0 has been applied to the
> > source
> > > > for
> > > > > > > > this release in github, you can see it at
> > > > > > > >
> https://github.com/apache/hive/tree/release-4.0.0-alpha-2-rc0
> > > > > > > >
> > > > > > > > The git commit hash
> > > > > > > > is:
> > > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/apache/hive/commit/da146200e003712e324496bf560a1702485d231c
> > > > > > > >
> > > > > > > >
> > > > > > > > Please check again.
> > > > > > > >
> > > > > > > >
> > > > > > > > Thanks,
> > > > > > > > Denys
> > > > > > > >
> > > > > > > > On Thu, Oct 27, 2022 at 2:53 PM Ádám Szita <sz...@apache.org
> >
> > > > wrote:
> > > > > > > >
> > > > > > > > > Hi Denys,
> > > > > > > > >
> > > > > > > > > Unfortutantely I can't give a plus 1 on this yet, as the
> > > Iceberg
> > > > > > > > artifacts
> > > > > > > > > are missing from the binary tar.gz. Perhaps -Piceberg flag
> > was
> > > > > > missing
> > > > > > > > > during build, can you please rebuild?
> > > > > > > > >
> > > > > > > > > Thanks,
> > > > > > > > > Adam
> > > > > > > > >
> > > > > > > > > On 2022/10/25 11:20:23 Denys Kuzmenko wrote:
> > > > > > > > > > Hi team,
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > Apache Hive 4.0.0-alpha-2 Release Candidate 0 is
> available
> > > > > > > > > > here:
> > > > > > https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > The checksums are these:
> > > > > > > > > > -
> > > > > 7d4c54ecfe2b04cabc283a84defcc1e8a02eed0e13baba2a2c91ae882b6bfaf7
> > > > > > > > > > apache-hive-4.0.0-alpha-2-bin.tar.gz
> > > > > > > > > > -
> > > > > 8c4639915e9bf649f4a55cd9adb9d266aa15d8fa48ddfadb28ebead2c0aee4d0
> > > > > > > > > > apache-hive-4.0.0-alpha-2-src.tar.gz
> > > > > > > > > >
> > > > > > > > > > Maven artifacts are available
> > > > > > > > > > here:
> > > > > > > > >
> > > > > >
> > > https://repository.apache.org/content/repositories/orgapachehive-1117/
> > > > > > > > > >
> > > > > > > > > > The tag release-4.0.0-alpha-2-rc0 has been applied to the
> > > > source
> > > > > > for
> > > > > > > > > > this release in github, you can see it at
> > > > > > > > > >
> > > https://github.com/apache/hive/tree/release-4.0.0-alpha-2-rc0
> > > > > > > > > >
> > > > > > > > > > The git commit hash
> > > > > > > > > > is:
> > > > > > > > >
> > > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/apache/hive/commit/da146200e003712e324496bf560a1702485d231c
> > > > > > > > > >
> > > > > > > > > > Voting will conclude in 72 hours.
> > > > > > > > > >
> > > > > > > > > > Hive PMC Members: Please test and vote.
> > > > > > > > > >
> > > > > > > > > > Thanks
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
>

Reply via email to