Hi Alessandro,
>From this:

> $ sw hadoop 3.1.0
> $ sw tez 0.10.0 (tried also 0.10.1)


I guess you are using the wrong versions, The Hadoop version to be used
should be 3.3.1[1] and the Tez version should be 0.10.2[2]

The error also seems to be coming from Hadoop code

> vertex=vertex_1666888075798_0001_1_00 [Map 1],
> java.lang.NoSuchMethodError:
> > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I


The compareTo method in Hadoop was changed in HADOOP-16196, which isn't
there in Hadoop-3.1.0, it is there post 3.2.1 [3]

Another stuff, TestBeelinePasswordOptions passes for me inside the source
directory.

[*INFO*] -------------------------------------------------------

[*INFO*]  T E S T S

[*INFO*] -------------------------------------------------------

[*INFO*] Running org.apache.hive.beeline.*TestBeelinePasswordOption*

[*INFO*] *Tests run: 10*, Failures: 0, Errors: 0, Skipped: 0, Time elapsed:
18.264 s - in org.apache.hive.beeline.*TestBeelinePasswordOption*

-Ayush

[1]
https://github.com/apache/hive/blob/release-4.0.0-alpha-2-rc0/pom.xml#L136
[2]
https://github.com/apache/hive/blob/release-4.0.0-alpha-2-rc0/pom.xml#L197
[3] https://issues.apache.org/jira/browse/HADOOP-16196

On Thu, 27 Oct 2022 at 23:15, Alessandro Solimando <
alessandro.solima...@gmail.com> wrote:

> Hi everyone,
>
> unfortunately my vote is -1 (although non-binding) due to a classpath error
> which prevents queries involving Tez to complete (all the details at the
> end of the email, apologies for the lengthy text but I wanted to provide
> all the context).
>
> - verified gpg signature: OK
>
> $ wget https://www.apache.org/dist/hive/KEYS
>
> $ gpg --import KEYS
>
> ...
>
> $ gpg --verify apache-hive-4.0.0-alpha-2-bin.tar.gz.asc
> apache-hive-4.0.0-alpha-2-bin.tar.gz
>
> gpg: Signature made Thu 27 Oct 15:11:48 2022 CEST
>
> gpg:                using RSA key 50606DE1BDBD5CF862A595A907C5682DAFC73125
>
> gpg:                issuer "dkuzme...@apache.org"
>
> gpg: Good signature from "Denys Kuzmenko (CODE SIGNING KEY) <
> dkuzme...@apache.org>" [unknown]
>
> gpg: WARNING: The key's User ID is not certified with a trusted signature!
>
> gpg:          There is no indication that the signature belongs to the
> owner.
>
> Primary key fingerprint: 5060 6DE1 BDBD 5CF8 62A5  95A9 07C5 682D AFC7 3125
>
> $ gpg --verify apache-hive-4.0.0-alpha-2-src.tar.gz.asc
> apache-hive-4.0.0-alpha-2-src.tar.gz
>
> gpg: Signature made Thu 27 Oct 15:12:08 2022 CEST
>
> gpg:                using RSA key 50606DE1BDBD5CF862A595A907C5682DAFC73125
>
> gpg:                issuer "dkuzme...@apache.org"
>
> gpg: Good signature from "Denys Kuzmenko (CODE SIGNING KEY) <
> dkuzme...@apache.org>" [unknown]
>
> gpg: WARNING: The key's User ID is not certified with a trusted signature!
>
> gpg:          There is no indication that the signature belongs to the
> owner.
>
> Primary key fingerprint: 5060 6DE1 BDBD 5CF8 62A5  95A9 07C5 682D AFC7 3125
>
> (AFAIK, this warning is OK)
>
> - verified package checksum: OK
>
> $ diff <(cat apache-hive-4.0.0-alpha-2-src.tar.gz.sha256) <(shasum -a 256
> apache-hive-4.0.0-alpha-2-src.tar.gz)
>
> $ diff <(cat apache-hive-4.0.0-alpha-2-bin.tar.gz.sha256) <(shasum -a 256
> apache-hive-4.0.0-alpha-2-bin.tar.gz)
>
> - verified maven build (no tests): OK
>
> $ mvn clean install -DskipTests
>
> ...
>
> [INFO]
> ------------------------------------------------------------------------
>
> [INFO] BUILD SUCCESS
>
> [INFO]
> ------------------------------------------------------------------------
>
> [INFO] Total time:  04:31 min
>
> - checked release notes: OK
>
> - checked few modules in Nexus: OK
>
> - environment used:
>
> $ sw_vers
>
> ProductName: macOS
>
> ProductVersion: 11.6.8
>
> BuildVersion: 20G730
>
> $ mvn --version
>
> Apache Maven 3.8.1 (05c21c65bdfed0f71a2f2ada8b84da59348c4c5d)
>
> Maven home: .../.sdkman/candidates/maven/current
>
> Java version: 1.8.0_292, vendor: AdoptOpenJDK, runtime:
> .../.sdkman/candidates/java/8.0.292.hs-adpt/jre
>
> Default locale: en_IE, platform encoding: UTF-8
>
> OS name: "mac os x", version: "10.16", arch: "x86_64", family: "mac"
>
> $ java -version
>
> openjdk version "1.8.0_292"
>
> OpenJDK Runtime Environment (AdoptOpenJDK)(build 1.8.0_292-b10)
>
> OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.292-b10, mixed mode)
>
>
> Testing in hive-dev-box (https://github.com/kgyrtkirk/hive-dev-box): KO
>
> This is the setup I have used:
>
> $ sw hadoop 3.1.0
>
> $ sw tez 0.10.0 (tried also 0.10.1)
>
> $ sw hive
>
> https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/apache-hive-4.0.0-alpha-2-bin.tar.gz
>
> In what follows the test data and query I have tried, with associated
> stacktrace for the error. It seems a classpath issue, probably there are
> multiple versions of the class ending up in the CP and the classloader
> happened to load the “wrong one”.
>
> CREATE TABLE test_stats_a (a int, b int) STORED AS ORC;
>
>
> > INSERT INTO test_stats_a (a, b) VALUES (0, 2);
> > INSERT INTO test_stats_a (a, b) VALUES (1, 2);
> > INSERT INTO test_stats_a (a, b) VALUES (2, 2);
> > INSERT INTO test_stats_a (a, b) VALUES (3, 2);
> > INSERT INTO test_stats_a (a, b) VALUES (4, 2);
> > INSERT INTO test_stats_a (a, b) VALUES (5, 2);
> > INSERT INTO test_stats_a (a, b) VALUES (6, 2);
> > INSERT INTO test_stats_a (a, b) VALUES (7, 2);
> > INSERT INTO test_stats_a (a, b) VALUES (8, 3);
> > INSERT INTO test_stats_a (a, b) VALUES (9, 4);
> > INSERT INTO test_stats_a (a, b) VALUES (10, 5);
> > INSERT INTO test_stats_a (a, b) VALUES (11, 6);
> > INSERT INTO test_stats_a (a, b) VALUES (12, 7);
> > INSERT INTO test_stats_a (a, b) VALUES (13, NULL);
> > INSERT INTO test_stats_a (a, b) VALUES (14, NULL);
>
>
> > CREATE TABLE test_stats_b (a int, b int) STORED AS ORC;
>
>
> > INSERT INTO test_stats_b (a, b) VALUES (0, 2);
> > INSERT INTO test_stats_b (a, b) VALUES (1, 2);
> > INSERT INTO test_stats_b (a, b) VALUES (2, 2);
> > INSERT INTO test_stats_b (a, b) VALUES (3, 2);
> > INSERT INTO test_stats_b (a, b) VALUES (4, 2);
> > INSERT INTO test_stats_b (a, b) VALUES (5, 2);
> > INSERT INTO test_stats_b (a, b) VALUES (6, 2);
> > INSERT INTO test_stats_b (a, b) VALUES (7, 2);
> > INSERT INTO test_stats_b (a, b) VALUES (8, 3);
> > INSERT INTO test_stats_b (a, b) VALUES (9, 4);
> > INSERT INTO test_stats_b (a, b) VALUES (10, 5);
> > INSERT INTO test_stats_b (a, b) VALUES (11, 6);
> > INSERT INTO test_stats_b (a, b) VALUES (12, 7);
> > INSERT INTO test_stats_b (a, b) VALUES (13, NULL);
> > INSERT INTO test_stats_b (a, b) VALUES (14, NULL);
>
>
>
> CREATE TABLE test_stats_c (a string, b int) STORED AS PARQUET;
>
>
> > INSERT INTO test_stats_c (a, b) VALUES ("a", 2);
> > INSERT INTO test_stats_c (a, b) VALUES ("b", 2);
> > INSERT INTO test_stats_c (a, b) VALUES ("c", 2);
> > INSERT INTO test_stats_c (a, b) VALUES ("d", 2);
> > INSERT INTO test_stats_c (a, b) VALUES ("e", 2);
> > INSERT INTO test_stats_c (a, b) VALUES ("f", 2);
> > INSERT INTO test_stats_c (a, b) VALUES ("g", 2);
> > INSERT INTO test_stats_c (a, b) VALUES ("h", 2);
> > INSERT INTO test_stats_c (a, b) VALUES ("i", 3);
> > INSERT INTO test_stats_c (a, b) VALUES ("j", 4);
> > INSERT INTO test_stats_c (a, b) VALUES ("k", 5);
> > INSERT INTO test_stats_c (a, b) VALUES ("l", 6);
> > INSERT INTO test_stats_c (a, b) VALUES ("m", 7);
> > INSERT INTO test_stats_c (a, b) VALUES ("n", NULL);
> > INSERT INTO test_stats_c (a, b) VALUES ("o", NULL);
>
>
> SELECT * FROM test_stats_a t1 JOIN test_stats_b t2 ON (t1.a = t2.a) WHERE
> > t1.b < 3 AND t2.b > 1;
>
>
> INFO  : Completed compiling
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b);
> > Time taken:
> > 4.171 seconds
> > INFO  : Operation QUERY obtained 0 locks
> > INFO  : Executing
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b):
> > SELECT * FROM test_sta
> > ts_a t1 JOIN test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b > 1
> > INFO  : Query ID =
> dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b
> > INFO  : Total jobs = 1
> > INFO  : Launching Job 1 out of 1
> > INFO  : Starting task [Stage-1:MAPRED] in serial mode
> > DEBUG : Task getting executed using mapred tag :
> > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b,userid=de
> > v
> > INFO  : Subscribed to counters: [] for queryId:
> > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b
> > INFO  : Tez session hasn't been created yet. Opening session
> > DEBUG : No local resources to process (other than hive-exec)
> > INFO  : Dag name: SELECT * FROM test_st...... < 3 AND t2.b > 1 (Stage-1)
> > DEBUG : DagInfo: {"context":"Hive","description":"SELECT * FROM
> > test_stats_a t1 JOIN test_stats_b t2 ON (t1.a = t2
> > .a) WHERE t1.b < 3 AND t2.b > 1"}
> > DEBUG : Setting Tez DAG access for
> > queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b with
> > viewAclStr
> > ing=dev, modifyStr=dev
> > INFO  : Setting tez.task.scale.memory.reserve-fraction to
> > 0.30000001192092896
> > INFO  : HS2 Host: [alpha2], Query ID:
> > [dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b], Dag ID:
> > [dag_1666888075798_0001_1], DAG Session ID:
> [application_1666888075798_0001]
> > INFO  : Status: Running (Executing on YARN cluster with App id
> > application_1666888075798_0001)
>
>
> > ERROR : Status: Failed
> > ERROR : Vertex failed, vertexName=Map 2,
> > vertexId=vertex_1666888075798_0001_1_01, diagnostics=[Vertex
> > vertex_1666888075798_0001_1_01 [Map 2] killed/failed due
> > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed,
> > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> java.lang.NoSuchMethodError:
> > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
>
>         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> >         at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> >         at java.util.TimSort.sort(TimSort.java:220)
> >         at java.util.Arrays.sort(Arrays.java:1438)
>
>         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
>
>         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:422)
> >         at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
>
>         at
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>
>         at java.lang.Thread.run(Thread.java:748)
> > ]
> > ERROR : Vertex failed, vertexName=Map 1,
> > vertexId=vertex_1666888075798_0001_1_00, diagnostics=[Vertex
> > vertex_1666888075798_0001_1_00 [Map 1] killed/failed due
> > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed,
> > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> java.lang.NoSuchMethodError:
> > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
>
>         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> >         at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> >         at java.util.TimSort.sort(TimSort.java:220)
> >         at java.util.Arrays.sort(Arrays.java:1438)
>
>         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
>
>         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:422)
> >         at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> >         at
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>
>         at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> >         at java.lang.Thread.run(Thread.java:748)
> > ]
> > ERROR : DAG did not succeed due to VERTEX_FAILURE. failedVertices:2
> > killedVertices:0
> > ERROR : FAILED: Execution Error, return code 2 from
> > org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed, vertexName=Map
> > 2, vertexId=vertex_1666888075798_0001_1_01, diagnostics=[Vertex
> > vertex_1666888075798_0001_1_01 [Map 2] killed/failed due
> > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed,
> > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> java.lang.NoSuchMethodError:
> > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
>
>         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> >         at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> >         at java.util.TimSort.sort(TimSort.java:220)
> >         at java.util.Arrays.sort(Arrays.java:1438)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
>
>         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> >         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:422)
> >         at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> >         at
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
>
>         at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> >         at java.lang.Thread.run(Thread.java:748)
> > ]Vertex failed, vertexName=Map 1,
> vertexId=vertex_1666888075798_0001_1_00,
> > diagnostics=[Vertex vertex_1666888075798_0001_1_00 [Map 1] killed/failed
> > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed,
> > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> java.lang.NoSuchMethodError:
> > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> >         at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> >         at java.util.TimSort.sort(TimSort.java:220)
> >         at java.util.Arrays.sort(Arrays.java:1438)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> >         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:422)
> >         at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> >         at
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> >         at java.lang.Thread.run(Thread.java:748)
> > ]DAG did not succeed due to VERTEX_FAILURE. failedVertices:2
> > killedVertices:0
> > DEBUG : Shutting down query SELECT * FROM test_stats_a t1 JOIN
> > test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b > 1
> > INFO  : Completed executing
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b);
> > Time taken: 6.983 seconds
> > DEBUG : Shutting down query SELECT * FROM test_stats_a t1 JOIN
> > test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b > 1
> > Error: Error while compiling statement: FAILED: Execution Error, return
> > code 2 from org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed,
> > vertexName=Map 2, vertexId=vertex_1666888075798_0001_1_01,
> > diagnostics=[Vertex vertex_1666888075798_0001_1_01 [Map 2] killed/failed
> > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed,
> > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> java.lang.NoSuchMethodError:
> > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> >         at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> >         at java.util.TimSort.sort(TimSort.java:220)
> >         at java.util.Arrays.sort(Arrays.java:1438)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> >         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:422)
> >         at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> >         at
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> >         at java.lang.Thread.run(Thread.java:748)
> > ]Vertex failed, vertexName=Map 1,
> vertexId=vertex_1666888075798_0001_1_00,
> > diagnostics=[Vertex vertex_1666888075798_0001_1_00 [Map 1] killed/failed
> > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed,
> > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> java.lang.NoSuchMethodError:
> > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> >         at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> >         at java.util.TimSort.sort(TimSort.java:220)
> >         at java.util.Arrays.sort(Arrays.java:1438)
> >         at
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> >         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:422)
> >         at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> >         at
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> >         at
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> >         at
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> >         at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> >         at java.lang.Thread.run(Thread.java:748)
> > ]DAG did not succeed due to VERTEX_FAILURE. failedVertices:2
> > killedVertices:0 (state=08S01,code=2)
>
>         at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> >
>
> Best regards,
> Alessandro
>
> On Thu, 27 Oct 2022 at 19:01, Ayush Saxena <ayush...@gmail.com> wrote:
>
> > Chris,
> > The KEYS file is at:
> > https://downloads.apache.org/hive/KEYS
> >
> > -Ayush
> >
> > On Thu, 27 Oct 2022 at 21:58, Chris Nauroth <cnaur...@apache.org> wrote:
> >
> > > Could someone please point me toward the right KEYS file to import so
> > that
> > > I can verify signatures? Thanks!
> > >
> > > I'm seeing numerous test failures due to "Insufficient configured
> > threads"
> > > while trying to start the HTTP server. One example is
> > > TestBeelinePasswordOption. Is anyone else seeing this? I noticed that
> > > HIVE-24484 set hive.server2.webui.max.threads to 4 in
> > > /data/conf/hive-site.xml. (The default in HiveConf.java is 50.)
> > >
> > > [INFO] Running org.apache.hive.beeline.TestBeelinePasswordOption
> > > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed:
> > > 11.742 s <<< FAILURE! - in
> > > org.apache.hive.beeline.TestBeelinePasswordOption
> > > [ERROR] org.apache.hive.beeline.TestBeelinePasswordOption  Time
> elapsed:
> > > 11.733 s  <<< ERROR!
> > > org.apache.hive.service.ServiceException:
> > java.lang.IllegalStateException:
> > > Insufficient configured threads: required=4 < max=4 for
> > >
> > >
> >
> QueuedThreadPool[hiveserver2-web]@628bd77e{STARTED,4<=4<=4,i=4,r=-1,q=0}[ReservedThreadExecutor@cfacf0
> > > {s=0/1,p=0}]
> > > at
> org.apache.hive.service.server.HiveServer2.start(HiveServer2.java:733)
> > > at org.apache.hive.jdbc.miniHS2.MiniHS2.start(MiniHS2.java:395)
> > > at
> > >
> > >
> >
> org.apache.hive.beeline.TestBeelinePasswordOption.preTests(TestBeelinePasswordOption.java:60)
> > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > > at
> > >
> > >
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> > > at
> > >
> > >
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> > > at java.lang.reflect.Method.invoke(Method.java:498)
> > > at
> > >
> > >
> >
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> > > at
> > >
> > >
> >
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> > > at
> > >
> > >
> >
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> > > at
> > >
> > >
> >
> org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
> > > at
> > >
> > >
> >
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
> > > at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
> > > at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
> > > at
> > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> > > at
> > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> > > at
> > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> > > at
> > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
> > > at
> > >
> > >
> >
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377)
> > > at
> > >
> > >
> >
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138)
> > > at
> > org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465)
> > > at
> > >
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451)
> > > Caused by: java.lang.IllegalStateException: Insufficient configured
> > > threads: required=4 < max=4 for
> > >
> > >
> >
> QueuedThreadPool[hiveserver2-web]@628bd77e{STARTED,4<=4<=4,i=4,r=-1,q=0}[ReservedThreadExecutor@cfacf0
> > > {s=0/1,p=0}]
> > > at
> > >
> > >
> >
> org.eclipse.jetty.util.thread.ThreadPoolBudget.check(ThreadPoolBudget.java:165)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.util.thread.ThreadPoolBudget.leaseTo(ThreadPoolBudget.java:141)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.util.thread.ThreadPoolBudget.leaseFrom(ThreadPoolBudget.java:191)
> > > at org.eclipse.jetty.io
> > .SelectorManager.doStart(SelectorManager.java:255)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.util.component.ContainerLifeCycle.start(ContainerLifeCycle.java:169)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.util.component.ContainerLifeCycle.doStart(ContainerLifeCycle.java:110)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.server.AbstractConnector.doStart(AbstractConnector.java:321)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.server.AbstractNetworkConnector.doStart(AbstractNetworkConnector.java:81)
> > > at
> > >
> >
> org.eclipse.jetty.server.ServerConnector.doStart(ServerConnector.java:234)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73)
> > > at org.eclipse.jetty.server.Server.doStart(Server.java:401)
> > > at
> > >
> > >
> >
> org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73)
> > > at org.apache.hive.http.HttpServer.start(HttpServer.java:335)
> > > at
> org.apache.hive.service.server.HiveServer2.start(HiveServer2.java:729)
> > > ... 21 more
> > >
> > > Chris Nauroth
> > >
> > >
> > > On Thu, Oct 27, 2022 at 7:48 AM Ádám Szita <sz...@apache.org> wrote:
> > >
> > > > Hi,
> > > >
> > > > Thanks for rebuilding this RC, Denys.
> > > >
> > > > Alessandro: IMHO since there was no vote cast yet and we're talking
> > about
> > > > a build option change only, I guess it just doesn't worth rebuilding
> > the
> > > > whole stuff from scratch to create a new RC.
> > > >
> > > > I give +1 (binding) to this RC, I verified the checksum, binary
> > content,
> > > > source content, built Hive from source and also tried out the
> artifacts
> > > in
> > > > a mini cluster environment. Built an HMS DB with the schema scripts
> > > > provided, did table creation, insert, delete, rollback (Iceberg).
> > > >
> > > > Thanks again, Denys for taking this up.
> > > >
> > > > On 2022/10/27 13:29:36 Alessandro Solimando wrote:
> > > > > Hi Denys,
> > > > > in other Apache communities I generally see that votes are
> cancelled
> > > and
> > > > a
> > > > > new RC is prepared when there are changes or blocking issues like
> in
> > > this
> > > > > case, not sure how things are done in Hive though.
> > > > >
> > > > > Best regards,
> > > > > Alessandro
> > > > >
> > > > > On Thu, 27 Oct 2022 at 15:22, Denys Kuzmenko <
> dkuzme...@cloudera.com
> > > > .invalid>
> > > > > wrote:
> > > > >
> > > > > > Hi Adam,
> > > > > >
> > > > > > Thanks for pointing that out! Upstream release guide is outdated.
> > > Once
> > > > I
> > > > > > receive the edit rights, I'll amend the instructions.
> > > > > > Updated the release artifacts and checksums:
> > > > > >
> > > > > > Apache Hive 4.0.0-alpha-2 Release Candidate 0 is available
> > > > > > here:
> > https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/
> > > > > >
> > > > > >
> > > > > > The checksums are these:
> > > > > > -
> b4dbaac5530694f631af13677ffe5443addc148bd94176b27a109a6da67f5e0f
> > > > > > apache-hive-4.0.0-alpha-2-bin.tar.gz
> > > > > > -
> 8c4639915e9bf649f4a55cd9adb9d266aa15d8fa48ddfadb28ebead2c0aee4d0
> > > > > > apache-hive-4.0.0-alpha-2-src.tar.gz
> > > > > >
> > > > > > Maven artifacts are available
> > > > > > here:
> > > > > >
> > > https://repository.apache.org/content/repositories/orgapachehive-1117/
> > > > > >
> > > > > > The tag release-4.0.0-alpha-2-rc0 has been applied to the source
> > for
> > > > > > this release in github, you can see it at
> > > > > > https://github.com/apache/hive/tree/release-4.0.0-alpha-2-rc0
> > > > > >
> > > > > > The git commit hash
> > > > > > is:
> > > > > >
> > > >
> > >
> >
> https://github.com/apache/hive/commit/da146200e003712e324496bf560a1702485d231c
> > > > > >
> > > > > >
> > > > > > Please check again.
> > > > > >
> > > > > >
> > > > > > Thanks,
> > > > > > Denys
> > > > > >
> > > > > > On Thu, Oct 27, 2022 at 2:53 PM Ádám Szita <sz...@apache.org>
> > wrote:
> > > > > >
> > > > > > > Hi Denys,
> > > > > > >
> > > > > > > Unfortutantely I can't give a plus 1 on this yet, as the
> Iceberg
> > > > > > artifacts
> > > > > > > are missing from the binary tar.gz. Perhaps -Piceberg flag was
> > > > missing
> > > > > > > during build, can you please rebuild?
> > > > > > >
> > > > > > > Thanks,
> > > > > > > Adam
> > > > > > >
> > > > > > > On 2022/10/25 11:20:23 Denys Kuzmenko wrote:
> > > > > > > > Hi team,
> > > > > > > >
> > > > > > > >
> > > > > > > > Apache Hive 4.0.0-alpha-2 Release Candidate 0 is available
> > > > > > > > here:
> > > > https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/
> > > > > > > >
> > > > > > > >
> > > > > > > > The checksums are these:
> > > > > > > > -
> > > 7d4c54ecfe2b04cabc283a84defcc1e8a02eed0e13baba2a2c91ae882b6bfaf7
> > > > > > > > apache-hive-4.0.0-alpha-2-bin.tar.gz
> > > > > > > > -
> > > 8c4639915e9bf649f4a55cd9adb9d266aa15d8fa48ddfadb28ebead2c0aee4d0
> > > > > > > > apache-hive-4.0.0-alpha-2-src.tar.gz
> > > > > > > >
> > > > > > > > Maven artifacts are available
> > > > > > > > here:
> > > > > > >
> > > >
> https://repository.apache.org/content/repositories/orgapachehive-1117/
> > > > > > > >
> > > > > > > > The tag release-4.0.0-alpha-2-rc0 has been applied to the
> > source
> > > > for
> > > > > > > > this release in github, you can see it at
> > > > > > > >
> https://github.com/apache/hive/tree/release-4.0.0-alpha-2-rc0
> > > > > > > >
> > > > > > > > The git commit hash
> > > > > > > > is:
> > > > > > >
> > > > > >
> > > >
> > >
> >
> https://github.com/apache/hive/commit/da146200e003712e324496bf560a1702485d231c
> > > > > > > >
> > > > > > > > Voting will conclude in 72 hours.
> > > > > > > >
> > > > > > > > Hive PMC Members: Please test and vote.
> > > > > > > >
> > > > > > > > Thanks
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
>

Reply via email to