Hi Alessandro, >From this: > $ sw hadoop 3.1.0 > $ sw tez 0.10.0 (tried also 0.10.1)
I guess you are using the wrong versions, The Hadoop version to be used should be 3.3.1[1] and the Tez version should be 0.10.2[2] The error also seems to be coming from Hadoop code > vertex=vertex_1666888075798_0001_1_00 [Map 1], > java.lang.NoSuchMethodError: > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I The compareTo method in Hadoop was changed in HADOOP-16196, which isn't there in Hadoop-3.1.0, it is there post 3.2.1 [3] Another stuff, TestBeelinePasswordOptions passes for me inside the source directory. [*INFO*] ------------------------------------------------------- [*INFO*] T E S T S [*INFO*] ------------------------------------------------------- [*INFO*] Running org.apache.hive.beeline.*TestBeelinePasswordOption* [*INFO*] *Tests run: 10*, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.264 s - in org.apache.hive.beeline.*TestBeelinePasswordOption* -Ayush [1] https://github.com/apache/hive/blob/release-4.0.0-alpha-2-rc0/pom.xml#L136 [2] https://github.com/apache/hive/blob/release-4.0.0-alpha-2-rc0/pom.xml#L197 [3] https://issues.apache.org/jira/browse/HADOOP-16196 On Thu, 27 Oct 2022 at 23:15, Alessandro Solimando < alessandro.solima...@gmail.com> wrote: > Hi everyone, > > unfortunately my vote is -1 (although non-binding) due to a classpath error > which prevents queries involving Tez to complete (all the details at the > end of the email, apologies for the lengthy text but I wanted to provide > all the context). > > - verified gpg signature: OK > > $ wget https://www.apache.org/dist/hive/KEYS > > $ gpg --import KEYS > > ... > > $ gpg --verify apache-hive-4.0.0-alpha-2-bin.tar.gz.asc > apache-hive-4.0.0-alpha-2-bin.tar.gz > > gpg: Signature made Thu 27 Oct 15:11:48 2022 CEST > > gpg: using RSA key 50606DE1BDBD5CF862A595A907C5682DAFC73125 > > gpg: issuer "dkuzme...@apache.org" > > gpg: Good signature from "Denys Kuzmenko (CODE SIGNING KEY) < > dkuzme...@apache.org>" [unknown] > > gpg: WARNING: The key's User ID is not certified with a trusted signature! > > gpg: There is no indication that the signature belongs to the > owner. > > Primary key fingerprint: 5060 6DE1 BDBD 5CF8 62A5 95A9 07C5 682D AFC7 3125 > > $ gpg --verify apache-hive-4.0.0-alpha-2-src.tar.gz.asc > apache-hive-4.0.0-alpha-2-src.tar.gz > > gpg: Signature made Thu 27 Oct 15:12:08 2022 CEST > > gpg: using RSA key 50606DE1BDBD5CF862A595A907C5682DAFC73125 > > gpg: issuer "dkuzme...@apache.org" > > gpg: Good signature from "Denys Kuzmenko (CODE SIGNING KEY) < > dkuzme...@apache.org>" [unknown] > > gpg: WARNING: The key's User ID is not certified with a trusted signature! > > gpg: There is no indication that the signature belongs to the > owner. > > Primary key fingerprint: 5060 6DE1 BDBD 5CF8 62A5 95A9 07C5 682D AFC7 3125 > > (AFAIK, this warning is OK) > > - verified package checksum: OK > > $ diff <(cat apache-hive-4.0.0-alpha-2-src.tar.gz.sha256) <(shasum -a 256 > apache-hive-4.0.0-alpha-2-src.tar.gz) > > $ diff <(cat apache-hive-4.0.0-alpha-2-bin.tar.gz.sha256) <(shasum -a 256 > apache-hive-4.0.0-alpha-2-bin.tar.gz) > > - verified maven build (no tests): OK > > $ mvn clean install -DskipTests > > ... > > [INFO] > ------------------------------------------------------------------------ > > [INFO] BUILD SUCCESS > > [INFO] > ------------------------------------------------------------------------ > > [INFO] Total time: 04:31 min > > - checked release notes: OK > > - checked few modules in Nexus: OK > > - environment used: > > $ sw_vers > > ProductName: macOS > > ProductVersion: 11.6.8 > > BuildVersion: 20G730 > > $ mvn --version > > Apache Maven 3.8.1 (05c21c65bdfed0f71a2f2ada8b84da59348c4c5d) > > Maven home: .../.sdkman/candidates/maven/current > > Java version: 1.8.0_292, vendor: AdoptOpenJDK, runtime: > .../.sdkman/candidates/java/8.0.292.hs-adpt/jre > > Default locale: en_IE, platform encoding: UTF-8 > > OS name: "mac os x", version: "10.16", arch: "x86_64", family: "mac" > > $ java -version > > openjdk version "1.8.0_292" > > OpenJDK Runtime Environment (AdoptOpenJDK)(build 1.8.0_292-b10) > > OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.292-b10, mixed mode) > > > Testing in hive-dev-box (https://github.com/kgyrtkirk/hive-dev-box): KO > > This is the setup I have used: > > $ sw hadoop 3.1.0 > > $ sw tez 0.10.0 (tried also 0.10.1) > > $ sw hive > > https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/apache-hive-4.0.0-alpha-2-bin.tar.gz > > In what follows the test data and query I have tried, with associated > stacktrace for the error. It seems a classpath issue, probably there are > multiple versions of the class ending up in the CP and the classloader > happened to load the “wrong one”. > > CREATE TABLE test_stats_a (a int, b int) STORED AS ORC; > > > > INSERT INTO test_stats_a (a, b) VALUES (0, 2); > > INSERT INTO test_stats_a (a, b) VALUES (1, 2); > > INSERT INTO test_stats_a (a, b) VALUES (2, 2); > > INSERT INTO test_stats_a (a, b) VALUES (3, 2); > > INSERT INTO test_stats_a (a, b) VALUES (4, 2); > > INSERT INTO test_stats_a (a, b) VALUES (5, 2); > > INSERT INTO test_stats_a (a, b) VALUES (6, 2); > > INSERT INTO test_stats_a (a, b) VALUES (7, 2); > > INSERT INTO test_stats_a (a, b) VALUES (8, 3); > > INSERT INTO test_stats_a (a, b) VALUES (9, 4); > > INSERT INTO test_stats_a (a, b) VALUES (10, 5); > > INSERT INTO test_stats_a (a, b) VALUES (11, 6); > > INSERT INTO test_stats_a (a, b) VALUES (12, 7); > > INSERT INTO test_stats_a (a, b) VALUES (13, NULL); > > INSERT INTO test_stats_a (a, b) VALUES (14, NULL); > > > > CREATE TABLE test_stats_b (a int, b int) STORED AS ORC; > > > > INSERT INTO test_stats_b (a, b) VALUES (0, 2); > > INSERT INTO test_stats_b (a, b) VALUES (1, 2); > > INSERT INTO test_stats_b (a, b) VALUES (2, 2); > > INSERT INTO test_stats_b (a, b) VALUES (3, 2); > > INSERT INTO test_stats_b (a, b) VALUES (4, 2); > > INSERT INTO test_stats_b (a, b) VALUES (5, 2); > > INSERT INTO test_stats_b (a, b) VALUES (6, 2); > > INSERT INTO test_stats_b (a, b) VALUES (7, 2); > > INSERT INTO test_stats_b (a, b) VALUES (8, 3); > > INSERT INTO test_stats_b (a, b) VALUES (9, 4); > > INSERT INTO test_stats_b (a, b) VALUES (10, 5); > > INSERT INTO test_stats_b (a, b) VALUES (11, 6); > > INSERT INTO test_stats_b (a, b) VALUES (12, 7); > > INSERT INTO test_stats_b (a, b) VALUES (13, NULL); > > INSERT INTO test_stats_b (a, b) VALUES (14, NULL); > > > > CREATE TABLE test_stats_c (a string, b int) STORED AS PARQUET; > > > > INSERT INTO test_stats_c (a, b) VALUES ("a", 2); > > INSERT INTO test_stats_c (a, b) VALUES ("b", 2); > > INSERT INTO test_stats_c (a, b) VALUES ("c", 2); > > INSERT INTO test_stats_c (a, b) VALUES ("d", 2); > > INSERT INTO test_stats_c (a, b) VALUES ("e", 2); > > INSERT INTO test_stats_c (a, b) VALUES ("f", 2); > > INSERT INTO test_stats_c (a, b) VALUES ("g", 2); > > INSERT INTO test_stats_c (a, b) VALUES ("h", 2); > > INSERT INTO test_stats_c (a, b) VALUES ("i", 3); > > INSERT INTO test_stats_c (a, b) VALUES ("j", 4); > > INSERT INTO test_stats_c (a, b) VALUES ("k", 5); > > INSERT INTO test_stats_c (a, b) VALUES ("l", 6); > > INSERT INTO test_stats_c (a, b) VALUES ("m", 7); > > INSERT INTO test_stats_c (a, b) VALUES ("n", NULL); > > INSERT INTO test_stats_c (a, b) VALUES ("o", NULL); > > > SELECT * FROM test_stats_a t1 JOIN test_stats_b t2 ON (t1.a = t2.a) WHERE > > t1.b < 3 AND t2.b > 1; > > > INFO : Completed compiling > > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b); > > Time taken: > > 4.171 seconds > > INFO : Operation QUERY obtained 0 locks > > INFO : Executing > > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b): > > SELECT * FROM test_sta > > ts_a t1 JOIN test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b > 1 > > INFO : Query ID = > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b > > INFO : Total jobs = 1 > > INFO : Launching Job 1 out of 1 > > INFO : Starting task [Stage-1:MAPRED] in serial mode > > DEBUG : Task getting executed using mapred tag : > > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b,userid=de > > v > > INFO : Subscribed to counters: [] for queryId: > > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b > > INFO : Tez session hasn't been created yet. Opening session > > DEBUG : No local resources to process (other than hive-exec) > > INFO : Dag name: SELECT * FROM test_st...... < 3 AND t2.b > 1 (Stage-1) > > DEBUG : DagInfo: {"context":"Hive","description":"SELECT * FROM > > test_stats_a t1 JOIN test_stats_b t2 ON (t1.a = t2 > > .a) WHERE t1.b < 3 AND t2.b > 1"} > > DEBUG : Setting Tez DAG access for > > queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b with > > viewAclStr > > ing=dev, modifyStr=dev > > INFO : Setting tez.task.scale.memory.reserve-fraction to > > 0.30000001192092896 > > INFO : HS2 Host: [alpha2], Query ID: > > [dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b], Dag ID: > > [dag_1666888075798_0001_1], DAG Session ID: > [application_1666888075798_0001] > > INFO : Status: Running (Executing on YARN cluster with App id > > application_1666888075798_0001) > > > > ERROR : Status: Failed > > ERROR : Vertex failed, vertexName=Map 2, > > vertexId=vertex_1666888075798_0001_1_01, diagnostics=[Vertex > > vertex_1666888075798_0001_1_01 [Map 2] killed/failed due > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed, > > vertex=vertex_1666888075798_0001_1_01 [Map 2], > java.lang.NoSuchMethodError: > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401) > > at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355) > > at java.util.TimSort.sort(TimSort.java:220) > > at java.util.Arrays.sort(Arrays.java:1438) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:422) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) > > at > > > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > > at java.lang.Thread.run(Thread.java:748) > > ] > > ERROR : Vertex failed, vertexName=Map 1, > > vertexId=vertex_1666888075798_0001_1_00, diagnostics=[Vertex > > vertex_1666888075798_0001_1_00 [Map 1] killed/failed due > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed, > > vertex=vertex_1666888075798_0001_1_00 [Map 1], > java.lang.NoSuchMethodError: > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401) > > at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355) > > at java.util.TimSort.sort(TimSort.java:220) > > at java.util.Arrays.sort(Arrays.java:1438) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:422) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) > > at > > > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > > at java.lang.Thread.run(Thread.java:748) > > ] > > ERROR : DAG did not succeed due to VERTEX_FAILURE. failedVertices:2 > > killedVertices:0 > > ERROR : FAILED: Execution Error, return code 2 from > > org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed, vertexName=Map > > 2, vertexId=vertex_1666888075798_0001_1_01, diagnostics=[Vertex > > vertex_1666888075798_0001_1_01 [Map 2] killed/failed due > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed, > > vertex=vertex_1666888075798_0001_1_01 [Map 2], > java.lang.NoSuchMethodError: > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401) > > at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355) > > at java.util.TimSort.sort(TimSort.java:220) > > at java.util.Arrays.sort(Arrays.java:1438) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:422) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) > > at > > > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > > at java.lang.Thread.run(Thread.java:748) > > ]Vertex failed, vertexName=Map 1, > vertexId=vertex_1666888075798_0001_1_00, > > diagnostics=[Vertex vertex_1666888075798_0001_1_00 [Map 1] killed/failed > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed, > > vertex=vertex_1666888075798_0001_1_00 [Map 1], > java.lang.NoSuchMethodError: > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401) > > at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355) > > at java.util.TimSort.sort(TimSort.java:220) > > at java.util.Arrays.sort(Arrays.java:1438) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:422) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) > > at > > > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > > at java.lang.Thread.run(Thread.java:748) > > ]DAG did not succeed due to VERTEX_FAILURE. failedVertices:2 > > killedVertices:0 > > DEBUG : Shutting down query SELECT * FROM test_stats_a t1 JOIN > > test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b > 1 > > INFO : Completed executing > > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b); > > Time taken: 6.983 seconds > > DEBUG : Shutting down query SELECT * FROM test_stats_a t1 JOIN > > test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b > 1 > > Error: Error while compiling statement: FAILED: Execution Error, return > > code 2 from org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed, > > vertexName=Map 2, vertexId=vertex_1666888075798_0001_1_01, > > diagnostics=[Vertex vertex_1666888075798_0001_1_01 [Map 2] killed/failed > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer failed, > > vertex=vertex_1666888075798_0001_1_01 [Map 2], > java.lang.NoSuchMethodError: > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401) > > at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355) > > at java.util.TimSort.sort(TimSort.java:220) > > at java.util.Arrays.sort(Arrays.java:1438) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:422) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) > > at > > > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > > at java.lang.Thread.run(Thread.java:748) > > ]Vertex failed, vertexName=Map 1, > vertexId=vertex_1666888075798_0001_1_00, > > diagnostics=[Vertex vertex_1666888075798_0001_1_00 [Map 1] killed/failed > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer failed, > > vertex=vertex_1666888075798_0001_1_00 [Map 1], > java.lang.NoSuchMethodError: > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401) > > at java.util.TimSort.countRunAndMakeAscending(TimSort.java:355) > > at java.util.TimSort.sort(TimSort.java:220) > > at java.util.Arrays.sort(Arrays.java:1438) > > at > > > org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:422) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272) > > at > > > org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) > > at > > > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57) > > at > > > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > > at java.lang.Thread.run(Thread.java:748) > > ]DAG did not succeed due to VERTEX_FAILURE. failedVertices:2 > > killedVertices:0 (state=08S01,code=2) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > > > > Best regards, > Alessandro > > On Thu, 27 Oct 2022 at 19:01, Ayush Saxena <ayush...@gmail.com> wrote: > > > Chris, > > The KEYS file is at: > > https://downloads.apache.org/hive/KEYS > > > > -Ayush > > > > On Thu, 27 Oct 2022 at 21:58, Chris Nauroth <cnaur...@apache.org> wrote: > > > > > Could someone please point me toward the right KEYS file to import so > > that > > > I can verify signatures? Thanks! > > > > > > I'm seeing numerous test failures due to "Insufficient configured > > threads" > > > while trying to start the HTTP server. One example is > > > TestBeelinePasswordOption. Is anyone else seeing this? I noticed that > > > HIVE-24484 set hive.server2.webui.max.threads to 4 in > > > /data/conf/hive-site.xml. (The default in HiveConf.java is 50.) > > > > > > [INFO] Running org.apache.hive.beeline.TestBeelinePasswordOption > > > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: > > > 11.742 s <<< FAILURE! - in > > > org.apache.hive.beeline.TestBeelinePasswordOption > > > [ERROR] org.apache.hive.beeline.TestBeelinePasswordOption Time > elapsed: > > > 11.733 s <<< ERROR! > > > org.apache.hive.service.ServiceException: > > java.lang.IllegalStateException: > > > Insufficient configured threads: required=4 < max=4 for > > > > > > > > > QueuedThreadPool[hiveserver2-web]@628bd77e{STARTED,4<=4<=4,i=4,r=-1,q=0}[ReservedThreadExecutor@cfacf0 > > > {s=0/1,p=0}] > > > at > org.apache.hive.service.server.HiveServer2.start(HiveServer2.java:733) > > > at org.apache.hive.jdbc.miniHS2.MiniHS2.start(MiniHS2.java:395) > > > at > > > > > > > > > org.apache.hive.beeline.TestBeelinePasswordOption.preTests(TestBeelinePasswordOption.java:60) > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > at > > > > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > > at > > > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > > at java.lang.reflect.Method.invoke(Method.java:498) > > > at > > > > > > > > > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) > > > at > > > > > > > > > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > > > at > > > > > > > > > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) > > > at > > > > > > > > > org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33) > > > at > > > > > > > > > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) > > > at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) > > > at org.junit.runners.ParentRunner.run(ParentRunner.java:413) > > > at > > > > > > > > > org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365) > > > at > > > > > > > > > org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273) > > > at > > > > > > > > > org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238) > > > at > > > > > > > > > org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159) > > > at > > > > > > > > > org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377) > > > at > > > > > > > > > org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138) > > > at > > org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465) > > > at > > > > org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451) > > > Caused by: java.lang.IllegalStateException: Insufficient configured > > > threads: required=4 < max=4 for > > > > > > > > > QueuedThreadPool[hiveserver2-web]@628bd77e{STARTED,4<=4<=4,i=4,r=-1,q=0}[ReservedThreadExecutor@cfacf0 > > > {s=0/1,p=0}] > > > at > > > > > > > > > org.eclipse.jetty.util.thread.ThreadPoolBudget.check(ThreadPoolBudget.java:165) > > > at > > > > > > > > > org.eclipse.jetty.util.thread.ThreadPoolBudget.leaseTo(ThreadPoolBudget.java:141) > > > at > > > > > > > > > org.eclipse.jetty.util.thread.ThreadPoolBudget.leaseFrom(ThreadPoolBudget.java:191) > > > at org.eclipse.jetty.io > > .SelectorManager.doStart(SelectorManager.java:255) > > > at > > > > > > > > > org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73) > > > at > > > > > > > > > org.eclipse.jetty.util.component.ContainerLifeCycle.start(ContainerLifeCycle.java:169) > > > at > > > > > > > > > org.eclipse.jetty.util.component.ContainerLifeCycle.doStart(ContainerLifeCycle.java:110) > > > at > > > > > > > > > org.eclipse.jetty.server.AbstractConnector.doStart(AbstractConnector.java:321) > > > at > > > > > > > > > org.eclipse.jetty.server.AbstractNetworkConnector.doStart(AbstractNetworkConnector.java:81) > > > at > > > > > > org.eclipse.jetty.server.ServerConnector.doStart(ServerConnector.java:234) > > > at > > > > > > > > > org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73) > > > at org.eclipse.jetty.server.Server.doStart(Server.java:401) > > > at > > > > > > > > > org.eclipse.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:73) > > > at org.apache.hive.http.HttpServer.start(HttpServer.java:335) > > > at > org.apache.hive.service.server.HiveServer2.start(HiveServer2.java:729) > > > ... 21 more > > > > > > Chris Nauroth > > > > > > > > > On Thu, Oct 27, 2022 at 7:48 AM Ádám Szita <sz...@apache.org> wrote: > > > > > > > Hi, > > > > > > > > Thanks for rebuilding this RC, Denys. > > > > > > > > Alessandro: IMHO since there was no vote cast yet and we're talking > > about > > > > a build option change only, I guess it just doesn't worth rebuilding > > the > > > > whole stuff from scratch to create a new RC. > > > > > > > > I give +1 (binding) to this RC, I verified the checksum, binary > > content, > > > > source content, built Hive from source and also tried out the > artifacts > > > in > > > > a mini cluster environment. Built an HMS DB with the schema scripts > > > > provided, did table creation, insert, delete, rollback (Iceberg). > > > > > > > > Thanks again, Denys for taking this up. > > > > > > > > On 2022/10/27 13:29:36 Alessandro Solimando wrote: > > > > > Hi Denys, > > > > > in other Apache communities I generally see that votes are > cancelled > > > and > > > > a > > > > > new RC is prepared when there are changes or blocking issues like > in > > > this > > > > > case, not sure how things are done in Hive though. > > > > > > > > > > Best regards, > > > > > Alessandro > > > > > > > > > > On Thu, 27 Oct 2022 at 15:22, Denys Kuzmenko < > dkuzme...@cloudera.com > > > > .invalid> > > > > > wrote: > > > > > > > > > > > Hi Adam, > > > > > > > > > > > > Thanks for pointing that out! Upstream release guide is outdated. > > > Once > > > > I > > > > > > receive the edit rights, I'll amend the instructions. > > > > > > Updated the release artifacts and checksums: > > > > > > > > > > > > Apache Hive 4.0.0-alpha-2 Release Candidate 0 is available > > > > > > here: > > https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/ > > > > > > > > > > > > > > > > > > The checksums are these: > > > > > > - > b4dbaac5530694f631af13677ffe5443addc148bd94176b27a109a6da67f5e0f > > > > > > apache-hive-4.0.0-alpha-2-bin.tar.gz > > > > > > - > 8c4639915e9bf649f4a55cd9adb9d266aa15d8fa48ddfadb28ebead2c0aee4d0 > > > > > > apache-hive-4.0.0-alpha-2-src.tar.gz > > > > > > > > > > > > Maven artifacts are available > > > > > > here: > > > > > > > > > https://repository.apache.org/content/repositories/orgapachehive-1117/ > > > > > > > > > > > > The tag release-4.0.0-alpha-2-rc0 has been applied to the source > > for > > > > > > this release in github, you can see it at > > > > > > https://github.com/apache/hive/tree/release-4.0.0-alpha-2-rc0 > > > > > > > > > > > > The git commit hash > > > > > > is: > > > > > > > > > > > > > > > > https://github.com/apache/hive/commit/da146200e003712e324496bf560a1702485d231c > > > > > > > > > > > > > > > > > > Please check again. > > > > > > > > > > > > > > > > > > Thanks, > > > > > > Denys > > > > > > > > > > > > On Thu, Oct 27, 2022 at 2:53 PM Ádám Szita <sz...@apache.org> > > wrote: > > > > > > > > > > > > > Hi Denys, > > > > > > > > > > > > > > Unfortutantely I can't give a plus 1 on this yet, as the > Iceberg > > > > > > artifacts > > > > > > > are missing from the binary tar.gz. Perhaps -Piceberg flag was > > > > missing > > > > > > > during build, can you please rebuild? > > > > > > > > > > > > > > Thanks, > > > > > > > Adam > > > > > > > > > > > > > > On 2022/10/25 11:20:23 Denys Kuzmenko wrote: > > > > > > > > Hi team, > > > > > > > > > > > > > > > > > > > > > > > > Apache Hive 4.0.0-alpha-2 Release Candidate 0 is available > > > > > > > > here: > > > > https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/ > > > > > > > > > > > > > > > > > > > > > > > > The checksums are these: > > > > > > > > - > > > 7d4c54ecfe2b04cabc283a84defcc1e8a02eed0e13baba2a2c91ae882b6bfaf7 > > > > > > > > apache-hive-4.0.0-alpha-2-bin.tar.gz > > > > > > > > - > > > 8c4639915e9bf649f4a55cd9adb9d266aa15d8fa48ddfadb28ebead2c0aee4d0 > > > > > > > > apache-hive-4.0.0-alpha-2-src.tar.gz > > > > > > > > > > > > > > > > Maven artifacts are available > > > > > > > > here: > > > > > > > > > > > > https://repository.apache.org/content/repositories/orgapachehive-1117/ > > > > > > > > > > > > > > > > The tag release-4.0.0-alpha-2-rc0 has been applied to the > > source > > > > for > > > > > > > > this release in github, you can see it at > > > > > > > > > https://github.com/apache/hive/tree/release-4.0.0-alpha-2-rc0 > > > > > > > > > > > > > > > > The git commit hash > > > > > > > > is: > > > > > > > > > > > > > > > > > > > > > > > https://github.com/apache/hive/commit/da146200e003712e324496bf560a1702485d231c > > > > > > > > > > > > > > > > Voting will conclude in 72 hours. > > > > > > > > > > > > > > > > Hive PMC Members: Please test and vote. > > > > > > > > > > > > > > > > Thanks > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > >