Hi team,

Thank you for taking time to verify this RC!
Unfortunately, we didn't get enough votes to go ahead with the release.

Closing this vote as unsuccessful.

Kind regards,
Denys

On Fri, Oct 28, 2022, 15:56 Stamatis Zampetakis <zabe...@gmail.com> wrote:

> I think that having a proper NOTICE file in jars is important to comply
> with the ASF release policy:
> * https://www.apache.org/legal/release-policy.html#licensing-documentation
> * https://www.apache.org/legal/src-headers.html#notice
> * https://www.apache.org/legal/src-headers.html#faq-binaries
> The fact that the NOTICE wasn't updated in alpha-1 is most likely an
> oversight.
>
> Having said that the final decision is up to the release manager.
>
> Best,
> Stamatis
>
> On Fri, Oct 28, 2022 at 1:57 PM Denys Kuzmenko
> <dkuzme...@cloudera.com.invalid> wrote:
>
> > Hi Stamatis,
> >
> > My bad, sorry. Removed the ".imp" files and updated the release
> artifacts.*
> > *** NO CODE CHANGES ****
> > I was following the alpha-1 release and the NOTICE wasn't updated there
> as
> > well. I don't think that should be a blocker. Noted that + javadoc
> > artifacts for the new RC.
> >
> > fc7908f40ec854671c6795acb525649d83c071d70cf62961dc90a251a0f45e47
> >  apache-hive-4.0.0-alpha-2-bin.tar.gz
> > f2814aadeca56ad1d8d9f7797b99d1670f6450f68ff6cae829384c9c102cd7a9
> >  apache-hive-4.0.0-alpha-2-src.tar.gz
> >
> > Thanks,
> > Denys
> >
> > On Fri, Oct 28, 2022 at 12:28 PM Stamatis Zampetakis <zabe...@gmail.com>
> > wrote:
> >
> > > -1 (non-binding)
> > >
> > > Ubuntu 20.04.5 LTS, java version "1.8.0_261", Apache Maven 3.6.3
> > >
> > > * Verified signatures and checksums OK
> > > * Checked diff between git repo and release sources (diff -qr hive-git
> > > hive-src) KO (among other *.iml files present in release sources but
> not
> > in
> > > git)
> > > * Checked LICENSE, NOTICE, and README.md file OK
> > > * Built from release sources (mvn clean install -DskipTests -Pitests)
> OK
> > > * Package binaries from release sources (mvn clean package -DskipTests)
> > OK
> > > * Built from git tag (mvn clean install -DskipTests -Pitests) OK
> > > * Run smoke tests on pseudo cluster using hive-dev-box [1] OK
> > > * Spot check maven artifacts for general structure, LICENSE, NOTICE,
> > > META-INF content KO (NOTICE file in hive-exec-4.0.0-alpha-2.jar has
> > > copyright for 2020)
> > >
> > > Smoke tests included: * Derby metastore initialization * simple CREATE
> > > TABLE statements; * basic INSERT INTO VALUES statements; * basic SELECT
> > > statements with simple INNER JOIN, WHERE, and GROUP BY variations; *
> > > EXPLAIN statement variations; * ANALYZE TABLE variations;
> > >
> > > The negative vote is for the spurious *.iml (IntelliJ project) files
> > > present in the release sources and the outdated NOTICE file in maven
> > > artifacts).
> > >
> > > Also javadoc artifacts are missing from maven staging repo. I checked
> > > previous releases and it seems that they were not there as well so this
> > is
> > > not blocking but may be worth fixing for the next release.
> > >
> > > Best,
> > > Stamatis
> > >
> > > [1] https://lists.apache.org/thread/7yqs7o6ncpottqx8txt0dtt9858ypsbb
> > >
> > >
> >
> https://repository.apache.org/content/repositories/orgapachehive-1117/org/apache/hive/hive-exec/4.0.0-alpha-2/hive-exec-4.0.0-alpha-2.jar
> > >
> > > On Fri, Oct 28, 2022 at 10:32 AM Ayush Saxena <ayush...@gmail.com>
> > wrote:
> > >
> > > > +1 (non-binding)
> > > > * Built from source.
> > > > * Verified Checksums.
> > > > * Verified Signatures
> > > > * Ran some basic unit tests.
> > > > * Ran some basic ACID & Iceberg related queries with Tez.
> > > > * Skimmed through the Maven Artifacts, Looks Good.
> > > >
> > > > Thanx Denys for driving the release. Good Luck!!!
> > > >
> > > > -Ayush
> > > >
> > > > On Fri, 28 Oct 2022 at 13:46, Denys Kuzmenko <dkuzme...@cloudera.com
> > > > .invalid>
> > > > wrote:
> > > >
> > > > > Extending voting for 24hr. 1 more +1 is needed from the PMC to
> > promote
> > > > the
> > > > > release.
> > > > > If not given, I'll be closing this vote as unsuccessful.
> > > > >
> > > > > On Thu, Oct 27, 2022 at 11:16 PM Chris Nauroth <
> cnaur...@apache.org>
> > > > > wrote:
> > > > >
> > > > > > +1 (non-binding)
> > > > > >
> > > > > > * Verified all checksums.
> > > > > > * Verified all signatures.
> > > > > > * Built from source.
> > > > > >     * mvn clean install -Piceberg -DskipTests
> > > > > > * Tests passed.
> > > > > >     * mvn --fail-never clean verify -Piceberg -Pitests
> > > > > > -Dmaven.test.jvm.args='-Xmx2048m -DJETTY_AVAILABLE_PROCESSORS=4'
> > > > > >
> > > > > > I figured out why my test runs were failing in HTTP server
> > > > > initialization.
> > > > > > Jetty enforces thread leasing to warn or abort if there aren't
> > enough
> > > > > > threads available [1]. During startup, it attempts to lease a
> > thread
> > > > per
> > > > > > NIO selector [2]. By default, the number of NIO selectors to use
> is
> > > > > > determined based on available CPUs [3]. This is mostly a
> > passthrough
> > > to
> > > > > > Runtime.availableProcessors() [4]. In my case, running on a
> machine
> > > > with
> > > > > 16
> > > > > > CPUs, this ended up creating more than 4 selectors, therefore
> > > requiring
> > > > > > more than 4 threads and violating the lease check. I was able to
> > work
> > > > > > around this by passing the JETTY_AVAILABLE_PROCESSORS system
> > property
> > > > to
> > > > > > constrain the number of CPUs available to Jetty.
> > > > > >
> > > > > > If we are intentionally constraining the pool to 4 threads during
> > > > itests,
> > > > > > then would it also make sense to limit JETTY_AVAILABLE_PROCESSORS
> > in
> > > > > > maven.test.jvm.args of the root pom.xml, so that others don't run
> > > into
> > > > > this
> > > > > > problem later? If so, I'll send a pull request.
> > > > > >
> > > > > > [1]
> > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/eclipse/jetty.project/blob/jetty-9.4.40.v20210413/jetty-util/src/main/java/org/eclipse/jetty/util/thread/ThreadPoolBudget.java#L165
> > > > > > [2]
> > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/eclipse/jetty.project/blob/jetty-9.4.40.v20210413/jetty-io/src/main/java/org/eclipse/jetty/io/SelectorManager.java#L255
> > > > > > [3]
> > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/eclipse/jetty.project/blob/jetty-9.4.40.v20210413/jetty-io/src/main/java/org/eclipse/jetty/io/SelectorManager.java#L79
> > > > > > [4]
> > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/eclipse/jetty.project/blob/jetty-9.4.40.v20210413/jetty-util/src/main/java/org/eclipse/jetty/util/ProcessorUtils.java#L45
> > > > > >
> > > > > > Chris Nauroth
> > > > > >
> > > > > >
> > > > > > On Thu, Oct 27, 2022 at 1:18 PM Alessandro Solimando <
> > > > > > alessandro.solima...@gmail.com> wrote:
> > > > > >
> > > > > > > You are right Ayush, I got sidetracked by the release notes (*
> > > > > > [HIVE-19217]
> > > > > > > - Upgrade to Hadoop 3.1.0) and I did not check the versions in
> > the
> > > > pom
> > > > > > > file, apologies for the false alarm but better safe than sorry.
> > > > > > >
> > > > > > > With the right versions in place (Hadoop 3.3.1 and Tez 10.0.2),
> > > tests
> > > > > > > including select, join, groupby, orderby, explain (ast, cbo,
> cbo
> > > > cost,
> > > > > > > vectorization) are working correctly, against data in ORC and
> > > parquet
> > > > > > > file format.
> > > > > > >
> > > > > > > No problem for me either when running TestBeelinePasswordOption
> > > > > locally.
> > > > > > >
> > > > > > > So my vote turns into a +1 (non-binding).
> > > > > > >
> > > > > > > Thanks a lot Denys for pushing the release process forward,
> sorry
> > > > again
> > > > > > you
> > > > > > > all for the oversight!
> > > > > > >
> > > > > > > Best regards,
> > > > > > > Alessandro
> > > > > > >
> > > > > > > On Thu, 27 Oct 2022 at 20:03, Ayush Saxena <ayush...@gmail.com
> >
> > > > wrote:
> > > > > > >
> > > > > > > > Hi Alessandro,
> > > > > > > > From this:
> > > > > > > >
> > > > > > > > > $ sw hadoop 3.1.0
> > > > > > > > > $ sw tez 0.10.0 (tried also 0.10.1)
> > > > > > > >
> > > > > > > >
> > > > > > > > I guess you are using the wrong versions, The Hadoop version
> to
> > > be
> > > > > used
> > > > > > > > should be 3.3.1[1] and the Tez version should be 0.10.2[2]
> > > > > > > >
> > > > > > > > The error also seems to be coming from Hadoop code
> > > > > > > >
> > > > > > > > > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> > > > > > > > > java.lang.NoSuchMethodError:
> > > > > > > > > >
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > > > > > >
> > > > > > > >
> > > > > > > > The compareTo method in Hadoop was changed in HADOOP-16196,
> > which
> > > > > isn't
> > > > > > > > there in Hadoop-3.1.0, it is there post 3.2.1 [3]
> > > > > > > >
> > > > > > > > Another stuff, TestBeelinePasswordOptions passes for me
> inside
> > > the
> > > > > > source
> > > > > > > > directory.
> > > > > > > >
> > > > > > > > [*INFO*]
> > -------------------------------------------------------
> > > > > > > >
> > > > > > > > [*INFO*]  T E S T S
> > > > > > > >
> > > > > > > > [*INFO*]
> > -------------------------------------------------------
> > > > > > > >
> > > > > > > > [*INFO*] Running
> > > > org.apache.hive.beeline.*TestBeelinePasswordOption*
> > > > > > > >
> > > > > > > > [*INFO*] *Tests run: 10*, Failures: 0, Errors: 0, Skipped: 0,
> > > Time
> > > > > > > elapsed:
> > > > > > > > 18.264 s - in
> > org.apache.hive.beeline.*TestBeelinePasswordOption*
> > > > > > > >
> > > > > > > > -Ayush
> > > > > > > >
> > > > > > > > [1]
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/apache/hive/blob/release-4.0.0-alpha-2-rc0/pom.xml#L136
> > > > > > > > [2]
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://github.com/apache/hive/blob/release-4.0.0-alpha-2-rc0/pom.xml#L197
> > > > > > > > [3] https://issues.apache.org/jira/browse/HADOOP-16196
> > > > > > > >
> > > > > > > > On Thu, 27 Oct 2022 at 23:15, Alessandro Solimando <
> > > > > > > > alessandro.solima...@gmail.com> wrote:
> > > > > > > >
> > > > > > > > > Hi everyone,
> > > > > > > > >
> > > > > > > > > unfortunately my vote is -1 (although non-binding) due to a
> > > > > classpath
> > > > > > > > error
> > > > > > > > > which prevents queries involving Tez to complete (all the
> > > details
> > > > > at
> > > > > > > the
> > > > > > > > > end of the email, apologies for the lengthy text but I
> wanted
> > > to
> > > > > > > provide
> > > > > > > > > all the context).
> > > > > > > > >
> > > > > > > > > - verified gpg signature: OK
> > > > > > > > >
> > > > > > > > > $ wget https://www.apache.org/dist/hive/KEYS
> > > > > > > > >
> > > > > > > > > $ gpg --import KEYS
> > > > > > > > >
> > > > > > > > > ...
> > > > > > > > >
> > > > > > > > > $ gpg --verify apache-hive-4.0.0-alpha-2-bin.tar.gz.asc
> > > > > > > > > apache-hive-4.0.0-alpha-2-bin.tar.gz
> > > > > > > > >
> > > > > > > > > gpg: Signature made Thu 27 Oct 15:11:48 2022 CEST
> > > > > > > > >
> > > > > > > > > gpg:                using RSA key
> > > > > > > > 50606DE1BDBD5CF862A595A907C5682DAFC73125
> > > > > > > > >
> > > > > > > > > gpg:                issuer "dkuzme...@apache.org"
> > > > > > > > >
> > > > > > > > > gpg: Good signature from "Denys Kuzmenko (CODE SIGNING
> KEY) <
> > > > > > > > > dkuzme...@apache.org>" [unknown]
> > > > > > > > >
> > > > > > > > > gpg: WARNING: The key's User ID is not certified with a
> > trusted
> > > > > > > > signature!
> > > > > > > > >
> > > > > > > > > gpg:          There is no indication that the signature
> > belongs
> > > > to
> > > > > > the
> > > > > > > > > owner.
> > > > > > > > >
> > > > > > > > > Primary key fingerprint: 5060 6DE1 BDBD 5CF8 62A5  95A9
> 07C5
> > > 682D
> > > > > > AFC7
> > > > > > > > 3125
> > > > > > > > >
> > > > > > > > > $ gpg --verify apache-hive-4.0.0-alpha-2-src.tar.gz.asc
> > > > > > > > > apache-hive-4.0.0-alpha-2-src.tar.gz
> > > > > > > > >
> > > > > > > > > gpg: Signature made Thu 27 Oct 15:12:08 2022 CEST
> > > > > > > > >
> > > > > > > > > gpg:                using RSA key
> > > > > > > > 50606DE1BDBD5CF862A595A907C5682DAFC73125
> > > > > > > > >
> > > > > > > > > gpg:                issuer "dkuzme...@apache.org"
> > > > > > > > >
> > > > > > > > > gpg: Good signature from "Denys Kuzmenko (CODE SIGNING
> KEY) <
> > > > > > > > > dkuzme...@apache.org>" [unknown]
> > > > > > > > >
> > > > > > > > > gpg: WARNING: The key's User ID is not certified with a
> > trusted
> > > > > > > > signature!
> > > > > > > > >
> > > > > > > > > gpg:          There is no indication that the signature
> > belongs
> > > > to
> > > > > > the
> > > > > > > > > owner.
> > > > > > > > >
> > > > > > > > > Primary key fingerprint: 5060 6DE1 BDBD 5CF8 62A5  95A9
> 07C5
> > > 682D
> > > > > > AFC7
> > > > > > > > 3125
> > > > > > > > >
> > > > > > > > > (AFAIK, this warning is OK)
> > > > > > > > >
> > > > > > > > > - verified package checksum: OK
> > > > > > > > >
> > > > > > > > > $ diff <(cat apache-hive-4.0.0-alpha-2-src.tar.gz.sha256)
> > > > <(shasum
> > > > > -a
> > > > > > > 256
> > > > > > > > > apache-hive-4.0.0-alpha-2-src.tar.gz)
> > > > > > > > >
> > > > > > > > > $ diff <(cat apache-hive-4.0.0-alpha-2-bin.tar.gz.sha256)
> > > > <(shasum
> > > > > -a
> > > > > > > 256
> > > > > > > > > apache-hive-4.0.0-alpha-2-bin.tar.gz)
> > > > > > > > >
> > > > > > > > > - verified maven build (no tests): OK
> > > > > > > > >
> > > > > > > > > $ mvn clean install -DskipTests
> > > > > > > > >
> > > > > > > > > ...
> > > > > > > > >
> > > > > > > > > [INFO]
> > > > > > > > >
> > > > > > >
> > > > >
> > >
> ------------------------------------------------------------------------
> > > > > > > > >
> > > > > > > > > [INFO] BUILD SUCCESS
> > > > > > > > >
> > > > > > > > > [INFO]
> > > > > > > > >
> > > > > > >
> > > > >
> > >
> ------------------------------------------------------------------------
> > > > > > > > >
> > > > > > > > > [INFO] Total time:  04:31 min
> > > > > > > > >
> > > > > > > > > - checked release notes: OK
> > > > > > > > >
> > > > > > > > > - checked few modules in Nexus: OK
> > > > > > > > >
> > > > > > > > > - environment used:
> > > > > > > > >
> > > > > > > > > $ sw_vers
> > > > > > > > >
> > > > > > > > > ProductName: macOS
> > > > > > > > >
> > > > > > > > > ProductVersion: 11.6.8
> > > > > > > > >
> > > > > > > > > BuildVersion: 20G730
> > > > > > > > >
> > > > > > > > > $ mvn --version
> > > > > > > > >
> > > > > > > > > Apache Maven 3.8.1
> (05c21c65bdfed0f71a2f2ada8b84da59348c4c5d)
> > > > > > > > >
> > > > > > > > > Maven home: .../.sdkman/candidates/maven/current
> > > > > > > > >
> > > > > > > > > Java version: 1.8.0_292, vendor: AdoptOpenJDK, runtime:
> > > > > > > > > .../.sdkman/candidates/java/8.0.292.hs-adpt/jre
> > > > > > > > >
> > > > > > > > > Default locale: en_IE, platform encoding: UTF-8
> > > > > > > > >
> > > > > > > > > OS name: "mac os x", version: "10.16", arch: "x86_64",
> > family:
> > > > > "mac"
> > > > > > > > >
> > > > > > > > > $ java -version
> > > > > > > > >
> > > > > > > > > openjdk version "1.8.0_292"
> > > > > > > > >
> > > > > > > > > OpenJDK Runtime Environment (AdoptOpenJDK)(build
> > 1.8.0_292-b10)
> > > > > > > > >
> > > > > > > > > OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.292-b10,
> > mixed
> > > > > mode)
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > Testing in hive-dev-box (
> > > > https://github.com/kgyrtkirk/hive-dev-box
> > > > > ):
> > > > > > > KO
> > > > > > > > >
> > > > > > > > > This is the setup I have used:
> > > > > > > > >
> > > > > > > > > $ sw hadoop 3.1.0
> > > > > > > > >
> > > > > > > > > $ sw tez 0.10.0 (tried also 0.10.1)
> > > > > > > > >
> > > > > > > > > $ sw hive
> > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> https://people.apache.org/~dkuzmenko/release-4.0.0-alpha-2-rc0/apache-hive-4.0.0-alpha-2-bin.tar.gz
> > > > > > > > >
> > > > > > > > > In what follows the test data and query I have tried, with
> > > > > associated
> > > > > > > > > stacktrace for the error. It seems a classpath issue,
> > probably
> > > > > there
> > > > > > > are
> > > > > > > > > multiple versions of the class ending up in the CP and the
> > > > > > classloader
> > > > > > > > > happened to load the “wrong one”.
> > > > > > > > >
> > > > > > > > > CREATE TABLE test_stats_a (a int, b int) STORED AS ORC;
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (0, 2);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (1, 2);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (2, 2);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (3, 2);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (4, 2);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (5, 2);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (6, 2);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (7, 2);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (8, 3);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (9, 4);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (10, 5);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (11, 6);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (12, 7);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (13, NULL);
> > > > > > > > > > INSERT INTO test_stats_a (a, b) VALUES (14, NULL);
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > > CREATE TABLE test_stats_b (a int, b int) STORED AS ORC;
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (0, 2);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (1, 2);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (2, 2);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (3, 2);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (4, 2);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (5, 2);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (6, 2);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (7, 2);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (8, 3);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (9, 4);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (10, 5);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (11, 6);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (12, 7);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (13, NULL);
> > > > > > > > > > INSERT INTO test_stats_b (a, b) VALUES (14, NULL);
> > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > CREATE TABLE test_stats_c (a string, b int) STORED AS
> > PARQUET;
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("a", 2);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("b", 2);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("c", 2);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("d", 2);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("e", 2);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("f", 2);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("g", 2);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("h", 2);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("i", 3);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("j", 4);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("k", 5);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("l", 6);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("m", 7);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("n", NULL);
> > > > > > > > > > INSERT INTO test_stats_c (a, b) VALUES ("o", NULL);
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > SELECT * FROM test_stats_a t1 JOIN test_stats_b t2 ON
> (t1.a =
> > > > t2.a)
> > > > > > > WHERE
> > > > > > > > > > t1.b < 3 AND t2.b > 1;
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > INFO  : Completed compiling
> > > > > > > > > >
> > > > > > > >
> > > > > >
> > > >
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b);
> > > > > > > > > > Time taken:
> > > > > > > > > > 4.171 seconds
> > > > > > > > > > INFO  : Operation QUERY obtained 0 locks
> > > > > > > > > > INFO  : Executing
> > > > > > > > > >
> > > > > > > >
> > > > > >
> > > >
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b):
> > > > > > > > > > SELECT * FROM test_sta
> > > > > > > > > > ts_a t1 JOIN test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b
> <
> > 3
> > > > AND
> > > > > > t2.b
> > > > > > > > > 1
> > > > > > > > > > INFO  : Query ID =
> > > > > > > > > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b
> > > > > > > > > > INFO  : Total jobs = 1
> > > > > > > > > > INFO  : Launching Job 1 out of 1
> > > > > > > > > > INFO  : Starting task [Stage-1:MAPRED] in serial mode
> > > > > > > > > > DEBUG : Task getting executed using mapred tag :
> > > > > > > > > >
> > > > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b,userid=de
> > > > > > > > > > v
> > > > > > > > > > INFO  : Subscribed to counters: [] for queryId:
> > > > > > > > > > dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b
> > > > > > > > > > INFO  : Tez session hasn't been created yet. Opening
> > session
> > > > > > > > > > DEBUG : No local resources to process (other than
> > hive-exec)
> > > > > > > > > > INFO  : Dag name: SELECT * FROM test_st...... < 3 AND
> t2.b
> > >
> > > 1
> > > > > > > > (Stage-1)
> > > > > > > > > > DEBUG : DagInfo: {"context":"Hive","description":"SELECT
> *
> > > FROM
> > > > > > > > > > test_stats_a t1 JOIN test_stats_b t2 ON (t1.a = t2
> > > > > > > > > > .a) WHERE t1.b < 3 AND t2.b > 1"}
> > > > > > > > > > DEBUG : Setting Tez DAG access for
> > > > > > > > > >
> > > queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b
> > > > > > with
> > > > > > > > > > viewAclStr
> > > > > > > > > > ing=dev, modifyStr=dev
> > > > > > > > > > INFO  : Setting tez.task.scale.memory.reserve-fraction to
> > > > > > > > > > 0.30000001192092896
> > > > > > > > > > INFO  : HS2 Host: [alpha2], Query ID:
> > > > > > > > > >
> [dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b],
> > > Dag
> > > > > ID:
> > > > > > > > > > [dag_1666888075798_0001_1], DAG Session ID:
> > > > > > > > > [application_1666888075798_0001]
> > > > > > > > > > INFO  : Status: Running (Executing on YARN cluster with
> App
> > > id
> > > > > > > > > > application_1666888075798_0001)
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > > ERROR : Status: Failed
> > > > > > > > > > ERROR : Vertex failed, vertexName=Map 2,
> > > > > > > > > > vertexId=vertex_1666888075798_0001_1_01,
> > diagnostics=[Vertex
> > > > > > > > > > vertex_1666888075798_0001_1_01 [Map 2] killed/failed due
> > > > > > > > > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer
> > > > failed,
> > > > > > > > > > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> > > > > > > > > java.lang.NoSuchMethodError:
> > > > > > > > > >
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > > > > > > > >         at
> > > > > > > java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > > > > > > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > > > > > > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > > > > > > >
> > > > > > > > >         at
> java.security.AccessController.doPrivileged(Native
> > > > > Method)
> > > > > > > > > >         at
> > javax.security.auth.Subject.doAs(Subject.java:422)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > > > > > > >
> > > > > > > > >         at java.lang.Thread.run(Thread.java:748)
> > > > > > > > > > ]
> > > > > > > > > > ERROR : Vertex failed, vertexName=Map 1,
> > > > > > > > > > vertexId=vertex_1666888075798_0001_1_00,
> > diagnostics=[Vertex
> > > > > > > > > > vertex_1666888075798_0001_1_00 [Map 1] killed/failed due
> > > > > > > > > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1 initializer
> > > > failed,
> > > > > > > > > > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> > > > > > > > > java.lang.NoSuchMethodError:
> > > > > > > > > >
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > > > > > > > >         at
> > > > > > > java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > > > > > > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > > > > > > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > > > > > > >
> > > > > > > > >         at
> java.security.AccessController.doPrivileged(Native
> > > > > Method)
> > > > > > > > > >         at
> > javax.security.auth.Subject.doAs(Subject.java:422)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > > > > > > > >         at java.lang.Thread.run(Thread.java:748)
> > > > > > > > > > ]
> > > > > > > > > > ERROR : DAG did not succeed due to VERTEX_FAILURE.
> > > > > failedVertices:2
> > > > > > > > > > killedVertices:0
> > > > > > > > > > ERROR : FAILED: Execution Error, return code 2 from
> > > > > > > > > > org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex
> failed,
> > > > > > > > vertexName=Map
> > > > > > > > > > 2, vertexId=vertex_1666888075798_0001_1_01,
> > > diagnostics=[Vertex
> > > > > > > > > > vertex_1666888075798_0001_1_01 [Map 2] killed/failed due
> > > > > > > > > > to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2 initializer
> > > > failed,
> > > > > > > > > > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> > > > > > > > > java.lang.NoSuchMethodError:
> > > > > > > > > >
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > > > > > > > >         at
> > > > > > > java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > > > > > > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > > > > > > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > java.security.AccessController.doPrivileged(Native
> > > > > > Method)
> > > > > > > > > >         at
> > javax.security.auth.Subject.doAs(Subject.java:422)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > > > > > > > >         at java.lang.Thread.run(Thread.java:748)
> > > > > > > > > > ]Vertex failed, vertexName=Map 1,
> > > > > > > > > vertexId=vertex_1666888075798_0001_1_00,
> > > > > > > > > > diagnostics=[Vertex vertex_1666888075798_0001_1_00 [Map
> 1]
> > > > > > > > killed/failed
> > > > > > > > > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1
> > initializer
> > > > > > failed,
> > > > > > > > > > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> > > > > > > > > java.lang.NoSuchMethodError:
> > > > > > > > > >
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > > > > > > > >         at
> > > > > > > java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > > > > > > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > > > > > > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > java.security.AccessController.doPrivileged(Native
> > > > > > Method)
> > > > > > > > > >         at
> > javax.security.auth.Subject.doAs(Subject.java:422)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > > > > > > > >         at java.lang.Thread.run(Thread.java:748)
> > > > > > > > > > ]DAG did not succeed due to VERTEX_FAILURE.
> > failedVertices:2
> > > > > > > > > > killedVertices:0
> > > > > > > > > > DEBUG : Shutting down query SELECT * FROM test_stats_a t1
> > > JOIN
> > > > > > > > > > test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b
> >
> > 1
> > > > > > > > > > INFO  : Completed executing
> > > > > > > > > >
> > > > > > > >
> > > > > >
> > > >
> > command(queryId=dev_20221027162822_de90e30b-a80a-427e-869a-b71799222f4b);
> > > > > > > > > > Time taken: 6.983 seconds
> > > > > > > > > > DEBUG : Shutting down query SELECT * FROM test_stats_a t1
> > > JOIN
> > > > > > > > > > test_stats_b t2 ON (t1.a = t2.a) WHERE t1.b < 3 AND t2.b
> >
> > 1
> > > > > > > > > > Error: Error while compiling statement: FAILED: Execution
> > > > Error,
> > > > > > > return
> > > > > > > > > > code 2 from org.apache.hadoop.hive.ql.exec.tez.TezTask.
> > > Vertex
> > > > > > > failed,
> > > > > > > > > > vertexName=Map 2,
> vertexId=vertex_1666888075798_0001_1_01,
> > > > > > > > > > diagnostics=[Vertex vertex_1666888075798_0001_1_01 [Map
> 2]
> > > > > > > > killed/failed
> > > > > > > > > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t2
> > initializer
> > > > > > failed,
> > > > > > > > > > vertex=vertex_1666888075798_0001_1_01 [Map 2],
> > > > > > > > > java.lang.NoSuchMethodError:
> > > > > > > > > >
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > > > > > > > >         at
> > > > > > > java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > > > > > > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > > > > > > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > java.security.AccessController.doPrivileged(Native
> > > > > > Method)
> > > > > > > > > >         at
> > javax.security.auth.Subject.doAs(Subject.java:422)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > > > > > > > >         at java.lang.Thread.run(Thread.java:748)
> > > > > > > > > > ]Vertex failed, vertexName=Map 1,
> > > > > > > > > vertexId=vertex_1666888075798_0001_1_00,
> > > > > > > > > > diagnostics=[Vertex vertex_1666888075798_0001_1_00 [Map
> 1]
> > > > > > > > killed/failed
> > > > > > > > > > due to:ROOT_INPUT_INIT_FAILURE, Vertex Input: t1
> > initializer
> > > > > > failed,
> > > > > > > > > > vertex=vertex_1666888075798_0001_1_00 [Map 1],
> > > > > > > > > java.lang.NoSuchMethodError:
> > > > > > > > > >
> > > > org.apache.hadoop.fs.Path.compareTo(Lorg/apache/hadoop/fs/Path;)I
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:415)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator$InputSplitComparator.compare(HiveSplitGenerator.java:401)
> > > > > > > > > >         at
> > > > > > > java.util.TimSort.countRunAndMakeAscending(TimSort.java:355)
> > > > > > > > > >         at java.util.TimSort.sort(TimSort.java:220)
> > > > > > > > > >         at java.util.Arrays.sort(Arrays.java:1438)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.hive.ql.exec.tez.HiveSplitGenerator.initialize(HiveSplitGenerator.java:254)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:281)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable$1.run(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > java.security.AccessController.doPrivileged(Native
> > > > > > Method)
> > > > > > > > > >         at
> > javax.security.auth.Subject.doAs(Subject.java:422)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:272)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.tez.dag.app.dag.RootInputInitializerManager$InputInitializerCallable.call(RootInputInitializerManager.java:256)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > > > > > > > > >         at java.lang.Thread.run(Thread.java:748)
> > > > > > > > > > ]DAG did not succeed due to VERTEX_FAILURE.
> > failedVertices:2
> > > > > > > > > > killedVertices:0 (state=08S01,code=2)
> > > > > > > > >
> > > > > > > > >         at
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > > > > > > > > >
> > > > > > > > >
> > > > > > > > > Best regards,
> > > > > > > > > Alessandro
> > > > > > > > >
> > > > > > > > > On Thu, 27 Oct 2022 at 19:01, Ayush Saxena <
> > ayush...@gmail.com
> > > >
> > > > > > wrote:
> > > > > > > > >
> > > > > > > > > > Chris,
> > > > > > > > > > The KEYS file is at:
> > > > > > > > > > https://downloads.apache.org/hive/KEYS
> > > > > > > > > >
> > > > > > > > > > -Ayush
> > > > > > > > > >
> > > > > > > > > > On Thu, 27 Oct 2022 at 21:58, Chris Nauroth <
> > > > cnaur...@apache.org
> > > > > >
> > > > > > > > wrote:
> > > > > > > > > >
> > > > > > > > > > > Could someone please point me toward the right KEYS
> file
> > to
> > > > > > import
> > > > > > > so
> > > > > > > > > > that
> > > > > > > > > > > I can verify signatures? Thanks!
> > > > > > > > > > >
> > > > > > > > > > > I'm seeing numerous test failures due to "Insufficient
> > > > > configured
> > > > > > > > > > threads"
> > > > > > > > > > > while trying to start the HTTP server. One example is
> > > > > > > > > > > TestBeelinePasswordOption. Is anyone else seeing this?
> I
> > > > > noticed
> > > > > > > that
> > > > > > > > > > > HIVE-24484 set hive.server2.webui.max.threads to 4 in
> > > > > > > > > > > /data/conf/hive-site.xml. (The default in HiveConf.java
> > is
> > > > 50.)
> > > > > > > > > > >
> > > > > > > > > > > [INFO] Running
> > > > > org.apache.hive.beeline.TestBeelinePasswordOption
> > > > > > > > > > > [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped:
> 0,
> > > > Time
> > > > > > > > elapsed:
> > > > > > > > > > > 11.742 s <<< FAILURE! - in
> > > > > > > > > > > org.apache.hive.beeline.TestBeelinePasswordOption
> > > > > > > > > > > [ERROR]
> org.apache.hive.beeline.TestBeelinePasswordOption
> > > > Time
> > > > > > > > > elapsed:
> > > > > > > > > > > 11.733 s  <<< ERROR!
> > > > > > > > > > > org.apache.hive.service.ServiceException:
> > > > > > > > > > java.lang.IllegalStateException:
> > > > > > > > > > > Insufficient configured threads: required=4 < max=4 for
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> QueuedThreadPool[hiveserver2-web]@628bd77e{STARTED,4<=4<=4,i=4,r=-1,q=0}[ReservedThreadExecutor@cfacf0
> > > > > > > > > > > {s=0/1,p=0}]
> > > > > > > > > > > at
> > > > > > > > >
> > > > > >
> > > org.apache.hive.service.server.HiveServer2.start(HiveServer2.java:733)
> > > > > > > > > > > at
> > > > org.apache.hive.jdbc.miniHS2.MiniHS2.start(MiniHS2.java:395)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.hive.beeline.TestBeelinePasswordOption.preTests(TestBeelinePasswordOption.java:60)
> > > > > > > > > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > > > Method)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> > > > > > > > > > > at java.lang.reflect.Method.invoke(Method.java:498)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
> > > > > > > > > > > at
> > > > > > org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
> > > > > > > > > > > at
> > > org.junit.runners.ParentRunner.run(ParentRunner.java:413)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138)
> > > > > > > > > > > at
> > > > > > > > > >
> > > > > > > >
> > > > > >
> > > >
> > org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > >
> > > > > > >
> > > > >
> > >
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451)
> > > > > > > > > > > Caused by: java.lang.IllegalStateException:
> Insufficient
> > > > > > configured
> > > > > > > > > > > threads: required=4 < max=4 for
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> QueuedThreadPool[hiveserver2-web]@628bd77e{STARTED,4<=4<=4,i=4,r=-1,q=0}[ReservedThreadExecutor@cfacf0
> > > > > > > > > > > {s=0/1,p=0}]
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> org.eclipse.jetty.util.thread.ThreadPoolBudget.check(ThreadPoolBudget.java:165)
> > > > > > > > > > > at
> > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > >

Reply via email to