Hmm interesting. Let me check on this error. Thanks Mukund. Chao
On Fri, Jan 21, 2022 at 4:42 AM Mukund Madhav Thakur <mtha...@cloudera.com.invalid> wrote: > Checked out the release tag. commit *6da346a358c * > Seeing below error while compiling : > > Duplicate classes found: > > > Found in: > > org.apache.hadoop:hadoop-client-api:jar:3.3.2:compile > > org.apache.hadoop:hadoop-client-minicluster:jar:3.3.2:compile > > Duplicate classes: > > org/apache/hadoop/io/serializer/avro/AvroRecord.class > > org/apache/hadoop/io/serializer/avro/AvroRecord$Builder.class > > org/apache/hadoop/io/serializer/avro/AvroRecord$1.class > > > [*INFO*] > *------------------------------------------------------------------------* > > [*INFO*] *Reactor Summary for Apache Hadoop Client Test Minicluster 3.3.2:* > > [*INFO*] > > [*INFO*] Apache Hadoop Client Test Minicluster .............. *SUCCESS* > [02:17 min] > > [*INFO*] Apache Hadoop Client Packaging Invariants for Test . > *FAILURE* [ 0.221 > s] > > [*INFO*] Apache Hadoop Client Packaging Integration Tests ... *SKIPPED* > > [*INFO*] Apache Hadoop Distribution ......................... *SKIPPED* > > [*INFO*] Apache Hadoop Client Modules ....................... *SKIPPED* > > [*INFO*] Apache Hadoop Tencent COS Support .................. *SKIPPED* > > [*INFO*] Apache Hadoop Cloud Storage ........................ *SKIPPED* > > [*INFO*] Apache Hadoop Cloud Storage Project ................ *SKIPPED* > > [*INFO*] > *------------------------------------------------------------------------* > > [*INFO*] *BUILD FAILURE* > > [*INFO*] > *------------------------------------------------------------------------* > > [*INFO*] Total time: 02:18 min > > [*INFO*] Finished at: 2022-01-21T18:06:11+05:30 > > [*INFO*] > *------------------------------------------------------------------------* > > [*ERROR*] Failed to execute goal > org.apache.maven.plugins:maven-enforcer-plugin:3.0.0-M1:enforce > *(enforce-banned-dependencies)* on project > hadoop-client-check-test-invariants: *Some Enforcer rules have failed. Look > above for specific messages explaining why the rule failed.* -> *[Help 1]* > > > On Fri, Jan 21, 2022 at 9:38 AM Wei-Chiu Chuang <weic...@apache.org> > wrote: > > > I'll find time to check out the RC bits. > > I just feel bad that the tarball is now more than 600MB in size. > > > > On Fri, Jan 21, 2022 at 2:23 AM Steve Loughran > <ste...@cloudera.com.invalid > > > > > wrote: > > > > > *+1 binding.* > > > > > > reviewed binaries, source, artifacts in the staging maven repository in > > > downstream builds. all good. > > > > > > *## test run* > > > > > > checked out the asf github repo at commit 6da346a358c into a location > > > already set up with aws and azure test credentials > > > > > > ran the hadoop-aws tests with -Dparallel-tests -DtestsThreadCount=6 > > > -Dmarkers=delete -Dscale > > > and hadoop-azure against azure cardiff with -Dparallel-tests=abfs > > > -DtestsThreadCount=6 > > > > > > all happy > > > > > > > > > > > > *## binary* > > > downloaded KEYS and imported, so adding your key to my list (also > signed > > > this and updated the key servers) > > > > > > downloaded rc tar and verified > > > ``` > > > > gpg2 --verify hadoop-3.3.2.tar.gz.asc hadoop-3.3.2.tar.gz > > > gpg: Signature made Sat Jan 15 23:41:10 2022 GMT > > > gpg: using RSA key > > DE7FA241EB298D027C97B2A1D8F1A97BE51ECA98 > > > gpg: Good signature from "Chao Sun (CODE SIGNING KEY) < > > sunc...@apache.org > > > >" > > > [full] > > > > > > > > > > cat hadoop-3.3.2.tar.gz.sha512 > > > SHA512 (hadoop-3.3.2.tar.gz) = > > > > > > > > > cdd3d9298ba7d6e63ed63f93c159729ea14d2b7d5e3a0640b1761c86c7714a721f88bdfa8cb1d8d3da316f616e4f0ceaace4f32845ee4441e6aaa7a12b8c647d > > > > > > > shasum -a 512 hadoop-3.3.2.tar.gz > > > > > > > > > cdd3d9298ba7d6e63ed63f93c159729ea14d2b7d5e3a0640b1761c86c7714a721f88bdfa8cb1d8d3da316f616e4f0ceaace4f32845ee4441e6aaa7a12b8c647d > > > hadoop-3.3.2.tar.gz > > > ``` > > > > > > > > > *# cloudstore against staged artifacts* > > > ``` > > > cd ~/.m2/repository/org/apache/hadoop > > > find . -name \*3.3.2\* -print | xargs rm -r > > > ``` > > > ensures no local builds have tainted the repo. > > > > > > in cloudstore mvn build without tests > > > ``` > > > mci -Pextra -Phadoop-3.3.2 -Psnapshots-and-staging > > > ``` > > > this fetches all from asf staging > > > > > > ``` > > > Downloading from ASF Staging: > > > > > > > > > https://repository.apache.org/content/groups/staging/org/apache/hadoop/hadoop-client/3.3.2/hadoop-client-3.3.2.pom > > > Downloaded from ASF Staging: > > > > > > > > > https://repository.apache.org/content/groups/staging/org/apache/hadoop/hadoop-client/3.3.2/hadoop-client-3.3.2.pom > > > (11 kB at 20 kB/s) > > > ``` > > > there's no tests there, but it did audit the download process. FWIW, > that > > > project has switched to logback, so I now have all hadoop imports > > excluding > > > slf4j and log4j. it takes too much effort right now. > > > > > > build works. > > > > > > tested abfs and s3a storediags, all happy > > > > > > > > > > > > > > > *### google GCS against staged artifacts* > > > > > > gcs is now java 11 only, so I had to switch JVMs here. > > > > > > had to add a snapshots and staging profile, after which I could build > and > > > test. > > > > > > ``` > > > -Dhadoop.three.version=3.3.2 -Psnapshots-and-staging > > > ``` > > > two test failures were related to auth failures where the tests were > > trying > > > to raise exceptions but things failed differently > > > ``` > > > [ERROR] Failures: > > > [ERROR] > > > > > > > > > GoogleHadoopFileSystemTest.eagerInitialization_fails_withInvalidCredentialsConfiguration:122 > > > unexpected exception type thrown; expected:<java.io > > .FileNotFoundException> > > > but was:<java.lang.IllegalArgumentException> > > > [ERROR] > > > > > > > > > GoogleHadoopFileSystemTest.lazyInitialization_deleteCall_fails_withInvalidCredentialsConfiguration:100 > > > value of: throwable.getMessage() > > > expected: Failed to create GCS FS > > > but was : A JSON key file may not be specified at the same time as > > > credentials via configuration. > > > > > > ``` > > > > > > I'm not worried here. > > > > > > ran cloudstore's diagnostics against gcs. > > > > > > Nice to see they are now collecting IOStatistics on their input > streams. > > we > > > really need to get this collected through the parquet/orc libs and then > > > through the query engines. > > > > > > ``` > > > > bin/hadoop jar $CLOUDSTORE storediag gs://stevel-london/ > > > > > > ... > > > 2022-01-20 17:52:47,447 [main] INFO diag.StoreDiag > > > (StoreDurationInfo.java:<init>(56)) - Starting: Reading a file > > > gs://stevel-london/dir-9cbfc774-76ff-49c0-b216-d7800369c3e1/file > > > input stream summary: org.apache.hadoop.fs.FSDataInputStream@6cfd9a54: > > > com.google.cloud.hadoop.fs.gcs.GoogleHadoopFSInputStream@78c1372d > > > {counters=((stream_read_close_operations=1) > > > (stream_read_seek_backward_operations=0) (stream_read_total_bytes=7) > > > (stream_read_bytes=7) (stream_read_exceptions=0) > > > (stream_read_seek_operations=0) (stream_read_seek_bytes_skipped=0) > > > (stream_read_operations=3) (stream_read_bytes_backwards_on_seek=0) > > > (stream_read_seek_forward_operations=0) > > > (stream_read_operations_incomplete=1)); > > > gauges=(); > > > minimums=(); > > > maximums=(); > > > means=(); > > > } > > > ... > > > ``` > > > > > > *### source* > > > > > > once I'd done builds and tests which fetched from staging, I did a > local > > > build and test > > > > > > repeated download/validate of source tarball, unzip/untar > > > > > > build with java11. > > > > > > I've not done the test run there, because that directory tree doesn't > > have > > > the credentials, and this mornings run was good. > > > > > > altogether then: very happy. tests good, downstream libraries building > > and > > > linking. > > > > > > On Wed, 19 Jan 2022 at 17:50, Chao Sun <sunc...@apache.org> wrote: > > > > > > > Hi all, > > > > > > > > I've put together Hadoop 3.3.2 RC2 below: > > > > > > > > The RC is available at: > > > > http://people.apache.org/~sunchao/hadoop-3.3.2-RC2/ > > > > The RC tag is at: > > > > https://github.com/apache/hadoop/releases/tag/release-3.3.2-RC2 > > > > The Maven artifacts are staged at: > > > > > > https://repository.apache.org/content/repositories/orgapachehadoop-1332 > > > > > > > > You can find my public key at: > > > > https://downloads.apache.org/hadoop/common/KEYS > > > > > > > > I've done the following tests and they look good: > > > > - Ran all the unit tests > > > > - Started a single node HDFS cluster and tested a few simple commands > > > > - Ran all the tests in Spark using the RC2 artifacts > > > > > > > > Please evaluate the RC and vote, thanks! > > > > > > > > Best, > > > > Chao > > > > > > > > > >