Overall so far so good. hadoop-api-shim: built, tested successfully.
cloudstore: built successfully. Spark: built successfully. Passed hadoop-cloud tests. Ozone: One test failure due to unrelated Ozone issue. This test is being disabled in the latest Ozone code. org.apache.hadoop.hdds.utils.NativeLibraryNotLoadedException: Unable to load library ozone_rocksdb_tools from both java.library.path & resource file libozone_rocksdb_t ools.so from jar. at org.apache.hadoop.hdds.utils.db.managed.ManagedSSTDumpTool.<init>(ManagedSSTDumpTool.java:49) Google gcs: There are two test failures. The tests were added recently by HADOOP-18724 <https://issues.apache.org/jira/browse/HADOOP-18724> in Hadoop 3.3.6. This is okay. Not production code problem. Can be addressed in GCS code. [ERROR] Errors: [ERROR] TestInMemoryGoogleContractOpen>AbstractContractOpenTest.testFloatingPointLength:403 » IllegalArgument Unknown mandatory key for gs://fake-in-memory-test-buck et/contract-test/testFloatingPointLength "fs.option.openfile.length" [ERROR] TestInMemoryGoogleContractOpen>AbstractContractOpenTest.testOpenFileApplyAsyncRead:341 » IllegalArgument Unknown mandatory key for gs://fake-in-memory-test-b ucket/contract-test/testOpenFileApplyAsyncRead "fs.option.openfile.length" On Wed, Jun 14, 2023 at 5:01 PM Wei-Chiu Chuang <weic...@apache.org> wrote: > The hbase-filesystem tests passed after reverting HADOOP-18596 > <https://issues.apache.org/jira/browse/HADOOP-18596> and HADOOP-18633 > <https://issues.apache.org/jira/browse/HADOOP-18633> from my local tree. > So I think it's a matter of the default behavior being changed. It's not > the end of the world. I think we can address it by adding an incompatible > change flag and a release note. > > On Wed, Jun 14, 2023 at 3:55 PM Wei-Chiu Chuang <weic...@apache.org> > wrote: > >> Cross referenced git history and jira. Changelog needs some update >> >> Not in the release >> >> 1. HDFS-16858 <https://issues.apache.org/jira/browse/HDFS-16858> >> >> >> 1. HADOOP-18532 <https://issues.apache.org/jira/browse/HADOOP-18532> >> 2. >> 1. HDFS-16861 <https://issues.apache.org/jira/browse/HDFS-16861> >> 2. >> 1. HDFS-16866 >> <https://issues.apache.org/jira/browse/HDFS-16866> >> 2. >> 1. HADOOP-18320 >> <https://issues.apache.org/jira/browse/HADOOP-18320> >> 2. >> >> Updated fixed version. Will generate. new Changelog in the next RC. >> >> Was able to build HBase and hbase-filesystem without any code change. >> >> hbase has one unit test failure. This one is reproducible even with >> Hadoop 3.3.5, so maybe a red herring. Local env or something. >> >> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: >> 9.007 s <<< FAILURE! - in >> org.apache.hadoop.hbase.regionserver.TestSyncTimeRangeTracker >> [ERROR] >> org.apache.hadoop.hbase.regionserver.TestSyncTimeRangeTracker.testConcurrentIncludeTimestampCorrectness >> Time elapsed: 3.13 s <<< ERROR! >> java.lang.OutOfMemoryError: Java heap space >> at >> org.apache.hadoop.hbase.regionserver.TestSyncTimeRangeTracker$RandomTestData.<init>(TestSyncTimeRangeTracker.java:91) >> at >> org.apache.hadoop.hbase.regionserver.TestSyncTimeRangeTracker.testConcurrentIncludeTimestampCorrectness(TestSyncTimeRangeTracker.java:156) >> >> hbase-filesystem has three test failures in TestHBOSSContractDistCp, and >> is not reproducible with Hadoop 3.3.5. >> [ERROR] Failures: [ERROR] >> TestHBOSSContractDistCp>AbstractContractDistCpTest.testDistCpUpdateCheckFileSkip:976->Assert.fail:88 >> 10 errors in file of length 10 >> [ERROR] >> TestHBOSSContractDistCp>AbstractContractDistCpTest.testUpdateDeepDirectoryStructureNoChange:270->AbstractContractDistCpTest.assertCounterInRange:290->Assert.assertTrue:41->Assert.fail:88 >> Files Skipped value 0 too below minimum 1 >> [ERROR] >> TestHBOSSContractDistCp>AbstractContractDistCpTest.testUpdateDeepDirectoryStructureToRemote:259->AbstractContractDistCpTest.distCpUpdateDeepDirectoryStructure:334->AbstractContractDistCpTest.assertCounterInRange:294->Assert.assertTrue:41->Assert.fail:88 >> Files Copied value 2 above maximum 1 >> [INFO] >> [ERROR] Tests run: 240, Failures: 3, Errors: 0, Skipped: 58 >> >> >> Ozone >> test in progress. Will report back. >> >> >> On Tue, Jun 13, 2023 at 11:27 PM Wei-Chiu Chuang <weic...@apache.org> >> wrote: >> >>> I am inviting anyone to try and vote on this release candidate. >>> >>> Note: >>> This is built off branch-3.3.6 plus PR#5741 (aws sdk update) and PR#5740 >>> (LICENSE file update) >>> >>> The RC is available at: >>> https://home.apache.org/~weichiu/hadoop-3.3.6-RC0-amd64/ (for amd64) >>> https://home.apache.org/~weichiu/hadoop-3.3.6-RC0-arm64/ (for arm64) >>> >>> Git tag: release-3.3.6-RC0 >>> https://github.com/apache/hadoop/releases/tag/release-3.3.6-RC0 >>> >>> Maven artifacts is built by x86 machine and are staged at >>> https://repository.apache.org/content/repositories/orgapachehadoop-1378/ >>> >>> My public key: >>> https://dist.apache.org/repos/dist/release/hadoop/common/KEYS >>> >>> Changelog: >>> https://home.apache.org/~weichiu/hadoop-3.3.6-RC0-amd64/CHANGELOG.md >>> >>> Release notes: >>> https://home.apache.org/~weichiu/hadoop-3.3.6-RC0-amd64/RELEASENOTES.md >>> >>> This is a relatively small release (by Hadoop standard) containing about >>> 120 commits. >>> Please give it a try, this RC vote will run for 7 days. >>> >>> >>> Feature highlights: >>> >>> SBOM artifacts >>> ---------------------------------------- >>> Starting from this release, Hadoop publishes Software Bill of Materials >>> (SBOM) using >>> CycloneDX Maven plugin. For more information about SBOM, please go to >>> [SBOM](https://cwiki.apache.org/confluence/display/COMDEV/SBOM). >>> >>> HDFS RBF: RDBMS based token storage support >>> ---------------------------------------- >>> HDFS Router-Router Based Federation now supports storing delegation >>> tokens on MySQL, >>> [HADOOP-18535](https://issues.apache.org/jira/browse/HADOOP-18535) >>> which improves token operation through over the original Zookeeper-based >>> implementation. >>> >>> >>> New File System APIs >>> ---------------------------------------- >>> [HADOOP-18671](https://issues.apache.org/jira/browse/HADOOP-18671) >>> moved a number of >>> HDFS-specific APIs to Hadoop Common to make it possible for certain >>> applications that >>> depend on HDFS semantics to run on other Hadoop compatible file systems. >>> >>> In particular, recoverLease() and isFileClosed() are exposed through >>> LeaseRecoverable >>> interface. While setSafeMode() is exposed through SafeMode interface. >>> >>> >>>