Cross referenced git history and jira. Changelog needs some update Not in the release
1. HDFS-16858 <https://issues.apache.org/jira/browse/HDFS-16858> 1. HADOOP-18532 <https://issues.apache.org/jira/browse/HADOOP-18532> 2. 1. HDFS-16861 <https://issues.apache.org/jira/browse/HDFS-16861> 2. 1. HDFS-16866 <https://issues.apache.org/jira/browse/HDFS-16866> 2. 1. HADOOP-18320 <https://issues.apache.org/jira/browse/HADOOP-18320> 2. Updated fixed version. Will generate. new Changelog in the next RC. Was able to build HBase and hbase-filesystem without any code change. hbase has one unit test failure. This one is reproducible even with Hadoop 3.3.5, so maybe a red herring. Local env or something. [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 9.007 s <<< FAILURE! - in org.apache.hadoop.hbase.regionserver.TestSyncTimeRangeTracker [ERROR] org.apache.hadoop.hbase.regionserver.TestSyncTimeRangeTracker.testConcurrentIncludeTimestampCorrectness Time elapsed: 3.13 s <<< ERROR! java.lang.OutOfMemoryError: Java heap space at org.apache.hadoop.hbase.regionserver.TestSyncTimeRangeTracker$RandomTestData.<init>(TestSyncTimeRangeTracker.java:91) at org.apache.hadoop.hbase.regionserver.TestSyncTimeRangeTracker.testConcurrentIncludeTimestampCorrectness(TestSyncTimeRangeTracker.java:156) hbase-filesystem has three test failures in TestHBOSSContractDistCp, and is not reproducible with Hadoop 3.3.5. [ERROR] Failures: [ERROR] TestHBOSSContractDistCp>AbstractContractDistCpTest.testDistCpUpdateCheckFileSkip:976->Assert.fail:88 10 errors in file of length 10 [ERROR] TestHBOSSContractDistCp>AbstractContractDistCpTest.testUpdateDeepDirectoryStructureNoChange:270->AbstractContractDistCpTest.assertCounterInRange:290->Assert.assertTrue:41->Assert.fail:88 Files Skipped value 0 too below minimum 1 [ERROR] TestHBOSSContractDistCp>AbstractContractDistCpTest.testUpdateDeepDirectoryStructureToRemote:259->AbstractContractDistCpTest.distCpUpdateDeepDirectoryStructure:334->AbstractContractDistCpTest.assertCounterInRange:294->Assert.assertTrue:41->Assert.fail:88 Files Copied value 2 above maximum 1 [INFO] [ERROR] Tests run: 240, Failures: 3, Errors: 0, Skipped: 58 Ozone test in progress. Will report back. On Tue, Jun 13, 2023 at 11:27 PM Wei-Chiu Chuang <weic...@apache.org> wrote: > I am inviting anyone to try and vote on this release candidate. > > Note: > This is built off branch-3.3.6 plus PR#5741 (aws sdk update) and PR#5740 > (LICENSE file update) > > The RC is available at: > https://home.apache.org/~weichiu/hadoop-3.3.6-RC0-amd64/ (for amd64) > https://home.apache.org/~weichiu/hadoop-3.3.6-RC0-arm64/ (for arm64) > > Git tag: release-3.3.6-RC0 > https://github.com/apache/hadoop/releases/tag/release-3.3.6-RC0 > > Maven artifacts is built by x86 machine and are staged at > https://repository.apache.org/content/repositories/orgapachehadoop-1378/ > > My public key: > https://dist.apache.org/repos/dist/release/hadoop/common/KEYS > > Changelog: > https://home.apache.org/~weichiu/hadoop-3.3.6-RC0-amd64/CHANGELOG.md > > Release notes: > https://home.apache.org/~weichiu/hadoop-3.3.6-RC0-amd64/RELEASENOTES.md > > This is a relatively small release (by Hadoop standard) containing about > 120 commits. > Please give it a try, this RC vote will run for 7 days. > > > Feature highlights: > > SBOM artifacts > ---------------------------------------- > Starting from this release, Hadoop publishes Software Bill of Materials > (SBOM) using > CycloneDX Maven plugin. For more information about SBOM, please go to > [SBOM](https://cwiki.apache.org/confluence/display/COMDEV/SBOM). > > HDFS RBF: RDBMS based token storage support > ---------------------------------------- > HDFS Router-Router Based Federation now supports storing delegation tokens > on MySQL, > [HADOOP-18535](https://issues.apache.org/jira/browse/HADOOP-18535) > which improves token operation through over the original Zookeeper-based > implementation. > > > New File System APIs > ---------------------------------------- > [HADOOP-18671](https://issues.apache.org/jira/browse/HADOOP-18671) moved > a number of > HDFS-specific APIs to Hadoop Common to make it possible for certain > applications that > depend on HDFS semantics to run on other Hadoop compatible file systems. > > In particular, recoverLease() and isFileClosed() are exposed through > LeaseRecoverable > interface. While setSafeMode() is exposed through SafeMode interface. > > >