This is an automated email from the ASF dual-hosted git repository. stevel pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/hadoop-release-support.git
commit cb8efc3ddb18f7ba9b905c1d839c502667bc018b Author: Steve Loughran <ste...@cloudera.com> AuthorDate: Fri Dec 16 16:46:08 2022 +0000 HADOOP-18470. rc work --- build.xml | 35 ++++++++++++++++++++++++++--------- src/text/email.txt | 4 +++- 2 files changed, 29 insertions(+), 10 deletions(-) diff --git a/build.xml b/build.xml index 6dd0e3e..d6cd783 100644 --- a/build.xml +++ b/build.xml @@ -353,17 +353,24 @@ Message is in file ${message.out} description="build the spark release in spark.dir"> <echo> - Note: this build includes kinesis and ASL artifacts + Note: this build includes kinesis and hadoop cloud artifacts; + builds the full tarball in assembly/target which can be used for + spark-standalone and cluster testing. + + Note, SPARK-41392 highlights some build problems on trunk; + until fixed expect failures to occur on the trunk branch. + the branch-3.3 line MUST work to qualify the build. </echo> <mvn dir="${spark.dir}"> <arg value="-Psnapshots-and-staging"/> - <arg value="-Phadoop-cloud"/> + <arg value="-Phadoop-cloud"/> <!-- cloud module --> <arg value="-Pyarn"/> - <arg value="-Pkinesis-asl"/> + <arg value="-Pkinesis-asl"/> <!-- kinesis --> + <arg value="-Pbigtop-dist"/> <!-- build the distribution --> <arg value="-DskipTests"/> <arg value="-Dmaven.javadoc.skip=true"/> <arg value="-Dhadoop.version=${hadoop.version}"/> - <arg value="-Daws.java.sdk.version=${spark.aws.version}"/> + <arg value="-Daws.java.sdk.version=${spark.aws.version}"/> <!-- kinesis module to use the same aws jar--> <arg value="clean"/> <arg value="install"/> </mvn> @@ -375,19 +382,30 @@ Message is in file ${message.out} depends="init" description="build the cloud examples release"> <echo> - Build the cloud examples modules + Build the cloud examples; + spark MUST have been built against this release first. </echo> - + <mvn dir="${cloud-examples.dir}"> + <arg value="-Psnapshots-and-staging"/> + <arg value="-Dspark-3.4"/> + <arg value="-Dspark.version=${spark.version}"/> + <arg value="-Dhadoop.version=${hadoop.version}"/> + <arg value="clean"/> + <arg value="install"/> + <arg value="-DskipTests"/> + </mvn> </target> <target name="cloud-examples.test" if="cloud-examples.dir" - depends="init" + depends="cloud-examples.build" description="test the cloud examples"> <echo> Test the cloud examples; - cloud.test.configuration.file must point to the xml file with binding info + 1. cloud.test.configuration.file MUST point to the xml file with binding info. + 2. spark MUST have been built against this release first. + 3. Java 11+ only. </echo> <require p="cloud.test.configuration.file" /> <mvn dir="${cloud-examples.dir}"> @@ -396,7 +414,6 @@ Message is in file ${message.out} <arg value="-Dspark.version=${spark.version}"/> <arg value="-Dhadoop.version=${hadoop.version}"/> <arg value="-Dcloud.test.configuration.file=${cloud.test.configuration.file}"/> - <arg value="clean"/> <arg value="test"/> </mvn> </target> diff --git a/src/text/email.txt b/src/text/email.txt index 907e2c8..94b53b8 100644 --- a/src/text/email.txt +++ b/src/text/email.txt @@ -35,11 +35,13 @@ See the release notes for details. Key changes * Big update of dependencies to try and keep those reports of - transitive CVEs under control. + transitive CVEs under control -both genuine and false positive. * Vectored IO API for all FSDataInputStream implementations, with high-performance versions for file:// and s3a:// filesystems. file:// through java native io s3a:// parallel GET requests. +* This release includes arm64 binaries. Please can anyone with + compatible systems validate these. Please try the release and vote. The vote will run for 5 days. --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org