Okay. Thankfully the binary release is fine per https://github.com/apache/spark/blob/v3.3.0-rc5/dev/create-release/release-build.sh#L268 . The source package (and GitHub tag) has 3.3.0.dev0, and the binary package has 3.3.0. Technically this is not a blocker now because PyPI upload will be able to be made correctly. I lowered the priority to critical. I switch my -1 to 0.
On Wed, 8 Jun 2022 at 15:17, Hyukjin Kwon <gurwls...@gmail.com> wrote: > Arrrgh .. I am very sorry that I found this problem late. > RC 5 does not have the correct version of PySpark, see > https://github.com/apache/spark/blob/v3.3.0-rc5/python/pyspark/version.py#L19 > I think the release script was broken because the version now has 'str' > type, see > https://github.com/apache/spark/blob/v3.3.0-rc5/dev/create-release/release-tag.sh#L88 > I filed a JIRA at https://issues.apache.org/jira/browse/SPARK-39411 > > -1 from me > > > > On Wed, 8 Jun 2022 at 13:16, Cheng Pan <pan3...@gmail.com> wrote: > >> +1 (non-binding) >> >> * Verified SPARK-39313 has been address[1] >> * Passed integration test w/ Apache Kyuubi (Incubating)[2] >> >> [1] https://github.com/housepower/spark-clickhouse-connector/pull/123 >> [2] https://github.com/apache/incubator-kyuubi/pull/2817 >> >> Thanks, >> Cheng Pan >> >> On Wed, Jun 8, 2022 at 7:04 AM Chris Nauroth <cnaur...@apache.org> wrote: >> > >> > +1 (non-binding) >> > >> > * Verified all checksums. >> > * Verified all signatures. >> > * Built from source, with multiple profiles, to full success, for Java >> 11 and Scala 2.13: >> > * build/mvn -Phadoop-3 -Phadoop-cloud -Phive-thriftserver >> -Pkubernetes -Pscala-2.13 -Psparkr -Pyarn -DskipTests clean package >> > * Tests passed. >> > * Ran several examples successfully: >> > * bin/spark-submit --class org.apache.spark.examples.SparkPi >> examples/jars/spark-examples_2.12-3.3.0.jar >> > * bin/spark-submit --class >> org.apache.spark.examples.sql.hive.SparkHiveExample >> examples/jars/spark-examples_2.12-3.3.0.jar >> > * bin/spark-submit >> examples/src/main/python/streaming/network_wordcount.py localhost 9999 >> > * Tested some of the issues that blocked prior release candidates: >> > * bin/spark-sql -e 'SELECT (SELECT IF(x, 1, 0)) AS a FROM (SELECT >> true) t(x) UNION SELECT 1 AS a;' >> > * bin/spark-sql -e "select date '2018-11-17' > 1" >> > * SPARK-39293 ArrayAggregate fix >> > >> > Chris Nauroth >> > >> > >> > On Tue, Jun 7, 2022 at 1:30 PM Cheng Su <chen...@fb.com.invalid> wrote: >> >> >> >> +1 (non-binding). Built and ran some internal test for Spark SQL. >> >> >> >> >> >> >> >> Thanks, >> >> >> >> Cheng Su >> >> >> >> >> >> >> >> From: L. C. Hsieh <vii...@gmail.com> >> >> Date: Tuesday, June 7, 2022 at 1:23 PM >> >> To: dev <dev@spark.apache.org> >> >> Subject: Re: [VOTE] Release Spark 3.3.0 (RC5) >> >> >> >> +1 >> >> >> >> Liang-Chi >> >> >> >> On Tue, Jun 7, 2022 at 1:03 PM Gengliang Wang <ltn...@gmail.com> >> wrote: >> >> > >> >> > +1 (non-binding) >> >> > >> >> > Gengliang >> >> > >> >> > On Tue, Jun 7, 2022 at 12:24 PM Thomas Graves <tgraves...@gmail.com> >> wrote: >> >> >> >> >> >> +1 >> >> >> >> >> >> Tom Graves >> >> >> >> >> >> On Sat, Jun 4, 2022 at 9:50 AM Maxim Gekk >> >> >> <maxim.g...@databricks.com.invalid> wrote: >> >> >> > >> >> >> > Please vote on releasing the following candidate as Apache Spark >> version 3.3.0. >> >> >> > >> >> >> > The vote is open until 11:59pm Pacific time June 8th and passes >> if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes. >> >> >> > >> >> >> > [ ] +1 Release this package as Apache Spark 3.3.0 >> >> >> > [ ] -1 Do not release this package because ... >> >> >> > >> >> >> > To learn more about Apache Spark, please see >> http://spark.apache.org/ >> >> >> > >> >> >> > The tag to be voted on is v3.3.0-rc5 (commit >> 7cf29705272ab8e8c70e8885a3664ad8ae3cd5e9): >> >> >> > https://github.com/apache/spark/tree/v3.3.0-rc5 >> >> >> > >> >> >> > The release files, including signatures, digests, etc. can be >> found at: >> >> >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc5-bin/ >> >> >> > >> >> >> > Signatures used for Spark RCs can be found in this file: >> >> >> > https://dist.apache.org/repos/dist/dev/spark/KEYS >> >> >> > >> >> >> > The staging repository for this release can be found at: >> >> >> > >> https://repository.apache.org/content/repositories/orgapachespark-1406 >> >> >> > >> >> >> > The documentation corresponding to this release can be found at: >> >> >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc5-docs/ >> >> >> > >> >> >> > The list of bug fixes going into 3.3.0 can be found at the >> following URL: >> >> >> > https://issues.apache.org/jira/projects/SPARK/versions/12350369 >> >> >> > >> >> >> > This release is using the release script of the tag v3.3.0-rc5. >> >> >> > >> >> >> > >> >> >> > FAQ >> >> >> > >> >> >> > ========================= >> >> >> > How can I help test this release? >> >> >> > ========================= >> >> >> > If you are a Spark user, you can help us test this release by >> taking >> >> >> > an existing Spark workload and running on this release candidate, >> then >> >> >> > reporting any regressions. >> >> >> > >> >> >> > If you're working in PySpark you can set up a virtual env and >> install >> >> >> > the current RC and see if anything important breaks, in the >> Java/Scala >> >> >> > you can add the staging repository to your projects resolvers and >> test >> >> >> > with the RC (make sure to clean up the artifact cache >> before/after so >> >> >> > you don't end up building with a out of date RC going forward). >> >> >> > >> >> >> > =========================================== >> >> >> > What should happen to JIRA tickets still targeting 3.3.0? >> >> >> > =========================================== >> >> >> > The current list of open tickets targeted at 3.3.0 can be found >> at: >> >> >> > https://issues.apache.org/jira/projects/SPARK and search for >> "Target Version/s" = 3.3.0 >> >> >> > >> >> >> > Committers should look at those and triage. Extremely important >> bug >> >> >> > fixes, documentation, and API tweaks that impact compatibility >> should >> >> >> > be worked on immediately. Everything else please retarget to an >> >> >> > appropriate release. >> >> >> > >> >> >> > ================== >> >> >> > But my bug isn't fixed? >> >> >> > ================== >> >> >> > In order to make timely releases, we will typically not hold the >> >> >> > release unless the bug in question is a regression from the >> previous >> >> >> > release. That being said, if there is something which is a >> regression >> >> >> > that has not been correctly targeted please ping me or a >> committer to >> >> >> > help target the issue. >> >> >> > >> >> >> > Maxim Gekk >> >> >> > >> >> >> > Software Engineer >> >> >> > >> >> >> > Databricks, Inc. >> >> >> >> >> >> >> --------------------------------------------------------------------- >> >> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> >> >> >> >> >> >> --------------------------------------------------------------------- >> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> >> --------------------------------------------------------------------- >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> >>