Arrrgh  .. I am very sorry that I found this problem late.
RC 5 does not have the correct version of PySpark, see
https://github.com/apache/spark/blob/v3.3.0-rc5/python/pyspark/version.py#L19
I think the release script was broken because the version now has 'str'
type, see
https://github.com/apache/spark/blob/v3.3.0-rc5/dev/create-release/release-tag.sh#L88
I filed a JIRA at https://issues.apache.org/jira/browse/SPARK-39411

-1 from me



On Wed, 8 Jun 2022 at 13:16, Cheng Pan <pan3...@gmail.com> wrote:

> +1 (non-binding)
>
> * Verified SPARK-39313 has been address[1]
> * Passed integration test w/ Apache Kyuubi (Incubating)[2]
>
> [1] https://github.com/housepower/spark-clickhouse-connector/pull/123
> [2] https://github.com/apache/incubator-kyuubi/pull/2817
>
> Thanks,
> Cheng Pan
>
> On Wed, Jun 8, 2022 at 7:04 AM Chris Nauroth <cnaur...@apache.org> wrote:
> >
> > +1 (non-binding)
> >
> > * Verified all checksums.
> > * Verified all signatures.
> > * Built from source, with multiple profiles, to full success, for Java
> 11 and Scala 2.13:
> >     * build/mvn -Phadoop-3 -Phadoop-cloud -Phive-thriftserver
> -Pkubernetes -Pscala-2.13 -Psparkr -Pyarn -DskipTests clean package
> > * Tests passed.
> > * Ran several examples successfully:
> >     * bin/spark-submit --class org.apache.spark.examples.SparkPi
> examples/jars/spark-examples_2.12-3.3.0.jar
> >     * bin/spark-submit --class
> org.apache.spark.examples.sql.hive.SparkHiveExample
> examples/jars/spark-examples_2.12-3.3.0.jar
> >     * bin/spark-submit
> examples/src/main/python/streaming/network_wordcount.py localhost 9999
> > * Tested some of the issues that blocked prior release candidates:
> >     * bin/spark-sql -e 'SELECT (SELECT IF(x, 1, 0)) AS a FROM (SELECT
> true) t(x) UNION SELECT 1 AS a;'
> >     * bin/spark-sql -e "select date '2018-11-17' > 1"
> >     * SPARK-39293 ArrayAggregate fix
> >
> > Chris Nauroth
> >
> >
> > On Tue, Jun 7, 2022 at 1:30 PM Cheng Su <chen...@fb.com.invalid> wrote:
> >>
> >> +1 (non-binding). Built and ran some internal test for Spark SQL.
> >>
> >>
> >>
> >> Thanks,
> >>
> >> Cheng Su
> >>
> >>
> >>
> >> From: L. C. Hsieh <vii...@gmail.com>
> >> Date: Tuesday, June 7, 2022 at 1:23 PM
> >> To: dev <dev@spark.apache.org>
> >> Subject: Re: [VOTE] Release Spark 3.3.0 (RC5)
> >>
> >> +1
> >>
> >> Liang-Chi
> >>
> >> On Tue, Jun 7, 2022 at 1:03 PM Gengliang Wang <ltn...@gmail.com> wrote:
> >> >
> >> > +1 (non-binding)
> >> >
> >> > Gengliang
> >> >
> >> > On Tue, Jun 7, 2022 at 12:24 PM Thomas Graves <tgraves...@gmail.com>
> wrote:
> >> >>
> >> >> +1
> >> >>
> >> >> Tom Graves
> >> >>
> >> >> On Sat, Jun 4, 2022 at 9:50 AM Maxim Gekk
> >> >> <maxim.g...@databricks.com.invalid> wrote:
> >> >> >
> >> >> > Please vote on releasing the following candidate as Apache Spark
> version 3.3.0.
> >> >> >
> >> >> > The vote is open until 11:59pm Pacific time June 8th and passes if
> a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
> >> >> >
> >> >> > [ ] +1 Release this package as Apache Spark 3.3.0
> >> >> > [ ] -1 Do not release this package because ...
> >> >> >
> >> >> > To learn more about Apache Spark, please see
> http://spark.apache.org/
> >> >> >
> >> >> > The tag to be voted on is v3.3.0-rc5 (commit
> 7cf29705272ab8e8c70e8885a3664ad8ae3cd5e9):
> >> >> > https://github.com/apache/spark/tree/v3.3.0-rc5
> >> >> >
> >> >> > The release files, including signatures, digests, etc. can be
> found at:
> >> >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc5-bin/
> >> >> >
> >> >> > Signatures used for Spark RCs can be found in this file:
> >> >> > https://dist.apache.org/repos/dist/dev/spark/KEYS
> >> >> >
> >> >> > The staging repository for this release can be found at:
> >> >> >
> https://repository.apache.org/content/repositories/orgapachespark-1406
> >> >> >
> >> >> > The documentation corresponding to this release can be found at:
> >> >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc5-docs/
> >> >> >
> >> >> > The list of bug fixes going into 3.3.0 can be found at the
> following URL:
> >> >> > https://issues.apache.org/jira/projects/SPARK/versions/12350369
> >> >> >
> >> >> > This release is using the release script of the tag v3.3.0-rc5.
> >> >> >
> >> >> >
> >> >> > FAQ
> >> >> >
> >> >> > =========================
> >> >> > How can I help test this release?
> >> >> > =========================
> >> >> > If you are a Spark user, you can help us test this release by
> taking
> >> >> > an existing Spark workload and running on this release candidate,
> then
> >> >> > reporting any regressions.
> >> >> >
> >> >> > If you're working in PySpark you can set up a virtual env and
> install
> >> >> > the current RC and see if anything important breaks, in the
> Java/Scala
> >> >> > you can add the staging repository to your projects resolvers and
> test
> >> >> > with the RC (make sure to clean up the artifact cache before/after
> so
> >> >> > you don't end up building with a out of date RC going forward).
> >> >> >
> >> >> > ===========================================
> >> >> > What should happen to JIRA tickets still targeting 3.3.0?
> >> >> > ===========================================
> >> >> > The current list of open tickets targeted at 3.3.0 can be found at:
> >> >> > https://issues.apache.org/jira/projects/SPARK  and search for
> "Target Version/s" = 3.3.0
> >> >> >
> >> >> > Committers should look at those and triage. Extremely important bug
> >> >> > fixes, documentation, and API tweaks that impact compatibility
> should
> >> >> > be worked on immediately. Everything else please retarget to an
> >> >> > appropriate release.
> >> >> >
> >> >> > ==================
> >> >> > But my bug isn't fixed?
> >> >> > ==================
> >> >> > In order to make timely releases, we will typically not hold the
> >> >> > release unless the bug in question is a regression from the
> previous
> >> >> > release. That being said, if there is something which is a
> regression
> >> >> > that has not been correctly targeted please ping me or a committer
> to
> >> >> > help target the issue.
> >> >> >
> >> >> > Maxim Gekk
> >> >> >
> >> >> > Software Engineer
> >> >> >
> >> >> > Databricks, Inc.
> >> >>
> >> >> ---------------------------------------------------------------------
> >> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
> >> >>
> >>
> >> ---------------------------------------------------------------------
> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to