Hi All,

Results of voting for Spark 3.3.0 RC5 are:

+1:
Sean Owen (*)
Dongjoon Hyun (*)
Yuming Wang
Yikun Jiang
Martin Grigorov
Thomas Graves (*)
Gengliang Wang
L. C. Hsieh (*)
Cheng Su
Chris Nauroth
Cheng Pan

0:
Hyukjin Kwon (*)

-1:
Jungtaek Lim
Jerry Peng
Prashant Singh
Huaxin Gao

I consider the voting as *failed*. I will prepare RC6 as soon as the issues
mentioned in the thread are solved.

Maxim Gekk

Software Engineer

Databricks, Inc.


On Wed, Jun 8, 2022 at 9:18 PM huaxin gao <huaxin.ga...@gmail.com> wrote:

> I agree with Prashant, -1 from me too because this may break iceberg
> usage.
>
> Thanks,
> Huaxin
>
> On Wed, Jun 8, 2022 at 10:07 AM Prashant Singh <prashant010...@gmail.com>
> wrote:
>
>> -1 from my side as well, found this today.
>>
>> While testing Apache iceberg with 3.3 found this bug where a table with
>> partitions with null values we get a NPE on partition discovery, earlier we
>> use to get `DEFAULT_PARTITION_NAME`
>>
>> Please look into : https://issues.apache.org/jira/browse/SPARK-39417 for
>> more details
>>
>> Regards,
>> Prashant Singh
>>
>> On Wed, Jun 8, 2022 at 10:27 PM Jerry Peng <jerry.boyang.p...@gmail.com>
>> wrote:
>>
>>>
>>>
>>> I agree with Jungtaek,  -1 from me because of the issue of Kafka source
>>> throwing an error with an incorrect error message that was introduced
>>> recently.  This may mislead users and cause unnecessary confusion.
>>>
>>> On Wed, Jun 8, 2022 at 12:04 AM Jungtaek Lim <
>>> kabhwan.opensou...@gmail.com> wrote:
>>>
>>>> Apologize for late participation.
>>>>
>>>> I'm sorry, but -1 (non-binding) from me.
>>>>
>>>> Unfortunately I found a major user-facing issue which hurts UX
>>>> seriously on Kafka data source usage.
>>>>
>>>> In some cases, Kafka data source can throw IllegalStateException for
>>>> the case of failOnDataLoss=true which condition is bound to the state of
>>>> Kafka topic (not Spark's issue). With the recent change of Spark,
>>>> IllegalStateException is now bound to the "internal error", and Spark gives
>>>> incorrect guidance to the end users, telling to end users that Spark has a
>>>> bug and they are encouraged to file a JIRA ticket which is simply wrong.
>>>>
>>>> Previously, Kafka data source provided the error message with the
>>>> context why it failed, and how to workaround it. I feel this is a serious
>>>> regression on UX.
>>>>
>>>> Please look into https://issues.apache.org/jira/browse/SPARK-39412 for
>>>> more details.
>>>>
>>>>
>>>> On Wed, Jun 8, 2022 at 3:40 PM Hyukjin Kwon <gurwls...@gmail.com>
>>>> wrote:
>>>>
>>>>> Okay. Thankfully the binary release is fine per
>>>>> https://github.com/apache/spark/blob/v3.3.0-rc5/dev/create-release/release-build.sh#L268
>>>>> .
>>>>> The source package (and GitHub tag) has 3.3.0.dev0, and the binary
>>>>> package has 3.3.0. Technically this is not a blocker now because PyPI
>>>>> upload will be able to be made correctly.
>>>>> I lowered the priority to critical. I switch my -1 to 0.
>>>>>
>>>>> On Wed, 8 Jun 2022 at 15:17, Hyukjin Kwon <gurwls...@gmail.com> wrote:
>>>>>
>>>>>> Arrrgh  .. I am very sorry that I found this problem late.
>>>>>> RC 5 does not have the correct version of PySpark, see
>>>>>> https://github.com/apache/spark/blob/v3.3.0-rc5/python/pyspark/version.py#L19
>>>>>> I think the release script was broken because the version now has
>>>>>> 'str' type, see
>>>>>> https://github.com/apache/spark/blob/v3.3.0-rc5/dev/create-release/release-tag.sh#L88
>>>>>> I filed a JIRA at https://issues.apache.org/jira/browse/SPARK-39411
>>>>>>
>>>>>> -1 from me
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, 8 Jun 2022 at 13:16, Cheng Pan <pan3...@gmail.com> wrote:
>>>>>>
>>>>>>> +1 (non-binding)
>>>>>>>
>>>>>>> * Verified SPARK-39313 has been address[1]
>>>>>>> * Passed integration test w/ Apache Kyuubi (Incubating)[2]
>>>>>>>
>>>>>>> [1]
>>>>>>> https://github.com/housepower/spark-clickhouse-connector/pull/123
>>>>>>> [2] https://github.com/apache/incubator-kyuubi/pull/2817
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Cheng Pan
>>>>>>>
>>>>>>> On Wed, Jun 8, 2022 at 7:04 AM Chris Nauroth <cnaur...@apache.org>
>>>>>>> wrote:
>>>>>>> >
>>>>>>> > +1 (non-binding)
>>>>>>> >
>>>>>>> > * Verified all checksums.
>>>>>>> > * Verified all signatures.
>>>>>>> > * Built from source, with multiple profiles, to full success, for
>>>>>>> Java 11 and Scala 2.13:
>>>>>>> >     * build/mvn -Phadoop-3 -Phadoop-cloud -Phive-thriftserver
>>>>>>> -Pkubernetes -Pscala-2.13 -Psparkr -Pyarn -DskipTests clean package
>>>>>>> > * Tests passed.
>>>>>>> > * Ran several examples successfully:
>>>>>>> >     * bin/spark-submit --class org.apache.spark.examples.SparkPi
>>>>>>> examples/jars/spark-examples_2.12-3.3.0.jar
>>>>>>> >     * bin/spark-submit --class
>>>>>>> org.apache.spark.examples.sql.hive.SparkHiveExample
>>>>>>> examples/jars/spark-examples_2.12-3.3.0.jar
>>>>>>> >     * bin/spark-submit
>>>>>>> examples/src/main/python/streaming/network_wordcount.py localhost 9999
>>>>>>> > * Tested some of the issues that blocked prior release candidates:
>>>>>>> >     * bin/spark-sql -e 'SELECT (SELECT IF(x, 1, 0)) AS a FROM
>>>>>>> (SELECT true) t(x) UNION SELECT 1 AS a;'
>>>>>>> >     * bin/spark-sql -e "select date '2018-11-17' > 1"
>>>>>>> >     * SPARK-39293 ArrayAggregate fix
>>>>>>> >
>>>>>>> > Chris Nauroth
>>>>>>> >
>>>>>>> >
>>>>>>> > On Tue, Jun 7, 2022 at 1:30 PM Cheng Su <chen...@fb.com.invalid>
>>>>>>> wrote:
>>>>>>> >>
>>>>>>> >> +1 (non-binding). Built and ran some internal test for Spark SQL.
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> Thanks,
>>>>>>> >>
>>>>>>> >> Cheng Su
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> From: L. C. Hsieh <vii...@gmail.com>
>>>>>>> >> Date: Tuesday, June 7, 2022 at 1:23 PM
>>>>>>> >> To: dev <dev@spark.apache.org>
>>>>>>> >> Subject: Re: [VOTE] Release Spark 3.3.0 (RC5)
>>>>>>> >>
>>>>>>> >> +1
>>>>>>> >>
>>>>>>> >> Liang-Chi
>>>>>>> >>
>>>>>>> >> On Tue, Jun 7, 2022 at 1:03 PM Gengliang Wang <ltn...@gmail.com>
>>>>>>> wrote:
>>>>>>> >> >
>>>>>>> >> > +1 (non-binding)
>>>>>>> >> >
>>>>>>> >> > Gengliang
>>>>>>> >> >
>>>>>>> >> > On Tue, Jun 7, 2022 at 12:24 PM Thomas Graves <
>>>>>>> tgraves...@gmail.com> wrote:
>>>>>>> >> >>
>>>>>>> >> >> +1
>>>>>>> >> >>
>>>>>>> >> >> Tom Graves
>>>>>>> >> >>
>>>>>>> >> >> On Sat, Jun 4, 2022 at 9:50 AM Maxim Gekk
>>>>>>> >> >> <maxim.g...@databricks.com.invalid> wrote:
>>>>>>> >> >> >
>>>>>>> >> >> > Please vote on releasing the following candidate as Apache
>>>>>>> Spark version 3.3.0.
>>>>>>> >> >> >
>>>>>>> >> >> > The vote is open until 11:59pm Pacific time June 8th and
>>>>>>> passes if a majority +1 PMC votes are cast, with a minimum of 3 +1 
>>>>>>> votes.
>>>>>>> >> >> >
>>>>>>> >> >> > [ ] +1 Release this package as Apache Spark 3.3.0
>>>>>>> >> >> > [ ] -1 Do not release this package because ...
>>>>>>> >> >> >
>>>>>>> >> >> > To learn more about Apache Spark, please see
>>>>>>> http://spark.apache.org/
>>>>>>> >> >> >
>>>>>>> >> >> > The tag to be voted on is v3.3.0-rc5 (commit
>>>>>>> 7cf29705272ab8e8c70e8885a3664ad8ae3cd5e9):
>>>>>>> >> >> > https://github.com/apache/spark/tree/v3.3.0-rc5
>>>>>>> >> >> >
>>>>>>> >> >> > The release files, including signatures, digests, etc. can
>>>>>>> be found at:
>>>>>>> >> >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc5-bin/
>>>>>>> >> >> >
>>>>>>> >> >> > Signatures used for Spark RCs can be found in this file:
>>>>>>> >> >> > https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>> >> >> >
>>>>>>> >> >> > The staging repository for this release can be found at:
>>>>>>> >> >> >
>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1406
>>>>>>> >> >> >
>>>>>>> >> >> > The documentation corresponding to this release can be found
>>>>>>> at:
>>>>>>> >> >> >
>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc5-docs/
>>>>>>> >> >> >
>>>>>>> >> >> > The list of bug fixes going into 3.3.0 can be found at the
>>>>>>> following URL:
>>>>>>> >> >> >
>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>>> >> >> >
>>>>>>> >> >> > This release is using the release script of the tag
>>>>>>> v3.3.0-rc5.
>>>>>>> >> >> >
>>>>>>> >> >> >
>>>>>>> >> >> > FAQ
>>>>>>> >> >> >
>>>>>>> >> >> > =========================
>>>>>>> >> >> > How can I help test this release?
>>>>>>> >> >> > =========================
>>>>>>> >> >> > If you are a Spark user, you can help us test this release
>>>>>>> by taking
>>>>>>> >> >> > an existing Spark workload and running on this release
>>>>>>> candidate, then
>>>>>>> >> >> > reporting any regressions.
>>>>>>> >> >> >
>>>>>>> >> >> > If you're working in PySpark you can set up a virtual env
>>>>>>> and install
>>>>>>> >> >> > the current RC and see if anything important breaks, in the
>>>>>>> Java/Scala
>>>>>>> >> >> > you can add the staging repository to your projects
>>>>>>> resolvers and test
>>>>>>> >> >> > with the RC (make sure to clean up the artifact cache
>>>>>>> before/after so
>>>>>>> >> >> > you don't end up building with a out of date RC going
>>>>>>> forward).
>>>>>>> >> >> >
>>>>>>> >> >> > ===========================================
>>>>>>> >> >> > What should happen to JIRA tickets still targeting 3.3.0?
>>>>>>> >> >> > ===========================================
>>>>>>> >> >> > The current list of open tickets targeted at 3.3.0 can be
>>>>>>> found at:
>>>>>>> >> >> > https://issues.apache.org/jira/projects/SPARK  and search
>>>>>>> for "Target Version/s" = 3.3.0
>>>>>>> >> >> >
>>>>>>> >> >> > Committers should look at those and triage. Extremely
>>>>>>> important bug
>>>>>>> >> >> > fixes, documentation, and API tweaks that impact
>>>>>>> compatibility should
>>>>>>> >> >> > be worked on immediately. Everything else please retarget to
>>>>>>> an
>>>>>>> >> >> > appropriate release.
>>>>>>> >> >> >
>>>>>>> >> >> > ==================
>>>>>>> >> >> > But my bug isn't fixed?
>>>>>>> >> >> > ==================
>>>>>>> >> >> > In order to make timely releases, we will typically not hold
>>>>>>> the
>>>>>>> >> >> > release unless the bug in question is a regression from the
>>>>>>> previous
>>>>>>> >> >> > release. That being said, if there is something which is a
>>>>>>> regression
>>>>>>> >> >> > that has not been correctly targeted please ping me or a
>>>>>>> committer to
>>>>>>> >> >> > help target the issue.
>>>>>>> >> >> >
>>>>>>> >> >> > Maxim Gekk
>>>>>>> >> >> >
>>>>>>> >> >> > Software Engineer
>>>>>>> >> >> >
>>>>>>> >> >> > Databricks, Inc.
>>>>>>> >> >>
>>>>>>> >> >>
>>>>>>> ---------------------------------------------------------------------
>>>>>>> >> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>> >> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> ---------------------------------------------------------------------
>>>>>>> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>>
>>>>>>> ---------------------------------------------------------------------
>>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>>
>>>>>>>

Reply via email to