Actually, I want to add this for discussion because there were many
concerns about the new Kafka 3.1 client.

https://github.com/apache/spark/pull/36517
Revert "[SPARK-36837][BUILD] Upgrade Kafka to 3.1.0"

Dongjoon.


On Wed, May 11, 2022 at 7:55 AM Holden Karau <hol...@pigscanfly.ca> wrote:

> Do we have everything we want for RC2 targeted to 3.3.0 for tracking?
>
> On Wed, May 11, 2022 at 6:44 AM Maxim Gekk
> <maxim.g...@databricks.com.invalid> wrote:
>
>> Hi All,
>>
>> The vote has failed. I will create RC2 in a couple of days.
>>
>> Maxim Gekk
>>
>> Software Engineer
>>
>> Databricks, Inc.
>>
>>
>> On Wed, May 11, 2022 at 4:23 AM Hyukjin Kwon <gurwls...@gmail.com> wrote:
>>
>>> I expect to see RC2 too. I guess he just sticks to the standard, leaving
>>> the vote open till the end.
>>> It hasn't got enough +1s anyway :-).
>>>
>>> On Wed, 11 May 2022 at 10:17, Holden Karau <hol...@pigscanfly.ca> wrote:
>>>
>>>> Technically release don't follow vetos (see
>>>> https://www.apache.org/foundation/voting.html ) it's up to the RM if
>>>> they get the minimum number of binding +1s (although they are encouraged to
>>>> cancel the release if any serious issues are raised).
>>>>
>>>> That being said I'll add my -1 based on the issues reported in this
>>>> thread.
>>>>
>>>> On Tue, May 10, 2022 at 6:07 PM Sean Owen <sro...@gmail.com> wrote:
>>>>
>>>>> There's a -1 vote here, so I think this RC fails anyway.
>>>>>
>>>>> On Fri, May 6, 2022 at 10:30 AM Gengliang Wang <ltn...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Maxim,
>>>>>>
>>>>>> Thanks for the work!
>>>>>> There is a bug fix from Bruce merged on branch-3.3 right after the
>>>>>> RC1 is cut:
>>>>>> SPARK-39093: Dividing interval by integral can result in codegen
>>>>>> compilation error
>>>>>> <https://github.com/apache/spark/commit/fd998c8a6783c0c8aceed8dcde4017cd479e42c8>
>>>>>>
>>>>>> So -1 from me. We should have RC2 to include the fix.
>>>>>>
>>>>>> Thanks
>>>>>> Gengliang
>>>>>>
>>>>>> On Fri, May 6, 2022 at 6:15 PM Maxim Gekk
>>>>>> <maxim.g...@databricks.com.invalid> wrote:
>>>>>>
>>>>>>> Hi Dongjoon,
>>>>>>>
>>>>>>>  > https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>>> > Since RC1 is started, could you move them out from the 3.3.0
>>>>>>> milestone?
>>>>>>>
>>>>>>> I have removed the 3.3.0 label from Fix version(s). Thank you,
>>>>>>> Dongjoon.
>>>>>>>
>>>>>>> Maxim Gekk
>>>>>>>
>>>>>>> Software Engineer
>>>>>>>
>>>>>>> Databricks, Inc.
>>>>>>>
>>>>>>>
>>>>>>> On Fri, May 6, 2022 at 11:06 AM Dongjoon Hyun <
>>>>>>> dongjoon.h...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi, Sean.
>>>>>>>> It's interesting. I didn't see those failures from my side.
>>>>>>>>
>>>>>>>> Hi, Maxim.
>>>>>>>> In the following link, there are 17 in-progress and 6 to-do JIRA
>>>>>>>> issues which look irrelevant to this RC1 vote.
>>>>>>>>
>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>>>>
>>>>>>>> Since RC1 is started, could you move them out from the 3.3.0
>>>>>>>> milestone?
>>>>>>>> Otherwise, we cannot distinguish new real blocker issues from those
>>>>>>>> obsolete JIRA issues.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Dongjoon.
>>>>>>>>
>>>>>>>>
>>>>>>>> On Thu, May 5, 2022 at 11:46 AM Adam Binford <adam...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> I looked back at the first one (SPARK-37618), it expects/assumes a
>>>>>>>>> 0022 umask to correctly test the behavior. I'm not sure how to get 
>>>>>>>>> that to
>>>>>>>>> not fail or be ignored with a more open umask.
>>>>>>>>>
>>>>>>>>> On Thu, May 5, 2022 at 1:56 PM Sean Owen <sro...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> I'm seeing test failures; is anyone seeing ones like this? This
>>>>>>>>>> is Java 8 / Scala 2.12 / Ubuntu 22.04:
>>>>>>>>>>
>>>>>>>>>> - SPARK-37618: Sub dirs are group writable when removing from
>>>>>>>>>> shuffle service enabled *** FAILED ***
>>>>>>>>>>   [OWNER_WRITE, GROUP_READ, GROUP_WRITE, GROUP_EXECUTE,
>>>>>>>>>> OTHERS_READ, OWNER_READ, OTHERS_EXECUTE, OWNER_EXECUTE] contained
>>>>>>>>>> GROUP_WRITE (DiskBlockManagerSuite.scala:155)
>>>>>>>>>>
>>>>>>>>>> - Check schemas for expression examples *** FAILED ***
>>>>>>>>>>   396 did not equal 398 Expected 396 blocks in result file but
>>>>>>>>>> got 398. Try regenerating the result files.
>>>>>>>>>> (ExpressionsSchemaSuite.scala:161)
>>>>>>>>>>
>>>>>>>>>>  Function 'bloom_filter_agg', Expression class
>>>>>>>>>> 'org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregate'
>>>>>>>>>> "" did not start with "
>>>>>>>>>>       Examples:
>>>>>>>>>>   " (ExpressionInfoSuite.scala:142)
>>>>>>>>>>
>>>>>>>>>> On Thu, May 5, 2022 at 6:01 AM Maxim Gekk
>>>>>>>>>> <maxim.g...@databricks.com.invalid> wrote:
>>>>>>>>>>
>>>>>>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>>>>>>>  version 3.3.0.
>>>>>>>>>>>
>>>>>>>>>>> The vote is open until 11:59pm Pacific time May 10th and passes
>>>>>>>>>>> if a majority +1 PMC votes are cast, with a minimum of 3 +1
>>>>>>>>>>> votes.
>>>>>>>>>>>
>>>>>>>>>>> [ ] +1 Release this package as Apache Spark 3.3.0
>>>>>>>>>>> [ ] -1 Do not release this package because ...
>>>>>>>>>>>
>>>>>>>>>>> To learn more about Apache Spark, please see http://spark
>>>>>>>>>>> .apache.org/
>>>>>>>>>>>
>>>>>>>>>>> The tag to be voted on is v3.3.0-rc1 (commit
>>>>>>>>>>> 482b7d54b522c4d1e25f3e84eabbc78126f22a3d):
>>>>>>>>>>> https://github.com/apache/spark/tree/v3.3.0-rc1
>>>>>>>>>>>
>>>>>>>>>>> The release files, including signatures, digests, etc. can be
>>>>>>>>>>> found at:
>>>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-bin/
>>>>>>>>>>>
>>>>>>>>>>> Signatures used for Spark RCs can be found in this file:
>>>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>>>>>>
>>>>>>>>>>> The staging repository for this release can be found at:
>>>>>>>>>>>
>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1402
>>>>>>>>>>>
>>>>>>>>>>> The documentation corresponding to this release can be found at:
>>>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-docs/
>>>>>>>>>>>
>>>>>>>>>>> The list of bug fixes going into 3.3.0 can be found at the
>>>>>>>>>>> following URL:
>>>>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>>>>>>>
>>>>>>>>>>> This release is using the release script of the tag v3.3.0-rc1.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> FAQ
>>>>>>>>>>>
>>>>>>>>>>> =========================
>>>>>>>>>>> How can I help test this release?
>>>>>>>>>>> =========================
>>>>>>>>>>> If you are a Spark user, you can help us test this release by
>>>>>>>>>>> taking
>>>>>>>>>>> an existing Spark workload and running on this release candidate,
>>>>>>>>>>> then
>>>>>>>>>>> reporting any regressions.
>>>>>>>>>>>
>>>>>>>>>>> If you're working in PySpark you can set up a virtual env and
>>>>>>>>>>> install
>>>>>>>>>>> the current RC and see if anything important breaks, in the
>>>>>>>>>>> Java/Scala
>>>>>>>>>>> you can add the staging repository to your projects resolvers
>>>>>>>>>>> and test
>>>>>>>>>>> with the RC (make sure to clean up the artifact cache
>>>>>>>>>>> before/after so
>>>>>>>>>>> you don't end up building with a out of date RC going forward).
>>>>>>>>>>>
>>>>>>>>>>> ===========================================
>>>>>>>>>>> What should happen to JIRA tickets still targeting 3.3.0?
>>>>>>>>>>> ===========================================
>>>>>>>>>>> The current list of open tickets targeted at 3.3.0 can be found
>>>>>>>>>>> at:
>>>>>>>>>>> https://issues.apache.org/jira/projects/SPARK and search for
>>>>>>>>>>> "Target Version/s" = 3.3.0
>>>>>>>>>>>
>>>>>>>>>>> Committers should look at those and triage. Extremely important
>>>>>>>>>>> bug
>>>>>>>>>>> fixes, documentation, and API tweaks that impact compatibility
>>>>>>>>>>> should
>>>>>>>>>>> be worked on immediately. Everything else please retarget to an
>>>>>>>>>>> appropriate release.
>>>>>>>>>>>
>>>>>>>>>>> ==================
>>>>>>>>>>> But my bug isn't fixed?
>>>>>>>>>>> ==================
>>>>>>>>>>> In order to make timely releases, we will typically not hold the
>>>>>>>>>>> release unless the bug in question is a regression from the
>>>>>>>>>>> previous
>>>>>>>>>>> release. That being said, if there is something which is a
>>>>>>>>>>> regression
>>>>>>>>>>> that has not been correctly targeted please ping me or a
>>>>>>>>>>> committer to
>>>>>>>>>>> help target the issue.
>>>>>>>>>>>
>>>>>>>>>>> Maxim Gekk
>>>>>>>>>>>
>>>>>>>>>>> Software Engineer
>>>>>>>>>>>
>>>>>>>>>>> Databricks, Inc.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Adam Binford
>>>>>>>>>
>>>>>>>>
>>>>
>>>> --
>>>> Twitter: https://twitter.com/holdenkarau
>>>> Books (Learning Spark, High Performance Spark, etc.):
>>>> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
>>>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>>>
>>> --
> Twitter: https://twitter.com/holdenkarau
> Books (Learning Spark, High Performance Spark, etc.):
> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>

Reply via email to