Technically release don't follow vetos (see
https://www.apache.org/foundation/voting.html ) it's up to the RM if they
get the minimum number of binding +1s (although they are encouraged to
cancel the release if any serious issues are raised).

That being said I'll add my -1 based on the issues reported in this thread.

On Tue, May 10, 2022 at 6:07 PM Sean Owen <sro...@gmail.com> wrote:

> There's a -1 vote here, so I think this RC fails anyway.
>
> On Fri, May 6, 2022 at 10:30 AM Gengliang Wang <ltn...@gmail.com> wrote:
>
>> Hi Maxim,
>>
>> Thanks for the work!
>> There is a bug fix from Bruce merged on branch-3.3 right after the RC1 is
>> cut:
>> SPARK-39093: Dividing interval by integral can result in codegen
>> compilation error
>> <https://github.com/apache/spark/commit/fd998c8a6783c0c8aceed8dcde4017cd479e42c8>
>>
>> So -1 from me. We should have RC2 to include the fix.
>>
>> Thanks
>> Gengliang
>>
>> On Fri, May 6, 2022 at 6:15 PM Maxim Gekk
>> <maxim.g...@databricks.com.invalid> wrote:
>>
>>> Hi Dongjoon,
>>>
>>>  > https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>> > Since RC1 is started, could you move them out from the 3.3.0 milestone?
>>>
>>> I have removed the 3.3.0 label from Fix version(s). Thank you, Dongjoon.
>>>
>>> Maxim Gekk
>>>
>>> Software Engineer
>>>
>>> Databricks, Inc.
>>>
>>>
>>> On Fri, May 6, 2022 at 11:06 AM Dongjoon Hyun <dongjoon.h...@gmail.com>
>>> wrote:
>>>
>>>> Hi, Sean.
>>>> It's interesting. I didn't see those failures from my side.
>>>>
>>>> Hi, Maxim.
>>>> In the following link, there are 17 in-progress and 6 to-do JIRA issues
>>>> which look irrelevant to this RC1 vote.
>>>>
>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>
>>>> Since RC1 is started, could you move them out from the 3.3.0 milestone?
>>>> Otherwise, we cannot distinguish new real blocker issues from those
>>>> obsolete JIRA issues.
>>>>
>>>> Thanks,
>>>> Dongjoon.
>>>>
>>>>
>>>> On Thu, May 5, 2022 at 11:46 AM Adam Binford <adam...@gmail.com> wrote:
>>>>
>>>>> I looked back at the first one (SPARK-37618), it expects/assumes a
>>>>> 0022 umask to correctly test the behavior. I'm not sure how to get that to
>>>>> not fail or be ignored with a more open umask.
>>>>>
>>>>> On Thu, May 5, 2022 at 1:56 PM Sean Owen <sro...@gmail.com> wrote:
>>>>>
>>>>>> I'm seeing test failures; is anyone seeing ones like this? This is
>>>>>> Java 8 / Scala 2.12 / Ubuntu 22.04:
>>>>>>
>>>>>> - SPARK-37618: Sub dirs are group writable when removing from shuffle
>>>>>> service enabled *** FAILED ***
>>>>>>   [OWNER_WRITE, GROUP_READ, GROUP_WRITE, GROUP_EXECUTE, OTHERS_READ,
>>>>>> OWNER_READ, OTHERS_EXECUTE, OWNER_EXECUTE] contained GROUP_WRITE
>>>>>> (DiskBlockManagerSuite.scala:155)
>>>>>>
>>>>>> - Check schemas for expression examples *** FAILED ***
>>>>>>   396 did not equal 398 Expected 396 blocks in result file but got
>>>>>> 398. Try regenerating the result files. 
>>>>>> (ExpressionsSchemaSuite.scala:161)
>>>>>>
>>>>>>  Function 'bloom_filter_agg', Expression class
>>>>>> 'org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregate'
>>>>>> "" did not start with "
>>>>>>       Examples:
>>>>>>   " (ExpressionInfoSuite.scala:142)
>>>>>>
>>>>>> On Thu, May 5, 2022 at 6:01 AM Maxim Gekk
>>>>>> <maxim.g...@databricks.com.invalid> wrote:
>>>>>>
>>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>>>  version 3.3.0.
>>>>>>>
>>>>>>> The vote is open until 11:59pm Pacific time May 10th and passes if
>>>>>>> a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>>>>
>>>>>>> [ ] +1 Release this package as Apache Spark 3.3.0
>>>>>>> [ ] -1 Do not release this package because ...
>>>>>>>
>>>>>>> To learn more about Apache Spark, please see http://spark
>>>>>>> .apache.org/
>>>>>>>
>>>>>>> The tag to be voted on is v3.3.0-rc1 (commit
>>>>>>> 482b7d54b522c4d1e25f3e84eabbc78126f22a3d):
>>>>>>> https://github.com/apache/spark/tree/v3.3.0-rc1
>>>>>>>
>>>>>>> The release files, including signatures, digests, etc. can be found
>>>>>>> at:
>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-bin/
>>>>>>>
>>>>>>> Signatures used for Spark RCs can be found in this file:
>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>>
>>>>>>> The staging repository for this release can be found at:
>>>>>>>
>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1402
>>>>>>>
>>>>>>> The documentation corresponding to this release can be found at:
>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-docs/
>>>>>>>
>>>>>>> The list of bug fixes going into 3.3.0 can be found at the
>>>>>>> following URL:
>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>>>
>>>>>>> This release is using the release script of the tag v3.3.0-rc1.
>>>>>>>
>>>>>>>
>>>>>>> FAQ
>>>>>>>
>>>>>>> =========================
>>>>>>> How can I help test this release?
>>>>>>> =========================
>>>>>>> If you are a Spark user, you can help us test this release by taking
>>>>>>> an existing Spark workload and running on this release candidate,
>>>>>>> then
>>>>>>> reporting any regressions.
>>>>>>>
>>>>>>> If you're working in PySpark you can set up a virtual env and install
>>>>>>> the current RC and see if anything important breaks, in the
>>>>>>> Java/Scala
>>>>>>> you can add the staging repository to your projects resolvers and
>>>>>>> test
>>>>>>> with the RC (make sure to clean up the artifact cache before/after so
>>>>>>> you don't end up building with a out of date RC going forward).
>>>>>>>
>>>>>>> ===========================================
>>>>>>> What should happen to JIRA tickets still targeting 3.3.0?
>>>>>>> ===========================================
>>>>>>> The current list of open tickets targeted at 3.3.0 can be found at:
>>>>>>> https://issues.apache.org/jira/projects/SPARK and search for
>>>>>>> "Target Version/s" = 3.3.0
>>>>>>>
>>>>>>> Committers should look at those and triage. Extremely important bug
>>>>>>> fixes, documentation, and API tweaks that impact compatibility should
>>>>>>> be worked on immediately. Everything else please retarget to an
>>>>>>> appropriate release.
>>>>>>>
>>>>>>> ==================
>>>>>>> But my bug isn't fixed?
>>>>>>> ==================
>>>>>>> In order to make timely releases, we will typically not hold the
>>>>>>> release unless the bug in question is a regression from the previous
>>>>>>> release. That being said, if there is something which is a
>>>>>>> regression
>>>>>>> that has not been correctly targeted please ping me or a committer to
>>>>>>> help target the issue.
>>>>>>>
>>>>>>> Maxim Gekk
>>>>>>>
>>>>>>> Software Engineer
>>>>>>>
>>>>>>> Databricks, Inc.
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> Adam Binford
>>>>>
>>>>

-- 
Twitter: https://twitter.com/holdenkarau
Books (Learning Spark, High Performance Spark, etc.):
https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
YouTube Live Streams: https://www.youtube.com/user/holdenkarau

Reply via email to