+1 (binding) in Spark SQL, Core and PySpark.

Xiao

2018-02-24 14:49 GMT-08:00 Ricardo Almeida <ricardo.alme...@actnowib.com>:

> +1 (non-binding)
>
> same as previous RC
>
> On 24 February 2018 at 11:10, Hyukjin Kwon <gurwls...@gmail.com> wrote:
>
>> +1
>>
>> 2018-02-24 16:57 GMT+09:00 Bryan Cutler <cutl...@gmail.com>:
>>
>>> +1
>>> Tests passed and additionally ran Arrow related tests and did some perf
>>> checks with python 2.7.14
>>>
>>> On Fri, Feb 23, 2018 at 6:18 PM, Holden Karau <hol...@pigscanfly.ca>
>>> wrote:
>>>
>>>> Note: given the state of Jenkins I'd love to see Bryan Cutler or
>>>> someone with Arrow experience sign off on this release.
>>>>
>>>> On Fri, Feb 23, 2018 at 6:13 PM, Cheng Lian <lian.cs....@gmail.com>
>>>> wrote:
>>>>
>>>>> +1 (binding)
>>>>>
>>>>> Passed all the tests, looks good.
>>>>>
>>>>> Cheng
>>>>>
>>>>> On 2/23/18 15:00, Holden Karau wrote:
>>>>>
>>>>> +1 (binding)
>>>>> PySpark artifacts install in a fresh Py3 virtual env
>>>>>
>>>>> On Feb 23, 2018 7:55 AM, "Denny Lee" <denny.g....@gmail.com> wrote:
>>>>>
>>>>>> +1 (non-binding)
>>>>>>
>>>>>> On Fri, Feb 23, 2018 at 07:08 Josh Goldsborough <
>>>>>> joshgoldsboroughs...@gmail.com> wrote:
>>>>>>
>>>>>>> New to testing out Spark RCs for the community but I was able to run
>>>>>>> some of the basic unit tests without error so for what it's worth, I'm 
>>>>>>> a +1.
>>>>>>>
>>>>>>> On Thu, Feb 22, 2018 at 4:23 PM, Sameer Agarwal <samee...@apache.org
>>>>>>> > wrote:
>>>>>>>
>>>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>>>> version 2.3.0. The vote is open until Tuesday February 27, 2018 at 
>>>>>>>> 8:00:00
>>>>>>>> am UTC and passes if a majority of at least 3 PMC +1 votes are cast.
>>>>>>>>
>>>>>>>>
>>>>>>>> [ ] +1 Release this package as Apache Spark 2.3.0
>>>>>>>>
>>>>>>>> [ ] -1 Do not release this package because ...
>>>>>>>>
>>>>>>>>
>>>>>>>> To learn more about Apache Spark, please see
>>>>>>>> https://spark.apache.org/
>>>>>>>>
>>>>>>>> The tag to be voted on is v2.3.0-rc5:
>>>>>>>> https://github.com/apache/spark/tree/v2.3.0-rc5
>>>>>>>> (992447fb30ee9ebb3cf794f2d06f4d63a2d792db)
>>>>>>>>
>>>>>>>> List of JIRA tickets resolved in this release can be found here:
>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12339551
>>>>>>>>
>>>>>>>> The release files, including signatures, digests, etc. can be found
>>>>>>>> at:
>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v2.3.0-rc5-bin/
>>>>>>>>
>>>>>>>> Release artifacts are signed with the following key:
>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>>>
>>>>>>>> The staging repository for this release can be found at:
>>>>>>>> https://repository.apache.org/content/repositories/orgapache
>>>>>>>> spark-1266/
>>>>>>>>
>>>>>>>> The documentation corresponding to this release can be found at:
>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v2.3.0-rc5-docs
>>>>>>>> /_site/index.html
>>>>>>>>
>>>>>>>>
>>>>>>>> FAQ
>>>>>>>>
>>>>>>>> =======================================
>>>>>>>> What are the unresolved issues targeted for 2.3.0?
>>>>>>>> =======================================
>>>>>>>>
>>>>>>>> Please see https://s.apache.org/oXKi. At the time of writing,
>>>>>>>> there are currently no known release blockers.
>>>>>>>>
>>>>>>>> =========================
>>>>>>>> How can I help test this release?
>>>>>>>> =========================
>>>>>>>>
>>>>>>>> If you are a Spark user, you can help us test this release by
>>>>>>>> taking an existing Spark workload and running on this release 
>>>>>>>> candidate,
>>>>>>>> then reporting any regressions.
>>>>>>>>
>>>>>>>> If you're working in PySpark you can set up a virtual env and
>>>>>>>> install the current RC and see if anything important breaks, in the
>>>>>>>> Java/Scala you can add the staging repository to your projects 
>>>>>>>> resolvers
>>>>>>>> and test with the RC (make sure to clean up the artifact cache 
>>>>>>>> before/after
>>>>>>>> so you don't end up building with a out of date RC going forward).
>>>>>>>>
>>>>>>>> ===========================================
>>>>>>>> What should happen to JIRA tickets still targeting 2.3.0?
>>>>>>>> ===========================================
>>>>>>>>
>>>>>>>> Committers should look at those and triage. Extremely important bug
>>>>>>>> fixes, documentation, and API tweaks that impact compatibility should 
>>>>>>>> be
>>>>>>>> worked on immediately. Everything else please retarget to 2.3.1 or 
>>>>>>>> 2.4.0 as
>>>>>>>> appropriate.
>>>>>>>>
>>>>>>>> ===================
>>>>>>>> Why is my bug not fixed?
>>>>>>>> ===================
>>>>>>>>
>>>>>>>> In order to make timely releases, we will typically not hold the
>>>>>>>> release unless the bug in question is a regression from 2.2.0. That 
>>>>>>>> being
>>>>>>>> said, if there is something which is a regression from 2.2.0 and has 
>>>>>>>> not
>>>>>>>> been correctly targeted please ping me or a committer to help target 
>>>>>>>> the
>>>>>>>> issue (you can see the open issues listed as impacting Spark 2.3.0 at
>>>>>>>> https://s.apache.org/WmoI).
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Twitter: https://twitter.com/holdenkarau
>>>>
>>>
>>>
>>
>

Reply via email to