+1 (non-binding)

RC2 is tested on CentOS, too.

Bests,
Dongjoon.

On Tue, Nov 28, 2017 at 4:35 PM, Hyukjin Kwon <gurwls...@gmail.com> wrote:

> +1
>
> 2017-11-29 8:18 GMT+09:00 Henry Robinson <he...@apache.org>:
>
>> (My vote is non-binding, of course).
>>
>> On 28 November 2017 at 14:53, Henry Robinson <he...@apache.org> wrote:
>>
>>> +1, tests all pass for me on Ubuntu 16.04.
>>>
>>> On 28 November 2017 at 10:36, Herman van Hövell tot Westerflier <
>>> hvanhov...@databricks.com> wrote:
>>>
>>>> +1
>>>>
>>>> On Tue, Nov 28, 2017 at 7:35 PM, Felix Cheung <felixche...@apache.org>
>>>> wrote:
>>>>
>>>>> +1
>>>>>
>>>>> Thanks Sean. Please vote!
>>>>>
>>>>> Tested various scenarios with R package. Ubuntu, Debian, Windows
>>>>> r-devel and release and on r-hub. Verified CRAN checks are clean
>>>>> (only 1 NOTE!) and no leaked files (.cache removed, /tmp clean)
>>>>>
>>>>>
>>>>> On Sun, Nov 26, 2017 at 11:55 AM Sean Owen <so...@cloudera.com> wrote:
>>>>>
>>>>>> Yes it downloads recent releases. The test worked for me on a second
>>>>>> try, so I suspect a bad mirror. If this comes up frequently we can just 
>>>>>> add
>>>>>> retry logic, as the closer.lua script will return different mirrors each
>>>>>> time.
>>>>>>
>>>>>> The tests all pass for me on the latest Debian, so +1 for this
>>>>>> release.
>>>>>>
>>>>>> (I committed the change to set -Xss4m for tests consistently, but
>>>>>> this shouldn't block a release.)
>>>>>>
>>>>>>
>>>>>> On Sat, Nov 25, 2017 at 12:47 PM Felix Cheung <felixche...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> Ah sorry digging through the history it looks like this is changed
>>>>>>> relatively recently and should only download previous releases.
>>>>>>>
>>>>>>> Perhaps we are intermittently hitting a mirror that doesn’t have the
>>>>>>> files?
>>>>>>>
>>>>>>>
>>>>>>> https://github.com/apache/spark/commit/daa838b8886496e64700b
>>>>>>> 55d1301d348f1d5c9ae
>>>>>>>
>>>>>>>
>>>>>>> On Sat, Nov 25, 2017 at 10:36 AM Felix Cheung <
>>>>>>> felixche...@apache.org> wrote:
>>>>>>>
>>>>>>>> Thanks Sean.
>>>>>>>>
>>>>>>>> For the second one, it looks like the  
>>>>>>>> HiveExternalCatalogVersionsSuite is
>>>>>>>> trying to download the release tgz from the official Apache mirror, 
>>>>>>>> which
>>>>>>>> won’t work unless the release is actually, released?
>>>>>>>>
>>>>>>>> val preferredMirror =
>>>>>>>> Seq("wget", "https://www.apache.org/dyn/closer.lua?preferred=true";,
>>>>>>>> "-q", "-O", "-").!!.trim
>>>>>>>> val url = s"$preferredMirror/spark/spark
>>>>>>>> -$version/spark-$version-bin-hadoop2.7.tgz"
>>>>>>>>
>>>>>>>> It’s proabbly getting an error page instead.
>>>>>>>>
>>>>>>>>
>>>>>>>> On Sat, Nov 25, 2017 at 10:28 AM Sean Owen <so...@cloudera.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> I hit the same StackOverflowError as in the previous RC test, but,
>>>>>>>>> pretty sure this is just because the increased thread stack size JVM 
>>>>>>>>> flag
>>>>>>>>> isn't applied consistently. This seems to resolve it:
>>>>>>>>>
>>>>>>>>> https://github.com/apache/spark/pull/19820
>>>>>>>>>
>>>>>>>>> This wouldn't block release IMHO.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> I am currently investigating this failure though -- seems like the
>>>>>>>>> mechanism that downloads Spark tarballs needs fixing, or updating, in 
>>>>>>>>> the
>>>>>>>>> 2.2 branch?
>>>>>>>>>
>>>>>>>>> HiveExternalCatalogVersionsSuite:
>>>>>>>>>
>>>>>>>>> gzip: stdin: not in gzip format
>>>>>>>>>
>>>>>>>>> tar: Child returned status 1
>>>>>>>>>
>>>>>>>>> tar: Error is not recoverable: exiting now
>>>>>>>>>
>>>>>>>>> *** RUN ABORTED ***
>>>>>>>>>
>>>>>>>>>   java.io.IOException: Cannot run program "./bin/spark-submit"
>>>>>>>>> (in directory "/tmp/test-spark/spark-2.0.2"): error=2, No such
>>>>>>>>> file or directory
>>>>>>>>>
>>>>>>>>> On Sat, Nov 25, 2017 at 12:34 AM Felix Cheung <
>>>>>>>>> felixche...@apache.org> wrote:
>>>>>>>>>
>>>>>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>>>>>> version 2.2.1. The vote is open until Friday December 1, 2017 at
>>>>>>>>>> 8:00:00 am UTC and passes if a majority of at least 3 PMC +1
>>>>>>>>>> votes are cast.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> [ ] +1 Release this package as Apache Spark 2.2.1
>>>>>>>>>>
>>>>>>>>>> [ ] -1 Do not release this package because ...
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> To learn more about Apache Spark, please see
>>>>>>>>>> https://spark.apache.org/
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> The tag to be voted on is v2.2.1-rc2 https://github.com/
>>>>>>>>>> apache/spark/tree/v2.2.1-rc2  (e30e2698a2193f0bbdcd4edb88471
>>>>>>>>>> 0819ab6397c)
>>>>>>>>>>
>>>>>>>>>> List of JIRA tickets resolved in this release can be found here
>>>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12340470
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> The release files, including signatures, digests, etc. can be
>>>>>>>>>> found at:
>>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2-bin/
>>>>>>>>>>
>>>>>>>>>> Release artifacts are signed with the following key:
>>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>>>>>
>>>>>>>>>> The staging repository for this release can be found at:
>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
>>>>>>>>>> spark-1257/
>>>>>>>>>>
>>>>>>>>>> The documentation corresponding to this release can be found at:
>>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2
>>>>>>>>>> -docs/_site/index.html
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> *FAQ*
>>>>>>>>>>
>>>>>>>>>> *How can I help test this release?*
>>>>>>>>>>
>>>>>>>>>> If you are a Spark user, you can help us test this release by
>>>>>>>>>> taking an existing Spark workload and running on this release 
>>>>>>>>>> candidate,
>>>>>>>>>> then reporting any regressions.
>>>>>>>>>>
>>>>>>>>>> If you're working in PySpark you can set up a virtual env and
>>>>>>>>>> install the current RC and see if anything important breaks, in the
>>>>>>>>>> Java/Scala you can add the staging repository to your projects 
>>>>>>>>>> resolvers
>>>>>>>>>> and test with the RC (make sure to clean up the artifact cache 
>>>>>>>>>> before/after
>>>>>>>>>> so you don't end up building with a out of date RC going forward).
>>>>>>>>>>
>>>>>>>>>> *What should happen to JIRA tickets still targeting 2.2.1?*
>>>>>>>>>>
>>>>>>>>>> Committers should look at those and triage. Extremely important
>>>>>>>>>> bug fixes, documentation, and API tweaks that impact compatibility 
>>>>>>>>>> should
>>>>>>>>>> be worked on immediately. Everything else please retarget to 2.2.2.
>>>>>>>>>>
>>>>>>>>>> *But my bug isn't fixed!??!*
>>>>>>>>>>
>>>>>>>>>> In order to make timely releases, we will typically not hold the
>>>>>>>>>> release unless the bug in question is a regression from 2.2.0. That 
>>>>>>>>>> being
>>>>>>>>>> said if there is something which is a regression form 2.2.0 that has 
>>>>>>>>>> not
>>>>>>>>>> been correctly targeted please ping a committer to help target the 
>>>>>>>>>> issue
>>>>>>>>>> (you can see the open issues listed as impacting Spark 2.2.1 / 2.2.2
>>>>>>>>>> here
>>>>>>>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20%3D%20OPEN%20AND%20(affectedVersion%20%3D%202.2.1%20OR%20affectedVersion%20%3D%202.2.2)>
>>>>>>>>>> .
>>>>>>>>>>
>>>>>>>>>> *What are the unresolved issues targeted for 2.2.1
>>>>>>>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20in%20(Open%2C%20%22In%20Progress%22%2C%20Reopened)%20AND%20%22Target%20Version%2Fs%22%20%3D%202.2.1>?*
>>>>>>>>>>
>>>>>>>>>> At the time of the writing, there is one intermited failure
>>>>>>>>>> SPARK-20201 <https://issues.apache.org/jira/browse/SPARK-20201> which
>>>>>>>>>> we are tracking since 2.2.0.
>>>>>>>>>>
>>>>>>>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to