+1

2017-11-29 8:18 GMT+09:00 Henry Robinson <he...@apache.org>:

> (My vote is non-binding, of course).
>
> On 28 November 2017 at 14:53, Henry Robinson <he...@apache.org> wrote:
>
>> +1, tests all pass for me on Ubuntu 16.04.
>>
>> On 28 November 2017 at 10:36, Herman van Hövell tot Westerflier <
>> hvanhov...@databricks.com> wrote:
>>
>>> +1
>>>
>>> On Tue, Nov 28, 2017 at 7:35 PM, Felix Cheung <felixche...@apache.org>
>>> wrote:
>>>
>>>> +1
>>>>
>>>> Thanks Sean. Please vote!
>>>>
>>>> Tested various scenarios with R package. Ubuntu, Debian, Windows
>>>> r-devel and release and on r-hub. Verified CRAN checks are clean (only
>>>> 1 NOTE!) and no leaked files (.cache removed, /tmp clean)
>>>>
>>>>
>>>> On Sun, Nov 26, 2017 at 11:55 AM Sean Owen <so...@cloudera.com> wrote:
>>>>
>>>>> Yes it downloads recent releases. The test worked for me on a second
>>>>> try, so I suspect a bad mirror. If this comes up frequently we can just 
>>>>> add
>>>>> retry logic, as the closer.lua script will return different mirrors each
>>>>> time.
>>>>>
>>>>> The tests all pass for me on the latest Debian, so +1 for this release.
>>>>>
>>>>> (I committed the change to set -Xss4m for tests consistently, but this
>>>>> shouldn't block a release.)
>>>>>
>>>>>
>>>>> On Sat, Nov 25, 2017 at 12:47 PM Felix Cheung <felixche...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Ah sorry digging through the history it looks like this is changed
>>>>>> relatively recently and should only download previous releases.
>>>>>>
>>>>>> Perhaps we are intermittently hitting a mirror that doesn’t have the
>>>>>> files?
>>>>>>
>>>>>>
>>>>>> https://github.com/apache/spark/commit/daa838b8886496e64700b
>>>>>> 55d1301d348f1d5c9ae
>>>>>>
>>>>>>
>>>>>> On Sat, Nov 25, 2017 at 10:36 AM Felix Cheung <felixche...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> Thanks Sean.
>>>>>>>
>>>>>>> For the second one, it looks like the  HiveExternalCatalogVersionsSuite 
>>>>>>> is
>>>>>>> trying to download the release tgz from the official Apache mirror, 
>>>>>>> which
>>>>>>> won’t work unless the release is actually, released?
>>>>>>>
>>>>>>> val preferredMirror =
>>>>>>> Seq("wget", "https://www.apache.org/dyn/closer.lua?preferred=true";,
>>>>>>> "-q", "-O", "-").!!.trim
>>>>>>> val url = s"$preferredMirror/spark/spark
>>>>>>> -$version/spark-$version-bin-hadoop2.7.tgz"
>>>>>>>
>>>>>>> It’s proabbly getting an error page instead.
>>>>>>>
>>>>>>>
>>>>>>> On Sat, Nov 25, 2017 at 10:28 AM Sean Owen <so...@cloudera.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I hit the same StackOverflowError as in the previous RC test, but,
>>>>>>>> pretty sure this is just because the increased thread stack size JVM 
>>>>>>>> flag
>>>>>>>> isn't applied consistently. This seems to resolve it:
>>>>>>>>
>>>>>>>> https://github.com/apache/spark/pull/19820
>>>>>>>>
>>>>>>>> This wouldn't block release IMHO.
>>>>>>>>
>>>>>>>>
>>>>>>>> I am currently investigating this failure though -- seems like the
>>>>>>>> mechanism that downloads Spark tarballs needs fixing, or updating, in 
>>>>>>>> the
>>>>>>>> 2.2 branch?
>>>>>>>>
>>>>>>>> HiveExternalCatalogVersionsSuite:
>>>>>>>>
>>>>>>>> gzip: stdin: not in gzip format
>>>>>>>>
>>>>>>>> tar: Child returned status 1
>>>>>>>>
>>>>>>>> tar: Error is not recoverable: exiting now
>>>>>>>>
>>>>>>>> *** RUN ABORTED ***
>>>>>>>>
>>>>>>>>   java.io.IOException: Cannot run program "./bin/spark-submit" (in
>>>>>>>> directory "/tmp/test-spark/spark-2.0.2"): error=2, No such file or
>>>>>>>> directory
>>>>>>>>
>>>>>>>> On Sat, Nov 25, 2017 at 12:34 AM Felix Cheung <
>>>>>>>> felixche...@apache.org> wrote:
>>>>>>>>
>>>>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>>>>> version 2.2.1. The vote is open until Friday December 1, 2017 at
>>>>>>>>> 8:00:00 am UTC and passes if a majority of at least 3 PMC +1
>>>>>>>>> votes are cast.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> [ ] +1 Release this package as Apache Spark 2.2.1
>>>>>>>>>
>>>>>>>>> [ ] -1 Do not release this package because ...
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> To learn more about Apache Spark, please see
>>>>>>>>> https://spark.apache.org/
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> The tag to be voted on is v2.2.1-rc2 https://github.com/
>>>>>>>>> apache/spark/tree/v2.2.1-rc2  (e30e2698a2193f0bbdcd4edb88471
>>>>>>>>> 0819ab6397c)
>>>>>>>>>
>>>>>>>>> List of JIRA tickets resolved in this release can be found here
>>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12340470
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> The release files, including signatures, digests, etc. can be
>>>>>>>>> found at:
>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2-bin/
>>>>>>>>>
>>>>>>>>> Release artifacts are signed with the following key:
>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>>>>
>>>>>>>>> The staging repository for this release can be found at:
>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
>>>>>>>>> spark-1257/
>>>>>>>>>
>>>>>>>>> The documentation corresponding to this release can be found at:
>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2
>>>>>>>>> -docs/_site/index.html
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> *FAQ*
>>>>>>>>>
>>>>>>>>> *How can I help test this release?*
>>>>>>>>>
>>>>>>>>> If you are a Spark user, you can help us test this release by
>>>>>>>>> taking an existing Spark workload and running on this release 
>>>>>>>>> candidate,
>>>>>>>>> then reporting any regressions.
>>>>>>>>>
>>>>>>>>> If you're working in PySpark you can set up a virtual env and
>>>>>>>>> install the current RC and see if anything important breaks, in the
>>>>>>>>> Java/Scala you can add the staging repository to your projects 
>>>>>>>>> resolvers
>>>>>>>>> and test with the RC (make sure to clean up the artifact cache 
>>>>>>>>> before/after
>>>>>>>>> so you don't end up building with a out of date RC going forward).
>>>>>>>>>
>>>>>>>>> *What should happen to JIRA tickets still targeting 2.2.1?*
>>>>>>>>>
>>>>>>>>> Committers should look at those and triage. Extremely important
>>>>>>>>> bug fixes, documentation, and API tweaks that impact compatibility 
>>>>>>>>> should
>>>>>>>>> be worked on immediately. Everything else please retarget to 2.2.2.
>>>>>>>>>
>>>>>>>>> *But my bug isn't fixed!??!*
>>>>>>>>>
>>>>>>>>> In order to make timely releases, we will typically not hold the
>>>>>>>>> release unless the bug in question is a regression from 2.2.0. That 
>>>>>>>>> being
>>>>>>>>> said if there is something which is a regression form 2.2.0 that has 
>>>>>>>>> not
>>>>>>>>> been correctly targeted please ping a committer to help target the 
>>>>>>>>> issue
>>>>>>>>> (you can see the open issues listed as impacting Spark 2.2.1 / 2.2.2
>>>>>>>>> here
>>>>>>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20%3D%20OPEN%20AND%20(affectedVersion%20%3D%202.2.1%20OR%20affectedVersion%20%3D%202.2.2)>
>>>>>>>>> .
>>>>>>>>>
>>>>>>>>> *What are the unresolved issues targeted for 2.2.1
>>>>>>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20in%20(Open%2C%20%22In%20Progress%22%2C%20Reopened)%20AND%20%22Target%20Version%2Fs%22%20%3D%202.2.1>?*
>>>>>>>>>
>>>>>>>>> At the time of the writing, there is one intermited failure
>>>>>>>>> SPARK-20201 <https://issues.apache.org/jira/browse/SPARK-20201> which
>>>>>>>>> we are tracking since 2.2.0.
>>>>>>>>>
>>>>>>>>>
>>>
>>>
>>>
>>>
>>
>

Reply via email to