+1 2017-11-29 8:18 GMT+09:00 Henry Robinson <[email protected]>:
> (My vote is non-binding, of course). > > On 28 November 2017 at 14:53, Henry Robinson <[email protected]> wrote: > >> +1, tests all pass for me on Ubuntu 16.04. >> >> On 28 November 2017 at 10:36, Herman van Hövell tot Westerflier < >> [email protected]> wrote: >> >>> +1 >>> >>> On Tue, Nov 28, 2017 at 7:35 PM, Felix Cheung <[email protected]> >>> wrote: >>> >>>> +1 >>>> >>>> Thanks Sean. Please vote! >>>> >>>> Tested various scenarios with R package. Ubuntu, Debian, Windows >>>> r-devel and release and on r-hub. Verified CRAN checks are clean (only >>>> 1 NOTE!) and no leaked files (.cache removed, /tmp clean) >>>> >>>> >>>> On Sun, Nov 26, 2017 at 11:55 AM Sean Owen <[email protected]> wrote: >>>> >>>>> Yes it downloads recent releases. The test worked for me on a second >>>>> try, so I suspect a bad mirror. If this comes up frequently we can just >>>>> add >>>>> retry logic, as the closer.lua script will return different mirrors each >>>>> time. >>>>> >>>>> The tests all pass for me on the latest Debian, so +1 for this release. >>>>> >>>>> (I committed the change to set -Xss4m for tests consistently, but this >>>>> shouldn't block a release.) >>>>> >>>>> >>>>> On Sat, Nov 25, 2017 at 12:47 PM Felix Cheung <[email protected]> >>>>> wrote: >>>>> >>>>>> Ah sorry digging through the history it looks like this is changed >>>>>> relatively recently and should only download previous releases. >>>>>> >>>>>> Perhaps we are intermittently hitting a mirror that doesn’t have the >>>>>> files? >>>>>> >>>>>> >>>>>> https://github.com/apache/spark/commit/daa838b8886496e64700b >>>>>> 55d1301d348f1d5c9ae >>>>>> >>>>>> >>>>>> On Sat, Nov 25, 2017 at 10:36 AM Felix Cheung <[email protected]> >>>>>> wrote: >>>>>> >>>>>>> Thanks Sean. >>>>>>> >>>>>>> For the second one, it looks like the HiveExternalCatalogVersionsSuite >>>>>>> is >>>>>>> trying to download the release tgz from the official Apache mirror, >>>>>>> which >>>>>>> won’t work unless the release is actually, released? >>>>>>> >>>>>>> val preferredMirror = >>>>>>> Seq("wget", "https://www.apache.org/dyn/closer.lua?preferred=true", >>>>>>> "-q", "-O", "-").!!.trim >>>>>>> val url = s"$preferredMirror/spark/spark >>>>>>> -$version/spark-$version-bin-hadoop2.7.tgz" >>>>>>> >>>>>>> It’s proabbly getting an error page instead. >>>>>>> >>>>>>> >>>>>>> On Sat, Nov 25, 2017 at 10:28 AM Sean Owen <[email protected]> >>>>>>> wrote: >>>>>>> >>>>>>>> I hit the same StackOverflowError as in the previous RC test, but, >>>>>>>> pretty sure this is just because the increased thread stack size JVM >>>>>>>> flag >>>>>>>> isn't applied consistently. This seems to resolve it: >>>>>>>> >>>>>>>> https://github.com/apache/spark/pull/19820 >>>>>>>> >>>>>>>> This wouldn't block release IMHO. >>>>>>>> >>>>>>>> >>>>>>>> I am currently investigating this failure though -- seems like the >>>>>>>> mechanism that downloads Spark tarballs needs fixing, or updating, in >>>>>>>> the >>>>>>>> 2.2 branch? >>>>>>>> >>>>>>>> HiveExternalCatalogVersionsSuite: >>>>>>>> >>>>>>>> gzip: stdin: not in gzip format >>>>>>>> >>>>>>>> tar: Child returned status 1 >>>>>>>> >>>>>>>> tar: Error is not recoverable: exiting now >>>>>>>> >>>>>>>> *** RUN ABORTED *** >>>>>>>> >>>>>>>> java.io.IOException: Cannot run program "./bin/spark-submit" (in >>>>>>>> directory "/tmp/test-spark/spark-2.0.2"): error=2, No such file or >>>>>>>> directory >>>>>>>> >>>>>>>> On Sat, Nov 25, 2017 at 12:34 AM Felix Cheung < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Please vote on releasing the following candidate as Apache Spark >>>>>>>>> version 2.2.1. The vote is open until Friday December 1, 2017 at >>>>>>>>> 8:00:00 am UTC and passes if a majority of at least 3 PMC +1 >>>>>>>>> votes are cast. >>>>>>>>> >>>>>>>>> >>>>>>>>> [ ] +1 Release this package as Apache Spark 2.2.1 >>>>>>>>> >>>>>>>>> [ ] -1 Do not release this package because ... >>>>>>>>> >>>>>>>>> >>>>>>>>> To learn more about Apache Spark, please see >>>>>>>>> https://spark.apache.org/ >>>>>>>>> >>>>>>>>> >>>>>>>>> The tag to be voted on is v2.2.1-rc2 https://github.com/ >>>>>>>>> apache/spark/tree/v2.2.1-rc2 (e30e2698a2193f0bbdcd4edb88471 >>>>>>>>> 0819ab6397c) >>>>>>>>> >>>>>>>>> List of JIRA tickets resolved in this release can be found here >>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12340470 >>>>>>>>> >>>>>>>>> >>>>>>>>> The release files, including signatures, digests, etc. can be >>>>>>>>> found at: >>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2-bin/ >>>>>>>>> >>>>>>>>> Release artifacts are signed with the following key: >>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS >>>>>>>>> >>>>>>>>> The staging repository for this release can be found at: >>>>>>>>> https://repository.apache.org/content/repositories/orgapache >>>>>>>>> spark-1257/ >>>>>>>>> >>>>>>>>> The documentation corresponding to this release can be found at: >>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2 >>>>>>>>> -docs/_site/index.html >>>>>>>>> >>>>>>>>> >>>>>>>>> *FAQ* >>>>>>>>> >>>>>>>>> *How can I help test this release?* >>>>>>>>> >>>>>>>>> If you are a Spark user, you can help us test this release by >>>>>>>>> taking an existing Spark workload and running on this release >>>>>>>>> candidate, >>>>>>>>> then reporting any regressions. >>>>>>>>> >>>>>>>>> If you're working in PySpark you can set up a virtual env and >>>>>>>>> install the current RC and see if anything important breaks, in the >>>>>>>>> Java/Scala you can add the staging repository to your projects >>>>>>>>> resolvers >>>>>>>>> and test with the RC (make sure to clean up the artifact cache >>>>>>>>> before/after >>>>>>>>> so you don't end up building with a out of date RC going forward). >>>>>>>>> >>>>>>>>> *What should happen to JIRA tickets still targeting 2.2.1?* >>>>>>>>> >>>>>>>>> Committers should look at those and triage. Extremely important >>>>>>>>> bug fixes, documentation, and API tweaks that impact compatibility >>>>>>>>> should >>>>>>>>> be worked on immediately. Everything else please retarget to 2.2.2. >>>>>>>>> >>>>>>>>> *But my bug isn't fixed!??!* >>>>>>>>> >>>>>>>>> In order to make timely releases, we will typically not hold the >>>>>>>>> release unless the bug in question is a regression from 2.2.0. That >>>>>>>>> being >>>>>>>>> said if there is something which is a regression form 2.2.0 that has >>>>>>>>> not >>>>>>>>> been correctly targeted please ping a committer to help target the >>>>>>>>> issue >>>>>>>>> (you can see the open issues listed as impacting Spark 2.2.1 / 2.2.2 >>>>>>>>> here >>>>>>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20%3D%20OPEN%20AND%20(affectedVersion%20%3D%202.2.1%20OR%20affectedVersion%20%3D%202.2.2)> >>>>>>>>> . >>>>>>>>> >>>>>>>>> *What are the unresolved issues targeted for 2.2.1 >>>>>>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20in%20(Open%2C%20%22In%20Progress%22%2C%20Reopened)%20AND%20%22Target%20Version%2Fs%22%20%3D%202.2.1>?* >>>>>>>>> >>>>>>>>> At the time of the writing, there is one intermited failure >>>>>>>>> SPARK-20201 <https://issues.apache.org/jira/browse/SPARK-20201> which >>>>>>>>> we are tracking since 2.2.0. >>>>>>>>> >>>>>>>>> >>> >>> >>> >>> >> >
