+1 (non-binding) RC2 is tested on CentOS, too.
Bests, Dongjoon. On Tue, Nov 28, 2017 at 4:35 PM, Hyukjin Kwon <gurwls...@gmail.com> wrote: > +1 > > 2017-11-29 8:18 GMT+09:00 Henry Robinson <he...@apache.org>: > >> (My vote is non-binding, of course). >> >> On 28 November 2017 at 14:53, Henry Robinson <he...@apache.org> wrote: >> >>> +1, tests all pass for me on Ubuntu 16.04. >>> >>> On 28 November 2017 at 10:36, Herman van Hövell tot Westerflier < >>> hvanhov...@databricks.com> wrote: >>> >>>> +1 >>>> >>>> On Tue, Nov 28, 2017 at 7:35 PM, Felix Cheung <felixche...@apache.org> >>>> wrote: >>>> >>>>> +1 >>>>> >>>>> Thanks Sean. Please vote! >>>>> >>>>> Tested various scenarios with R package. Ubuntu, Debian, Windows >>>>> r-devel and release and on r-hub. Verified CRAN checks are clean >>>>> (only 1 NOTE!) and no leaked files (.cache removed, /tmp clean) >>>>> >>>>> >>>>> On Sun, Nov 26, 2017 at 11:55 AM Sean Owen <so...@cloudera.com> wrote: >>>>> >>>>>> Yes it downloads recent releases. The test worked for me on a second >>>>>> try, so I suspect a bad mirror. If this comes up frequently we can just >>>>>> add >>>>>> retry logic, as the closer.lua script will return different mirrors each >>>>>> time. >>>>>> >>>>>> The tests all pass for me on the latest Debian, so +1 for this >>>>>> release. >>>>>> >>>>>> (I committed the change to set -Xss4m for tests consistently, but >>>>>> this shouldn't block a release.) >>>>>> >>>>>> >>>>>> On Sat, Nov 25, 2017 at 12:47 PM Felix Cheung <felixche...@apache.org> >>>>>> wrote: >>>>>> >>>>>>> Ah sorry digging through the history it looks like this is changed >>>>>>> relatively recently and should only download previous releases. >>>>>>> >>>>>>> Perhaps we are intermittently hitting a mirror that doesn’t have the >>>>>>> files? >>>>>>> >>>>>>> >>>>>>> https://github.com/apache/spark/commit/daa838b8886496e64700b >>>>>>> 55d1301d348f1d5c9ae >>>>>>> >>>>>>> >>>>>>> On Sat, Nov 25, 2017 at 10:36 AM Felix Cheung < >>>>>>> felixche...@apache.org> wrote: >>>>>>> >>>>>>>> Thanks Sean. >>>>>>>> >>>>>>>> For the second one, it looks like the >>>>>>>> HiveExternalCatalogVersionsSuite is >>>>>>>> trying to download the release tgz from the official Apache mirror, >>>>>>>> which >>>>>>>> won’t work unless the release is actually, released? >>>>>>>> >>>>>>>> val preferredMirror = >>>>>>>> Seq("wget", "https://www.apache.org/dyn/closer.lua?preferred=true", >>>>>>>> "-q", "-O", "-").!!.trim >>>>>>>> val url = s"$preferredMirror/spark/spark >>>>>>>> -$version/spark-$version-bin-hadoop2.7.tgz" >>>>>>>> >>>>>>>> It’s proabbly getting an error page instead. >>>>>>>> >>>>>>>> >>>>>>>> On Sat, Nov 25, 2017 at 10:28 AM Sean Owen <so...@cloudera.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> I hit the same StackOverflowError as in the previous RC test, but, >>>>>>>>> pretty sure this is just because the increased thread stack size JVM >>>>>>>>> flag >>>>>>>>> isn't applied consistently. This seems to resolve it: >>>>>>>>> >>>>>>>>> https://github.com/apache/spark/pull/19820 >>>>>>>>> >>>>>>>>> This wouldn't block release IMHO. >>>>>>>>> >>>>>>>>> >>>>>>>>> I am currently investigating this failure though -- seems like the >>>>>>>>> mechanism that downloads Spark tarballs needs fixing, or updating, in >>>>>>>>> the >>>>>>>>> 2.2 branch? >>>>>>>>> >>>>>>>>> HiveExternalCatalogVersionsSuite: >>>>>>>>> >>>>>>>>> gzip: stdin: not in gzip format >>>>>>>>> >>>>>>>>> tar: Child returned status 1 >>>>>>>>> >>>>>>>>> tar: Error is not recoverable: exiting now >>>>>>>>> >>>>>>>>> *** RUN ABORTED *** >>>>>>>>> >>>>>>>>> java.io.IOException: Cannot run program "./bin/spark-submit" >>>>>>>>> (in directory "/tmp/test-spark/spark-2.0.2"): error=2, No such >>>>>>>>> file or directory >>>>>>>>> >>>>>>>>> On Sat, Nov 25, 2017 at 12:34 AM Felix Cheung < >>>>>>>>> felixche...@apache.org> wrote: >>>>>>>>> >>>>>>>>>> Please vote on releasing the following candidate as Apache Spark >>>>>>>>>> version 2.2.1. The vote is open until Friday December 1, 2017 at >>>>>>>>>> 8:00:00 am UTC and passes if a majority of at least 3 PMC +1 >>>>>>>>>> votes are cast. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> [ ] +1 Release this package as Apache Spark 2.2.1 >>>>>>>>>> >>>>>>>>>> [ ] -1 Do not release this package because ... >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> To learn more about Apache Spark, please see >>>>>>>>>> https://spark.apache.org/ >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> The tag to be voted on is v2.2.1-rc2 https://github.com/ >>>>>>>>>> apache/spark/tree/v2.2.1-rc2 (e30e2698a2193f0bbdcd4edb88471 >>>>>>>>>> 0819ab6397c) >>>>>>>>>> >>>>>>>>>> List of JIRA tickets resolved in this release can be found here >>>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12340470 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> The release files, including signatures, digests, etc. can be >>>>>>>>>> found at: >>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2-bin/ >>>>>>>>>> >>>>>>>>>> Release artifacts are signed with the following key: >>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS >>>>>>>>>> >>>>>>>>>> The staging repository for this release can be found at: >>>>>>>>>> https://repository.apache.org/content/repositories/orgapache >>>>>>>>>> spark-1257/ >>>>>>>>>> >>>>>>>>>> The documentation corresponding to this release can be found at: >>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2 >>>>>>>>>> -docs/_site/index.html >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> *FAQ* >>>>>>>>>> >>>>>>>>>> *How can I help test this release?* >>>>>>>>>> >>>>>>>>>> If you are a Spark user, you can help us test this release by >>>>>>>>>> taking an existing Spark workload and running on this release >>>>>>>>>> candidate, >>>>>>>>>> then reporting any regressions. >>>>>>>>>> >>>>>>>>>> If you're working in PySpark you can set up a virtual env and >>>>>>>>>> install the current RC and see if anything important breaks, in the >>>>>>>>>> Java/Scala you can add the staging repository to your projects >>>>>>>>>> resolvers >>>>>>>>>> and test with the RC (make sure to clean up the artifact cache >>>>>>>>>> before/after >>>>>>>>>> so you don't end up building with a out of date RC going forward). >>>>>>>>>> >>>>>>>>>> *What should happen to JIRA tickets still targeting 2.2.1?* >>>>>>>>>> >>>>>>>>>> Committers should look at those and triage. Extremely important >>>>>>>>>> bug fixes, documentation, and API tweaks that impact compatibility >>>>>>>>>> should >>>>>>>>>> be worked on immediately. Everything else please retarget to 2.2.2. >>>>>>>>>> >>>>>>>>>> *But my bug isn't fixed!??!* >>>>>>>>>> >>>>>>>>>> In order to make timely releases, we will typically not hold the >>>>>>>>>> release unless the bug in question is a regression from 2.2.0. That >>>>>>>>>> being >>>>>>>>>> said if there is something which is a regression form 2.2.0 that has >>>>>>>>>> not >>>>>>>>>> been correctly targeted please ping a committer to help target the >>>>>>>>>> issue >>>>>>>>>> (you can see the open issues listed as impacting Spark 2.2.1 / 2.2.2 >>>>>>>>>> here >>>>>>>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20%3D%20OPEN%20AND%20(affectedVersion%20%3D%202.2.1%20OR%20affectedVersion%20%3D%202.2.2)> >>>>>>>>>> . >>>>>>>>>> >>>>>>>>>> *What are the unresolved issues targeted for 2.2.1 >>>>>>>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20in%20(Open%2C%20%22In%20Progress%22%2C%20Reopened)%20AND%20%22Target%20Version%2Fs%22%20%3D%202.2.1>?* >>>>>>>>>> >>>>>>>>>> At the time of the writing, there is one intermited failure >>>>>>>>>> SPARK-20201 <https://issues.apache.org/jira/browse/SPARK-20201> which >>>>>>>>>> we are tracking since 2.2.0. >>>>>>>>>> >>>>>>>>>> >>>> >>>> >>>> >>>> >>> >> >