+1 Thanks Sean. Please vote!
Tested various scenarios with R package. Ubuntu, Debian, Windows r-devel and release and on r-hub. Verified CRAN checks are clean (only 1 NOTE!) and no leaked files (.cache removed, /tmp clean) On Sun, Nov 26, 2017 at 11:55 AM Sean Owen <so...@cloudera.com> wrote: > Yes it downloads recent releases. The test worked for me on a second try, > so I suspect a bad mirror. If this comes up frequently we can just add > retry logic, as the closer.lua script will return different mirrors each > time. > > The tests all pass for me on the latest Debian, so +1 for this release. > > (I committed the change to set -Xss4m for tests consistently, but this > shouldn't block a release.) > > > On Sat, Nov 25, 2017 at 12:47 PM Felix Cheung <felixche...@apache.org> > wrote: > >> Ah sorry digging through the history it looks like this is changed >> relatively recently and should only download previous releases. >> >> Perhaps we are intermittently hitting a mirror that doesn’t have the >> files? >> >> >> >> https://github.com/apache/spark/commit/daa838b8886496e64700b55d1301d348f1d5c9ae >> >> >> On Sat, Nov 25, 2017 at 10:36 AM Felix Cheung <felixche...@apache.org> >> wrote: >> >>> Thanks Sean. >>> >>> For the second one, it looks like the >>> HiveExternalCatalogVersionsSuite is trying to download the release tgz >>> from the official Apache mirror, which won’t work unless the release is >>> actually, released? >>> >>> val preferredMirror = >>> Seq("wget", "https://www.apache.org/dyn/closer.lua?preferred=true", "-q", >>> "-O", "-").!!.trim >>> val url = s" >>> $preferredMirror/spark/spark-$version/spark-$version-bin-hadoop2.7.tgz" >>> >>> It’s proabbly getting an error page instead. >>> >>> >>> On Sat, Nov 25, 2017 at 10:28 AM Sean Owen <so...@cloudera.com> wrote: >>> >>>> I hit the same StackOverflowError as in the previous RC test, but, >>>> pretty sure this is just because the increased thread stack size JVM flag >>>> isn't applied consistently. This seems to resolve it: >>>> >>>> https://github.com/apache/spark/pull/19820 >>>> >>>> This wouldn't block release IMHO. >>>> >>>> >>>> I am currently investigating this failure though -- seems like the >>>> mechanism that downloads Spark tarballs needs fixing, or updating, in the >>>> 2.2 branch? >>>> >>>> HiveExternalCatalogVersionsSuite: >>>> >>>> gzip: stdin: not in gzip format >>>> >>>> tar: Child returned status 1 >>>> >>>> tar: Error is not recoverable: exiting now >>>> >>>> *** RUN ABORTED *** >>>> >>>> java.io.IOException: Cannot run program "./bin/spark-submit" (in >>>> directory "/tmp/test-spark/spark-2.0.2"): error=2, No such file or >>>> directory >>>> >>>> On Sat, Nov 25, 2017 at 12:34 AM Felix Cheung <felixche...@apache.org> >>>> wrote: >>>> >>>>> Please vote on releasing the following candidate as Apache Spark >>>>> version 2.2.1. The vote is open until Friday December 1, 2017 at >>>>> 8:00:00 am UTC and passes if a majority of at least 3 PMC +1 votes >>>>> are cast. >>>>> >>>>> >>>>> [ ] +1 Release this package as Apache Spark 2.2.1 >>>>> >>>>> [ ] -1 Do not release this package because ... >>>>> >>>>> >>>>> To learn more about Apache Spark, please see https://spark.apache.org/ >>>>> >>>>> >>>>> The tag to be voted on is v2.2.1-rc2 >>>>> https://github.com/apache/spark/tree/v2.2.1-rc2 ( >>>>> e30e2698a2193f0bbdcd4edb884710819ab6397c) >>>>> >>>>> List of JIRA tickets resolved in this release can be found here >>>>> https://issues.apache.org/jira/projects/SPARK/versions/12340470 >>>>> >>>>> >>>>> The release files, including signatures, digests, etc. can be found at: >>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2-bin/ >>>>> >>>>> Release artifacts are signed with the following key: >>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS >>>>> >>>>> The staging repository for this release can be found at: >>>>> https://repository.apache.org/content/repositories/orgapachespark-1257/ >>>>> >>>>> The documentation corresponding to this release can be found at: >>>>> >>>>> https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2-docs/_site/index.html >>>>> >>>>> >>>>> *FAQ* >>>>> >>>>> *How can I help test this release?* >>>>> >>>>> If you are a Spark user, you can help us test this release by taking >>>>> an existing Spark workload and running on this release candidate, then >>>>> reporting any regressions. >>>>> >>>>> If you're working in PySpark you can set up a virtual env and install >>>>> the current RC and see if anything important breaks, in the Java/Scala you >>>>> can add the staging repository to your projects resolvers and test with >>>>> the >>>>> RC (make sure to clean up the artifact cache before/after so you don't end >>>>> up building with a out of date RC going forward). >>>>> >>>>> *What should happen to JIRA tickets still targeting 2.2.1?* >>>>> >>>>> Committers should look at those and triage. Extremely important bug >>>>> fixes, documentation, and API tweaks that impact compatibility should be >>>>> worked on immediately. Everything else please retarget to 2.2.2. >>>>> >>>>> *But my bug isn't fixed!??!* >>>>> >>>>> In order to make timely releases, we will typically not hold the >>>>> release unless the bug in question is a regression from 2.2.0. That being >>>>> said if there is something which is a regression form 2.2.0 that has not >>>>> been correctly targeted please ping a committer to help target the issue >>>>> (you can see the open issues listed as impacting Spark 2.2.1 / 2.2.2 >>>>> here >>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20%3D%20OPEN%20AND%20(affectedVersion%20%3D%202.2.1%20OR%20affectedVersion%20%3D%202.2.2)> >>>>> . >>>>> >>>>> *What are the unresolved issues targeted for 2.2.1 >>>>> <https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20in%20(Open%2C%20%22In%20Progress%22%2C%20Reopened)%20AND%20%22Target%20Version%2Fs%22%20%3D%202.2.1>?* >>>>> >>>>> At the time of the writing, there is one intermited failure >>>>> SPARK-20201 <https://issues.apache.org/jira/browse/SPARK-20201> which >>>>> we are tracking since 2.2.0. >>>>> >>>>>