+1 (non-binding) Verified on OSX 10.10.2, built from source, spark-shell / spark-submit jobs ran various simple Spark / Scala queries ran various SparkSQL queries (including HiveContext) ran ThriftServer service and connected via beeline ran SparkSVD
On Mon Dec 01 2014 at 11:09:26 PM Patrick Wendell <[email protected]> wrote: > Hey All, > > Just an update. Josh, Andrew, and others are working to reproduce > SPARK-4498 and fix it. Other than that issue no serious regressions > have been reported so far. If we are able to get a fix in for that > soon, we'll likely cut another RC with the patch. > > Continued testing of RC1 is definitely appreciated! > > I'll leave this vote open to allow folks to continue posting comments. > It's fine to still give "+1" from your own testing... i.e. you can > assume at this point SPARK-4498 will be fixed before releasing. > > - Patrick > > On Mon, Dec 1, 2014 at 3:30 PM, Matei Zaharia <[email protected]> > wrote: > > +0.9 from me. Tested it on Mac and Windows (someone has to do it) and > while things work, I noticed a few recent scripts don't have Windows > equivalents, namely https://issues.apache.org/jira/browse/SPARK-4683 and > https://issues.apache.org/jira/browse/SPARK-4684. The first one at least > would be good to fix if we do another RC. Not blocking the release but > useful to fix in docs is https://issues.apache.org/jira/browse/SPARK-4685. > > > > Matei > > > > > >> On Dec 1, 2014, at 11:18 AM, Josh Rosen <[email protected]> wrote: > >> > >> Hi everyone, > >> > >> There's an open bug report related to Spark standalone which could be a > potential release-blocker (pending investigation / a bug fix): > https://issues.apache.org/jira/browse/SPARK-4498. This issue seems > non-deterministc and only affects long-running Spark standalone > deployments, so it may be hard to reproduce. I'm going to work on a patch > to add additional logging in order to help with debugging. > >> > >> I just wanted to give an early head's up about this issue and to get > more eyes on it in case anyone else has run into it or wants to help with > debugging. > >> > >> - Josh > >> > >> On November 28, 2014 at 9:18:09 PM, Patrick Wendell ([email protected]) > wrote: > >> > >> Please vote on releasing the following candidate as Apache Spark > version 1.2.0! > >> > >> The tag to be voted on is v1.2.0-rc1 (commit 1056e9ec1): > >> https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h= > 1056e9ec13203d0c51564265e94d77a054498fdb > >> > >> The release files, including signatures, digests, etc. can be found at: > >> http://people.apache.org/~pwendell/spark-1.2.0-rc1/ > >> > >> Release artifacts are signed with the following key: > >> https://people.apache.org/keys/committer/pwendell.asc > >> > >> The staging repository for this release can be found at: > >> https://repository.apache.org/content/repositories/orgapachespark-1048/ > >> > >> The documentation corresponding to this release can be found at: > >> http://people.apache.org/~pwendell/spark-1.2.0-rc1-docs/ > >> > >> Please vote on releasing this package as Apache Spark 1.2.0! > >> > >> The vote is open until Tuesday, December 02, at 05:15 UTC and passes > >> if a majority of at least 3 +1 PMC votes are cast. > >> > >> [ ] +1 Release this package as Apache Spark 1.1.0 > >> [ ] -1 Do not release this package because ... > >> > >> To learn more about Apache Spark, please see > >> http://spark.apache.org/ > >> > >> == What justifies a -1 vote for this release? == > >> This vote is happening very late into the QA period compared with > >> previous votes, so -1 votes should only occur for significant > >> regressions from 1.0.2. Bugs already present in 1.1.X, minor > >> regressions, or bugs related to new features will not block this > >> release. > >> > >> == What default changes should I be aware of? == > >> 1. The default value of "spark.shuffle.blockTransferService" has been > >> changed to "netty" > >> --> Old behavior can be restored by switching to "nio" > >> > >> 2. The default value of "spark.shuffle.manager" has been changed to > "sort". > >> --> Old behavior can be restored by setting "spark.shuffle.manager" to > "hash". > >> > >> == Other notes == > >> Because this vote is occurring over a weekend, I will likely extend > >> the vote if this RC survives until the end of the vote period. > >> > >> - Patrick > >> > >> --------------------------------------------------------------------- > >> To unsubscribe, e-mail: [email protected] > >> For additional commands, e-mail: [email protected] > >> > > > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: [email protected] > > For additional commands, e-mail: [email protected] > > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: [email protected] > For additional commands, e-mail: [email protected] > >
