Hey Sean,

Thanks for pointing this out.  Looks like a bad test where we should be
doing Set comparison instead of Array.

Michael

On Thu, Nov 13, 2014 at 2:05 AM, Sean Owen <so...@cloudera.com> wrote:

> LICENSE and NOTICE are fine. Signature and checksum is fine. I
> unzipped and built the plain source distribution, which built.
>
> However I am seeing a consistent test failure with "mvn -DskipTests
> clean package; mvn test". In the Hive module:
>
> - SET commands semantics for a HiveContext *** FAILED ***
>   Expected Array("spark.sql.key.usedfortestonly=test.val.0",
>
> "spark.sql.key.usedfortestonlyspark.sql.key.usedfortestonly=test.val.0test.val.0"),
> but got
> Array("spark.sql.key.usedfortestonlyspark.sql.key.usedfortestonly=test.val.0test.val.0",
> "spark.sql.key.usedfortestonly=test.val.0") (HiveQuerySuite.scala:544)
>
> Anyone else seeing this?
>
>
> On Thu, Nov 13, 2014 at 8:18 AM, Krishna Sankar <ksanka...@gmail.com>
> wrote:
> > +1
> > 1. Compiled OSX 10.10 (Yosemite) mvn -Pyarn -Phadoop-2.4
> > -Dhadoop.version=2.4.0 -DskipTests clean package 10:49 min
> > 2. Tested pyspark, mlib
> > 2.1. statistics OK
> > 2.2. Linear/Ridge/Laso Regression OK
> > 2.3. Decision Tree, Naive Bayes OK
> > 2.4. KMeans OK
> > 2.5. rdd operations OK
> > 2.6. recommendation OK
> > 2.7. Good work ! In 1.1.0, there was an error and my program used to hang
> > (over memory allocation) consistently running validation using itertools,
> > compute optimum rank, lambda,numofiterations/rmse; data - movielens
> medium
> > dataset (1 million records) . It works well in 1.1.1 !
> > Cheers
> > <k/>
> > P.S: Missed Reply all, first time
> >
> > On Wed, Nov 12, 2014 at 8:35 PM, Andrew Or <and...@databricks.com>
> wrote:
> >
> >> I will start the vote with a +1
> >>
> >> 2014-11-12 20:34 GMT-08:00 Andrew Or <and...@databricks.com>:
> >>
> >> > Please vote on releasing the following candidate as Apache Spark
> version
> >> 1
> >> > .1.1.
> >> >
> >> > This release fixes a number of bugs in Spark 1.1.0. Some of the
> notable
> >> > ones are
> >> > - [SPARK-3426] Sort-based shuffle compression settings are
> incompatible
> >> > - [SPARK-3948] Stream corruption issues in sort-based shuffle
> >> > - [SPARK-4107] Incorrect handling of Channel.read() led to data
> >> truncation
> >> > The full list is at http://s.apache.org/z9h and in the CHANGES.txt
> >> > attached.
> >> >
> >> > The tag to be voted on is v1.1.1-rc1 (commit 72a4fdbe):
> >> > http://s.apache.org/cZC
> >> >
> >> > The release files, including signatures, digests, etc can be found at:
> >> > http://people.apache.org/~andrewor14/spark-1.1.1-rc1/
> >> >
> >> > Release artifacts are signed with the following key:
> >> > https://people.apache.org/keys/committer/andrewor14.asc
> >> >
> >> > The staging repository for this release can be found at:
> >> >
> https://repository.apache.org/content/repositories/orgapachespark-1034/
> >> >
> >> > The documentation corresponding to this release can be found at:
> >> > http://people.apache.org/~andrewor14/spark-1.1.1-rc1-docs/
> >> >
> >> > Please vote on releasing this package as Apache Spark 1.1.1!
> >> >
> >> > The vote is open until Sunday, November 16, at 04:30 UTC and passes if
> >> > a majority of at least 3 +1 PMC votes are cast.
> >> > [ ] +1 Release this package as Apache Spark 1.1.1
> >> > [ ] -1 Do not release this package because ...
> >> >
> >> > To learn more about Apache Spark, please see
> >> > http://spark.apache.org/
> >> >
> >> > Cheers,
> >> > Andrew
> >> >
> >>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> For additional commands, e-mail: dev-h...@spark.apache.org
>
>

Reply via email to