@Ted, I ran the following two commands. mvn -Phadoop-2.4 -Dhadoop.version=2.7.0 -Pyarn -Phive -DskipTests clean package mvn -Phadoop-2.4 -Dhadoop.version=2.7.0 -Pyarn -Phive -DwildcardSuites=org.apache.spark.streaming.StreamingContextSuite test
Using Java version "1.7.0_51", the tests passed normally. On Mon, Jun 29, 2015 at 1:05 PM, Krishna Sankar <[email protected]> wrote: > +1 (non-binding, of course) > > 1. Compiled OSX 10.10 (Yosemite) OK Total time: 13:26 min > mvn clean package -Pyarn -Phadoop-2.6 -DskipTests > 2. Tested pyspark, mllib > 2.1. statistics (min,max,mean,Pearson,Spearman) OK > 2.2. Linear/Ridge/Laso Regression OK > 2.3. Decision Tree, Naive Bayes OK > 2.4. KMeans OK > Center And Scale OK > 2.5. RDD operations OK > State of the Union Texts - MapReduce, Filter,sortByKey (word count) > 2.6. Recommendation (Movielens medium dataset ~1 M ratings) OK > Model evaluation/optimization (rank, numIter, lambda) with > itertools OK > 3. Scala - MLlib > 3.1. statistics (min,max,mean,Pearson,Spearman) OK > 3.2. LinearRegressionWithSGD OK > 3.3. Decision Tree OK > 3.4. KMeans OK > 3.5. Recommendation (Movielens medium dataset ~1 M ratings) OK > 3.6. saveAsParquetFile OK > 3.7. Read and verify the 4.3 save(above) - sqlContext.parquetFile, > registerTempTable, sql OK > 3.8. result = sqlContext.sql("SELECT > OrderDetails.OrderID,ShipCountry,UnitPrice,Qty,Discount FROM Orders INNER > JOIN OrderDetails ON Orders.OrderID = OrderDetails.OrderID") OK > 4.0. Spark SQL from Python OK > 4.1. result = sqlContext.sql("SELECT * from people WHERE State = 'WA'") OK > 5.0. Packages > 5.1. com.databricks.spark.csv - read/write OK > > Cheers > <k/> > > On Tue, Jun 23, 2015 at 10:37 PM, Patrick Wendell <[email protected]> > wrote: > >> Please vote on releasing the following candidate as Apache Spark version >> 1.4.1! >> >> This release fixes a handful of known issues in Spark 1.4.0, listed here: >> http://s.apache.org/spark-1.4.1 >> >> The tag to be voted on is v1.4.1-rc1 (commit 60e08e5): >> https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h= >> 60e08e50751fe3929156de956d62faea79f5b801 >> >> The release files, including signatures, digests, etc. can be found at: >> http://people.apache.org/~pwendell/spark-releases/spark-1.4.1-rc1-bin/ >> >> Release artifacts are signed with the following key: >> https://people.apache.org/keys/committer/pwendell.asc >> >> The staging repository for this release can be found at: >> [published as version: 1.4.1] >> https://repository.apache.org/content/repositories/orgapachespark-1118/ >> [published as version: 1.4.1-rc1] >> https://repository.apache.org/content/repositories/orgapachespark-1119/ >> >> The documentation corresponding to this release can be found at: >> http://people.apache.org/~pwendell/spark-releases/spark-1.4.1-rc1-docs/ >> >> Please vote on releasing this package as Apache Spark 1.4.1! >> >> The vote is open until Saturday, June 27, at 06:32 UTC and passes >> if a majority of at least 3 +1 PMC votes are cast. >> >> [ ] +1 Release this package as Apache Spark 1.4.1 >> [ ] -1 Do not release this package because ... >> >> To learn more about Apache Spark, please see >> http://spark.apache.org/ >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: [email protected] >> For additional commands, e-mail: [email protected] >> >> >
