It is the LR over car-data at https://github.com/xsankar/cloaked-ironman.
1.2.0 gives Mean Squared Error = 40.8130551358
1.3.0 gives Mean Squared Error = 105.857603953

I will verify it one more time tomorrow.

Cheers
<k/>

On Tue, Mar 3, 2015 at 11:28 PM, Xiangrui Meng <men...@gmail.com> wrote:

> On Tue, Mar 3, 2015 at 11:15 PM, Krishna Sankar <ksanka...@gmail.com>
> wrote:
> > +1 (non-binding, of course)
> >
> > 1. Compiled OSX 10.10 (Yosemite) OK Total time: 13:53 min
> >      mvn clean package -Pyarn -Dyarn.version=2.6.0 -Phadoop-2.4
> > -Dhadoop.version=2.6.0 -Phive -DskipTests -Dscala-2.11
> > 2. Tested pyspark, mlib - running as well as compare results with 1.1.x &
> > 1.2.x
> > 2.1. statistics (min,max,mean,Pearson,Spearman) OK
> > 2.2. Linear/Ridge/Laso Regression OK
> > But MSE has increased from 40.81 to 105.86. Has some refactoring happened
> > on SGD/Linear Models ? Or do we have some extra parameters ? or change of
> > defaults ?
>
> Could you share the code you used? I don't remember any changes in
> linear regression. Thanks! -Xiangrui
>
> > 2.3. Decision Tree, Naive Bayes OK
> > 2.4. KMeans OK
> >        Center And Scale OK
> >        WSSSE has come down slightly
> > 2.5. rdd operations OK
> >       State of the Union Texts - MapReduce, Filter,sortByKey (word count)
> > 2.6. Recommendation (Movielens medium dataset ~1 M ratings) OK
> >        Model evaluation/optimization (rank, numIter, lmbda) with
> itertools
> > OK
> > 3. Scala - MLlib
> > 3.1. statistics (min,max,mean,Pearson,Spearman) OK
> > 3.2. LinearRegressionWIthSGD OK
> > 3.3. Decision Tree OK
> > 3.4. KMeans OK
> > 3.5. Recommendation (Movielens medium dataset ~1 M ratings) OK
> > 4.0. SQL from Python
> > 4.1. result = sqlContext.sql("SELECT * from Employees WHERE State =
> 'WA'")
> > OK
> >
> > Cheers
> > <k/>
> >
> > On Tue, Mar 3, 2015 at 8:19 PM, Patrick Wendell <pwend...@gmail.com>
> wrote:
> >
> >> Please vote on releasing the following candidate as Apache Spark version
> >> 1.3.0!
> >>
> >> The tag to be voted on is v1.3.0-rc2 (commit 3af2687):
> >>
> >>
> https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=3af26870e5163438868c4eb2df88380a533bb232
> >>
> >> The release files, including signatures, digests, etc. can be found at:
> >> http://people.apache.org/~pwendell/spark-1.3.0-rc2/
> >>
> >> Release artifacts are signed with the following key:
> >> https://people.apache.org/keys/committer/pwendell.asc
> >>
> >> Staging repositories for this release can be found at:
> >> https://repository.apache.org/content/repositories/orgapachespark-1074/
> >> (published with version '1.3.0')
> >> https://repository.apache.org/content/repositories/orgapachespark-1075/
> >> (published with version '1.3.0-rc2')
> >>
> >> The documentation corresponding to this release can be found at:
> >> http://people.apache.org/~pwendell/spark-1.3.0-rc2-docs/
> >>
> >> Please vote on releasing this package as Apache Spark 1.3.0!
> >>
> >> The vote is open until Saturday, March 07, at 04:17 UTC and passes if
> >> a majority of at least 3 +1 PMC votes are cast.
> >>
> >> [ ] +1 Release this package as Apache Spark 1.3.0
> >> [ ] -1 Do not release this package because ...
> >>
> >> To learn more about Apache Spark, please see
> >> http://spark.apache.org/
> >>
> >> == How does this compare to RC1 ==
> >> This patch includes a variety of bug fixes found in RC1.
> >>
> >> == How can I help test this release? ==
> >> If you are a Spark user, you can help us test this release by
> >> taking a Spark 1.2 workload and running on this release candidate,
> >> then reporting any regressions.
> >>
> >> If you are happy with this release based on your own testing, give a +1
> >> vote.
> >>
> >> == What justifies a -1 vote for this release? ==
> >> This vote is happening towards the end of the 1.3 QA period,
> >> so -1 votes should only occur for significant regressions from 1.2.1.
> >> Bugs already present in 1.2.X, minor regressions, or bugs related
> >> to new features will not block this release.
> >>
> >> ---------------------------------------------------------------------
> >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> >> For additional commands, e-mail: dev-h...@spark.apache.org
> >>
> >>
>

Reply via email to