Thanks for the pointers.

That's the plan...I am going to have default ALS, variable projection at
each ALS step, and then using the constrained first order solver that you
have...

I am hoping ECOS QP will beat all of them...I don't know...some of these
obvious assumptions does not work when the data is large...

About ECOS, we are trying to get it Apache licensed from Professor Boyd /
Alex Domahidi who is the primary author. Let's see how that goes...

I am not sure how easy it will be for projected gradient solver to handle
quadratic constraints...For simple bounds I feel it will work fine...On Mar
31, 2014 10:34 AM, "David Hall" <d...@cs.berkeley.edu> wrote:

> On Mon, Mar 31, 2014 at 10:15 AM, Debasish Das <debasish.da...@gmail.com
> >wrote:
>
> > I added eclipse support in my qp branch:
> >
> > https://github.com/debasish83/breeze/tree/qp
>
>
> Ok, great. Totally fine for this to come in with the QP.
>
>
> >
> >
> > For the QP solver I will look into this solver
> http://www.joptimizer.com/
> >
> > Right now my plan is to use Professor Boyd's ECOS solver which is also
> > designed in the very similar lines but has been tested to solve even cone
> > programs...
> >
> > https://github.com/ifa-ethz/ecos
> >
> > Any idea whether I should add C native code using jniloader as the first
> > version or rewrite using breeze.optimize style and call netlib-java calls
> > for native support (ldl, cholesky etc)...
> >
>
> I personally don't care much. A year ago, I would have much preferred pure
> JVM, and a few months ago I would have just said get it working with
> natives, since presumably that's easier. *However*, ECOS is GPL, and
> jniloader is LGPL. So, you can't require these dependencies in an Apache
> product.
>
> IANAL, but if you want this in Spark, I would suggest you stop looking at
> ECOS immediately, at least until you have something that isn't encumbered
> that works well enough.
>
>
>
> >
> > I still have to think how much cone support we will need...In ALS for
> > example X^TX = I and Y^Y=I are interesting constraints for
> > orthogonality...and they are quadratic constraints...With BFGS and CG, it
> > is difficult to handle quadratic constraints...
> >
>
>
> I wonder how our general projected gradient solver would do? Clearly having
> dedicated QP support is better, but in terms of just getting it working, it
> might be enough.
>
> -- David
>
>
> >
> >
> >
> > On Sun, Mar 30, 2014 at 4:40 PM, David Hall <d...@cs.berkeley.edu>
> wrote:
> >
> > > On Sun, Mar 30, 2014 at 2:01 PM, Debasish Das <
> debasish.da...@gmail.com
> > > >wrote:
> > >
> > > > Hi David,
> > > >
> > > > I have started to experiment with BFGS solvers for Spark GLM over
> large
> > > > scale data...
> > > >
> > > > I am also looking to add a good QP solver in breeze that can be used
> in
> > > > Spark ALS for constraint solves...More details on that soon...
> > > >
> > > > I could not load up breeze 0.7 code onto eclipse...There is a folder
> > > called
> > > > natives in the master but there is no code in that....all the code is
> > in
> > > > src/main/scala...
> > > >
> > > > I added the eclipse plugin:
> > > >
> > > > addSbtPlugin("com.github.mpeltonen" % "sbt-idea" % "1.6.0")
> > > >
> > > > addSbtPlugin("com.typesafe.sbteclipse" % "sbteclipse-plugin" %
> "2.2.0")
> > > >
> > > > But it seems the project is set to use idea...
> > > >
> > > > Could you please explain the dev methodology for breeze ? My idea is
> to
> > > do
> > > > solver work in breeze as that's the right place and get it into Spark
> > > > through Xiangrui's WIP on Sparse data and breeze support...
> > > >
> > >
> > > It would be great to have a QP Solver: I don't know if you know about
> > this
> > > library: http://www.joptimizer.com/
> > >
> > > I'm not quite sure what you mean by dev methodology. If you just mean
> how
> > > to get code into Breeze, just send a PR to scalanlp/breeze. Unit tests
> > are
> > > good for something nontrivial like this. Maybe some basic
> documentation.
> > >
> > >
> > > >
> > > > Thanks.
> > > > Deb
> > > >
> > > >
> > > >
> > > > On Fri, Mar 7, 2014 at 12:46 AM, DB Tsai <dbt...@alpinenow.com>
> wrote:
> > > >
> > > > > Hi Xiangrui,
> > > > >
> > > > > I think it doesn't matter whether we use Fortran/Breeze/RISO for
> > > > > optimizers since optimization only takes << 1% of time. Most of the
> > > > > time is in gradientSum and lossSum parallel computation.
> > > > >
> > > > > Sincerely,
> > > > >
> > > > > DB Tsai
> > > > > Machine Learning Engineer
> > > > > Alpine Data Labs
> > > > > --------------------------------------
> > > > > Web: http://alpinenow.com/
> > > > >
> > > > >
> > > > > On Thu, Mar 6, 2014 at 7:10 PM, Xiangrui Meng <men...@gmail.com>
> > > wrote:
> > > > > > Hi DB,
> > > > > >
> > > > > > Thanks for doing the comparison! What were the running times for
> > > > > > fortran/breeze/riso?
> > > > > >
> > > > > > Best,
> > > > > > Xiangrui
> > > > > >
> > > > > > On Thu, Mar 6, 2014 at 4:21 PM, DB Tsai <dbt...@alpinenow.com>
> > > wrote:
> > > > > >> Hi David,
> > > > > >>
> > > > > >> I can converge to the same result with your breeze LBFGS and
> > Fortran
> > > > > >> implementations now. Probably, I made some mistakes when I tried
> > > > > >> breeze before. I apologize that I claimed it's not stable.
> > > > > >>
> > > > > >> See the test case in BreezeLBFGSSuite.scala
> > > > > >> https://github.com/AlpineNow/spark/tree/dbtsai-breezeLBFGS
> > > > > >>
> > > > > >> This is training multinomial logistic regression against iris
> > > dataset,
> > > > > >> and both optimizers can train the models with 98% training
> > accuracy.
> > > > > >>
> > > > > >> There are two issues to use Breeze in Spark,
> > > > > >>
> > > > > >> 1) When the gradientSum and lossSum are computed distributively
> in
> > > > > >> custom defined DiffFunction which will be passed into your
> > > optimizer,
> > > > > >> Spark will complain LBFGS class is not serializable. In
> > > > > >> BreezeLBFGS.scala, I've to convert RDD to array to make it work
> > > > > >> locally. It should be easy to fix by just having LBFGS to
> > implement
> > > > > >> Serializable.
> > > > > >>
> > > > > >> 2) Breeze computes redundant gradient and loss. See the
> following
> > > log
> > > > > >> from both Fortran and Breeze implementations.
> > > > > >>
> > > > > >> Thanks.
> > > > > >>
> > > > > >> Fortran:
> > > > > >> Iteration -1: loss 1.3862943611198926, diff 1.0
> > > > > >> Iteration 0: loss 1.5846343143210866, diff 0.14307193024217352
> > > > > >> Iteration 1: loss 1.1242501524477688, diff 0.29053004039012126
> > > > > >> Iteration 2: loss 1.0930151243303563, diff 0.027782962952189336
> > > > > >> Iteration 3: loss 1.054036932835569, diff 0.03566113127440601
> > > > > >> Iteration 4: loss 0.9907956302751622, diff 0.05999907649459571
> > > > > >> Iteration 5: loss 0.9184205380342829, diff 0.07304737423337761
> > > > > >> Iteration 6: loss 0.8259870936519937, diff 0.10064381175132982
> > > > > >> Iteration 7: loss 0.6327447552109574, diff 0.23395293458364716
> > > > > >> Iteration 8: loss 0.5534101162436359, diff 0.1253815427665277
> > > > > >> Iteration 9: loss 0.4045020086612566, diff 0.26907321376758075
> > > > > >> Iteration 10: loss 0.3078824990823728, diff 0.23885980452569627
> > > > > >>
> > > > > >> Breeze:
> > > > > >> Iteration -1: loss 1.3862943611198926, diff 1.0
> > > > > >> Mar 6, 2014 3:59:11 PM com.github.fommil.netlib.BLAS <clinit>
> > > > > >> WARNING: Failed to load implementation from:
> > > > > >> com.github.fommil.netlib.NativeSystemBLAS
> > > > > >> Mar 6, 2014 3:59:11 PM com.github.fommil.netlib.BLAS <clinit>
> > > > > >> WARNING: Failed to load implementation from:
> > > > > >> com.github.fommil.netlib.NativeRefBLAS
> > > > > >> Iteration 0: loss 1.3862943611198926, diff 0.0
> > > > > >> Iteration 1: loss 1.5846343143210866, diff 0.14307193024217352
> > > > > >> Iteration 2: loss 1.1242501524477688, diff 0.29053004039012126
> > > > > >> Iteration 3: loss 1.1242501524477688, diff 0.0
> > > > > >> Iteration 4: loss 1.1242501524477688, diff 0.0
> > > > > >> Iteration 5: loss 1.0930151243303563, diff 0.027782962952189336
> > > > > >> Iteration 6: loss 1.0930151243303563, diff 0.0
> > > > > >> Iteration 7: loss 1.0930151243303563, diff 0.0
> > > > > >> Iteration 8: loss 1.054036932835569, diff 0.03566113127440601
> > > > > >> Iteration 9: loss 1.054036932835569, diff 0.0
> > > > > >> Iteration 10: loss 1.054036932835569, diff 0.0
> > > > > >> Iteration 11: loss 0.9907956302751622, diff 0.05999907649459571
> > > > > >> Iteration 12: loss 0.9907956302751622, diff 0.0
> > > > > >> Iteration 13: loss 0.9907956302751622, diff 0.0
> > > > > >> Iteration 14: loss 0.9184205380342829, diff 0.07304737423337761
> > > > > >> Iteration 15: loss 0.9184205380342829, diff 0.0
> > > > > >> Iteration 16: loss 0.9184205380342829, diff 0.0
> > > > > >> Iteration 17: loss 0.8259870936519939, diff 0.1006438117513297
> > > > > >> Iteration 18: loss 0.8259870936519939, diff 0.0
> > > > > >> Iteration 19: loss 0.8259870936519939, diff 0.0
> > > > > >> Iteration 20: loss 0.6327447552109576, diff 0.233952934583647
> > > > > >> Iteration 21: loss 0.6327447552109576, diff 0.0
> > > > > >> Iteration 22: loss 0.6327447552109576, diff 0.0
> > > > > >> Iteration 23: loss 0.5534101162436362, diff 0.12538154276652747
> > > > > >> Iteration 24: loss 0.5534101162436362, diff 0.0
> > > > > >> Iteration 25: loss 0.5534101162436362, diff 0.0
> > > > > >> Iteration 26: loss 0.40450200866125635, diff 0.2690732137675816
> > > > > >> Iteration 27: loss 0.40450200866125635, diff 0.0
> > > > > >> Iteration 28: loss 0.40450200866125635, diff 0.0
> > > > > >> Iteration 29: loss 0.30788249908237314, diff 0.23885980452569502
> > > > > >>
> > > > > >> Sincerely,
> > > > > >>
> > > > > >> DB Tsai
> > > > > >> Machine Learning Engineer
> > > > > >> Alpine Data Labs
> > > > > >> --------------------------------------
> > > > > >> Web: http://alpinenow.com/
> > > > > >>
> > > > > >>
> > > > > >> On Wed, Mar 5, 2014 at 2:00 PM, David Hall <
> d...@cs.berkeley.edu>
> > > > > wrote:
> > > > > >>> On Wed, Mar 5, 2014 at 1:57 PM, DB Tsai <dbt...@alpinenow.com>
> > > > wrote:
> > > > > >>>
> > > > > >>>> Hi David,
> > > > > >>>>
> > > > > >>>> On Tue, Mar 4, 2014 at 8:13 PM, dlwh <david.lw.h...@gmail.com
> >
> > > > wrote:
> > > > > >>>> > I'm happy to help fix any problems. I've verified at points
> > that
> > > > the
> > > > > >>>> > implementation gives the exact same sequence of iterates
> for a
> > > few
> > > > > >>>> different
> > > > > >>>> > functions (with a particular line search) as the c port of
> > > lbfgs.
> > > > > So I'm
> > > > > >>>> a
> > > > > >>>> > little surprised it fails where Fortran succeeds... but
> only a
> > > > > little.
> > > > > >>>> This
> > > > > >>>> > was fixed late last year.
> > > > > >>>> I'm working on a reproducible test case using breeze vs
> fortran
> > > > > >>>> implementation to show the problem I've run into. The test
> will
> > be
> > > > in
> > > > > >>>> one of the test cases in my Spark fork, is it okay for you to
> > > > > >>>> investigate the issue? Or do I need to make it as a standalone
> > > test?
> > > > > >>>>
> > > > > >>>
> > > > > >>>
> > > > > >>> Um, as long as it wouldn't be too hard to pull out.
> > > > > >>>
> > > > > >>>
> > > > > >>>>
> > > > > >>>> Will send you the test later today.
> > > > > >>>>
> > > > > >>>> Thanks.
> > > > > >>>>
> > > > > >>>> Sincerely,
> > > > > >>>>
> > > > > >>>> DB Tsai
> > > > > >>>> Machine Learning Engineer
> > > > > >>>> Alpine Data Labs
> > > > > >>>> --------------------------------------
> > > > > >>>> Web: http://alpinenow.com/
> > > > > >>>>
> > > > >
> > > >
> > >
> >
>

Reply via email to