I think Sean has put it well but just to emphasise the point: 

- optimisation problems are either convex or non-convex
- if you have a convex problem then the fundamental result that everyone should 
know is that if you find a local minimum it is also the global minimum
- if you don't have a convex problem (or you can't prove that it is convex) 
then you cannot be sure the minimum you have found is a global minimum

So if you know something is non-convex (e.g. ALS, k-means, convnets) you 
usually have to trade off something. You could try and find a global minimum 
but that may take an impossibly long-time. The other trade-off is to that you 
accept a local minimum and hope that it is close to the global. Which local 
minimum you get will depend on what your initialisation values are - if you try 
different initialisation values you may get a different local minimum. So a 
common practice is to run several sets of iterations using different initial 
values - often randomly generated - and take the best minimum. And of course 
you know the best one by evaluating the objective function and choosing the one 
that gives the lowest value.

Robin
On 24 Jun 2013, at 20:44, Michael Kazekin <kazm...@hotmail.com> wrote:

> Thanks a lot! 
> Do you know by any chance what are the underlying reasons for including such 
> mandatory random seed initialization?
> Do you see any sense in providing another option, such as filling them with 
> zeroes in order to ensure the consistency and repeatability? (for example we 
> might want to track and compare the generated recommendation lists for 
> different parameters, such as the number of features or number of iterations 
> etc.)
> M.
> 
> 
>> Date: Mon, 24 Jun 2013 19:51:44 +0200
>> Subject: Re: Consistent repeatable results for distributed ALS-WR recommender
>> From: s...@apache.org
>> To: user@mahout.apache.org
>> 
>> The matrices of the factorization are initalized randomly. If you fix the
>> random seed (would require modification of the code) you should get exactly
>> the same results.
>> Am 24.06.2013 13:49 schrieb "Michael Kazekin" <kazm...@hotmail.com>:
>> 
>>> Hi!
>>> Should I assume that under same dataset and same parameters for factorizer
>>> and recommender I will get the same results for any specific user?
>>> My current understanding that theoretically ALS-WR algorithm could
>>> guarantee this, but I was wondering could be there any numeric method
>>> issues and/or implementation-specific concerns.
>>> Would appreciate any highlight on this issue.
>>> Mike.
>>> 
>>> 
>>> 
>                                         

Reply via email to