I should have mentioned this is a synthetic dataset I create using some 
likelihood distributions of the rating values.  I am only experimenting / 
learning.  In practice though, the list of items is likely to be at least in 
the 10’s if not 100’s.  Are even this item numbers to low?

Thanks.

-S

From: Nick Pentreath <nick.pentre...@gmail.com>
Sent: Wednesday, June 26, 2019 9:09 AM
To: user@spark.apache.org
Subject: Re: [EXTERNAL] - Re: Problem with the ML ALS algorithm

If the number of items is indeed 4, then another issue is the rank of the 
factors defaults to 10. Setting the "rank" parameter < 4 will help.

However, if you only have 4 items, then I would propose that using ALS (or any 
recommendation model in fact) is not really necessary. There is not really 
enough information as well as sparsity, to make collaborative filtering useful. 
And you could simply recommend all items a user has not rated and the result 
would be the same essentially.


On Wed, Jun 26, 2019 at 3:03 PM Steve Pruitt 
<bpru...@opentext.com<mailto:bpru...@opentext.com>> wrote:
Number of users is 1055
Number of items is 4
Ratings values are either 120, 20, 0


From: Nick Pentreath <nick.pentre...@gmail.com<mailto:nick.pentre...@gmail.com>>
Sent: Wednesday, June 26, 2019 6:03 AM
To: user@spark.apache.org<mailto:user@spark.apache.org>
Subject: [EXTERNAL] - Re: Problem with the ML ALS algorithm

This means that the matrix that ALS is trying to factor is not positive 
definite. Try increasing regParam (try 0.1, 1.0 for example).

What does the data look like? e.g. number of users, number of items, number of 
ratings, etc?

On Wed, Jun 26, 2019 at 12:06 AM Steve Pruitt 
<bpru...@opentext.com<mailto:bpru...@opentext.com>> wrote:
I get an inexplicable exception when trying to build an ALSModel with the 
implicit set to true.  I can’t find any help online.

Thanks in advance.

My code is:

ALS als = new ALS()
                .setMaxIter(5)
                .setRegParam(0.01)
                .setUserCol("customer")
                .setItemCol("item")
                .setImplicitPrefs(true)
                .setRatingCol("rating");
ALSModel model = als.fit(training);

The exception is:
org.apache.spark.ml.optim.SingularMatrixException: LAPACK.dppsv returned 6 
because A is not positive definite. Is A derived from a singular matrix (e.g. 
collinear column values)?
                at 
org.apache.spark.mllib.linalg.CholeskyDecomposition$.checkReturnValue(CholeskyDecomposition.scala:65)
 ~[spark-mllib_2.11-2.3.1.jar:2.3.1]
                at 
org.apache.spark.mllib.linalg.CholeskyDecomposition$.solve(CholeskyDecomposition.scala:41)
 ~[spark-mllib_2.11-2.3.1.jar:2.3.1]
                at 
org.apache.spark.ml.recommendation.ALS$CholeskySolver.solve(ALS.scala:747) 
~[spark-mllib_2.11-2.3.1.jar:2.3.1]

Reply via email to