Thanks Sean, that makes sense. I'll look into the source and see if I
can find learn more.

Another question. I understand how the recommendations are created.
I'd like to wrap this all up as a web service, but I'm not sure I
understand how one would go about doing that? How would one app, fetch
recomendations for a user? Does my app need access to the HDFS file
system?

Thanks again.

On Fri, Aug 3, 2012 at 5:40 PM, Sean Owen <sro...@gmail.com> wrote:
> Is it reasonable to use 1.5GB of heap for recs? sure -- assuming you can
> allow the JVM to use, say, 2GB or more of heap total.
>
> There are more choices in Mahout for non-distributed recs. The primary
> distributed version is an item-similarity-based approach but you can choose
> from several similarity metrics. There is also a matrix-factorization-based
> approach in there.
>
> You can make the distributed version do anything you want, depends on how
> much code you want to write. You'd have to replace the similarity
> computation bits with your own logic, yes.
>
> Sean
>
> On Fri, Aug 3, 2012 at 5:16 PM, Matt Mitchell <goodie...@gmail.com> wrote:
>
>> Hi,
>>
>> I have a pretty neat, non-distributed recommender running here. In
>> doing some math on new user growth rate, I thought it might be wise
>> for me to turn to a distributed approach now, rather than later. Just
>> to make sure I'm on the right track, I calculated 1.5 GB for in-memory
>> prefs. Is that pushing it for a single machine recommender?
>>
>> Am I right in thinking that the distributed approach is a single
>> algorithm, unlike the many possible choices for non-distributed? Is it
>> possible to inject content-aware logic within a distributed
>> recommender? Would I inject that into the map/reduce phase, when the
>> recommendations are generated? Curious to know if there are any
>> examples out there?
>>
>> Thanks!
>>

Reply via email to