The code you cite is not Hadoop-related, so no you can't run it on
Hadoop. You say you want to use Hadoop without using Mappers and
Reducers -- this isn't possible.

The only thing you can do is to take a look at the pseudo-distributed
recommender, which just runs several non-distributed recommenders
using Hadoop. It's not really parallelizing the computation. But may
be what you need.

On Sun, Dec 25, 2011 at 11:49 PM, Jinyuan Zhou <[email protected]> wrote:
> Merry Chrisams Sean,
> Thanks for such fast response on this day.
> I think I misunderstood the following description on mahout's home page.
> For the "level of integration" I mentioned,  Here is what I am looking
> for.  I don't need to know Hadoop when I build recommender via mahout API.
> But when I run the recommender, I can ask mahout to run it via a Hadoop
> Cluster. But as you said hadoop based and non hadoop based recommeder are
> different. My question does not make any sense now.
>
> I am tring to bring Mahout to my working place and I am actually reading
> your book Mahout in Action. I think the support  for different kind of User
> similarities as well as support for evaluating a recommender  will  really
> save us a lot of time.  I  understand that  for a item based recommeder,
> one can do pre-computation for ,  say, co-occurance matix through Hadoop.
> I am still looking for a way to build a hadoop based recommender without
> writing mappers or reducers.  I mean I want to be able to write code
> almost as simple as the  following:
>
>  DataModel model = new FileDataModel(new File("myrating.csv"));
>
>    UserSimilarity similarity = new PearsonCorrelationSimilarity(model);
>    UserNeighborhood neighborhood =
>      new NearestNUserNeighborhood(2, similarity, model);
>
>    Recommender recommender = new GenericUserBasedRecommender(
>        model, neighborhood, similarity);
>
>
>
> I do believe that writing  customized  hadoop based components will most
> likely  be necessary even if  what I expected above does exists.
>
> Your time is  greatly appreciated.
>
> Thanks,
>
> Jack
>
>
> On Sun, Dec 25, 2011 at 8:12 PM, Sean Owen <[email protected]> wrote:
>
>> I'm not sure quite what you are asking. No, it is not all built on top
>> of Hadoop. If you run a Hadoop-based job on 1 node, it is easy to run
>> it on 100 nodes. The non-Hadoop-based recommender is completely
>> different from the Hadoop-based recommender and they are not
>> interchangeable. I am not sure what you mean by "level of
>> integration".
>>
>> On Sun, Dec 25, 2011 at 9:00 PM, Jinyuan Zhou <[email protected]>
>> wrote:
>> > Hi,
>> > I had a impression that mahout is build on top of Hadoop. For this I
>> expect
>> > that,  for  a recommender I build. After I run it successfuly with modest
>> > data on on mahcine, I should be able to  run the same recommender with
>> > Hadoop cluster for the purpose of handling huge data. What I expect is
>> > that  mahout will allow me do some configuration about my remcomender and
>> > Hadoop cluster and then it is good to run that with  power on Hadoop. Is
>> > this true?  I know Hbase or big they are build on top of Hadoop, when
>> they
>> > run command the  useage of Hadoop is transparent to user. That is , the
>> > contruction of hadoop job,  construction of job jar as well as hadoop
>> > command for running the job in Hadoop are all trasparent to user. Does
>> > Mahout support this level of  integeration with Hadoop.
>> >
>> > Thanks,
>> >
>> > Jack
>>

Reply via email to