PFP consists of three mapreduce jobs. which job did you modified? And do you mean the number of total reduce task or just reduce task that is running?
Regards, Q 2012/8/30 C.V.Krishnakumar Iyer <cvkrishnaku...@me.com> > Hi, > > I've already tried setting it in the code using job.setNumReduceTasks() > and conf.set("mapred.reduce.tasks","100"). > However, it does not seem to take the number of reducers at all, even for > the job that does parallel counting. Any advice would be appreciated. > Regards, > Krishnakumar. > On Aug 29, 2012, at 11:28 PM, 戴清灏 <rogerda...@gmail.com> wrote: > > > I doubt that you specify the config in hadoop config xml file. > > > > -- > > Regards, > > Q > > > > > > > > 2012/8/30 C.V. Krishnakumar Iyer <cvkrishnaku...@me.com> > > > >> Hi, > >> > >> Quick question regarding PFPGrowth in Mahout 0.6: > >> > >> I see that there are no options to set the number of reducers in the > >> parallel counting phase of PFP Growth. It is just simple word count - so > >> I'm guessing it should be parallelized. But for some reason it is not! > >> > >> Is that intentional? > >> > >> Regards, > >> Krishnakumar. > >> > >