But i'd have to find out that id manually by looking in the web
interface right?
I think it will be difficult to find the right hadoop job, because
multiple RecommenderJobs can run at the same time in my system.
So, just in theory, a way to make this work from java would mean to add
a method to RecommenderJob that calls killJob() on the currently running
hadoop job. For mahout jobs used in RecommenderJob (e.g.
RowSimilarityJob) the same had to be done so RecommenderJob could also
cancel mahout sub tasks.
Do you think it could work that way?
On 02/04/12 19:05, Sean Owen wrote:
You can use the Hadoop interface itself (like, the command-line hadoop
tool) to kill a job by its ID. If you kill one MapReduce job the
entire process should halt after that.
On Mon, Apr 2, 2012 at 6:44 PM, Sören Brunk<soren.br...@deri.org> wrote:
Hi,
I'm using the distributed RecommenderJob from within a Java program.
For that, in a separate thread, I'm creating a RecommenderJob object, call
setConf() for the hadoop configuration and then run() with the job
parameters.
This is working fine for me but now I would like to be able to stop a
running job.
Not sure if that's possible at all since RecommenderJob encapsulates several
Hadoop jobs (or even other Mahout jobs that call Hadoop in turn) and runs
them in a blocking way.
Would be interesting for other Mahout jobs as well.
Any ideas?
Thanks,
--
Sören Brunk
Research Assistant
Data Intensive Infrastructures Unit (DI2)
Digital Enterprise Research Institute
National University of Ireland Galway