Hi Saptarshi:
Please refer the following example code, I wish it can help you. JobConf grepJob = new JobConf(getConf(), Grep.class); try { grepJob.setJobName("search"); FileInputFormat.setInputPaths(grepJob, args[0]); ………………………………… FileOutputFormat.setOutputPath(grepJob, tempDir); ………………………………. JobClient.runJob(grepJob); JobConf sortJob = new JobConf(Grep.class); sortJob.setJobName("sort"); FileInputFormat.setInputPaths(sortJob, tempDir); ………………………………. FileOutputFormat.setOutputPath(sortJob, new Path(args[1])); …………………………………….. JobClient.runJob(sortJob); --Jerry -----邮件原件----- 发件人: Saptarshi Guha [mailto:[EMAIL PROTECTED] 发送时间: 2008年11月11日 12:06 收件人: core-user@hadoop.apache.org 主题: Passing information from one job to the next in a JobControl Hello, I am using JobControl to run a sequence of jobs(Job_1,Job_2,..Job_n) on after the other. Each job returns some information e.g key1 value1,value2 key2 value1,value2 and so on. This can be found in the outdir passed to the jar file. Is there a way for Job_1 to return some data (which can be passed onto the Job_2), without my main program having to read the information from the file in the HDFS? I could use things like Linda Spaces, however does MapReduce have a framework for this? Thanks Saptarshi -- Saptarshi Guha - [EMAIL PROTECTED]