Looks like it may be running in the local mode. Have you setup your
Eclipse configuration properly?

What version of Hadoop are you using?

On Wed, Jul 20, 2011 at 7:35 AM, foo_foo_foo <finallya...@gmail.com> wrote:
>
> I am a Hadoop novice so kindly pardon my ingorance.
>
> I am running the following Hadoop program in Fully Distributed Mode to count
> the number of lines in a file. I am running this job from eclipse and I see
> it running (based on the output to the eclipse console) but I do not see the
> tasks in the TaskTracker web interface. Also eventhough the data is
> distributed accross multiple hosts it doesnt seem to be distributing works
> accross hosts.
>
> Could someone pelase help me with this.
>
>
>        package LineCount;
>
>        import java.util.*;
>        import java.io.*;
>
>        import org.apache.hadoop.mapreduce.Mapper;
>        import org.apache.hadoop.mapreduce.Reducer;
>        import org.apache.hadoop.mapreduce.Job;
>        import org.apache.hadoop.io.*;
>        import org.apache.hadoop.conf.*;
>        import org.apache.hadoop.fs.Path;
>        import org.apache.hadoop.mapreduce.lib.input.*;
>        import org.apache.hadoop.mapreduce.lib.output.*;
>        import org.apache.hadoop.util.*;
>
>        public class LineCount extends Configured implements Tool {
>
>                public static class Map extends 
> Mapper<LongWritable,Text,Text,IntWritable>
> {
>                        private static int counter = 1;
>                        private static Text mapOpKey = new Text();
>                        private final static IntWritable mapOpValue = new 
> IntWritable(1);
>                        @Override
>                        public void map(LongWritable mapInpKey, Text 
> mapInpValue,
> Mapper<LongWritable,Text,Text,IntWritable>.Context context) throws
> IOException,InterruptedException{
>                                        System.out.println("Calling Map "+ 
> counter);
>                                        counter++;
>                                        mapOpKey.set("Number Of Lines");
>                                        context.write(mapOpKey, mapOpValue);
>                        }
>                }
>
>                public static class Reduce extends
> Reducer<Text,IntWritable,Text,IntWritable> {
>                        private static int counter = 1;
>                        @Override
>                        public void reduce(Text redIpKey, 
> Iterable<IntWritable> redIpValue,
> Reducer<Text,IntWritable,Text,IntWritable>.Context context) throws
> IOException,InterruptedException {
>                                int sum=0;
>                                System.out.println("Calling Reduce "+ counter);
>                                counter++;
>                                while(redIpValue.iterator().hasNext()){
>                                        sum = sum + 
> redIpValue.iterator().next().get();
>                                }
>                                context.write(redIpKey, new IntWritable(sum));
>                        }
>                }
>
>                @Override
>                public int run(String[] args) throws Exception{
>
>                        Configuration conf = new Configuration();
>                        conf.addResource(new 
> Path("/hadoop-0.20.2/conf/core-site.xml"));
>                        Job job = new Job(conf);
>                        job.setJobName("LineCount");
>                        job.setJarByClass(LineCount.class);
>                        job.setOutputKeyClass(Text.class);
>                        job.setOutputValueClass(IntWritable.class);
>                        job.setMapperClass(Map.class);
>                        //job.setCombinerClass(Reduce.class);
>                        job.setReducerClass(Reduce.class);
>                        FileInputFormat.setInputPaths(job, new 
> Path("/usr/foo/hadoopIP"));
>                        FileOutputFormat.setOutputPath(job, new 
> Path("/usr/foo/hadoopOP"));
>                        job.waitForCompletion(true);
>                        return 0;
>                }
>
>                public static void main(String[] args) throws Exception{
>                        ToolRunner.run(new LineCount(), args);
>                }
>        }
>
>
> --
> View this message in context: 
> http://old.nabble.com/Job-progress-not-showing-in-Hadoop-Tasktracker--web-interface-tp32096156p32096156.html
> Sent from the Hadoop core-user mailing list archive at Nabble.com.
>
>



-- 
Harsh J

Reply via email to