I am still getting the same exception. This is the stack trace of it. java.io.IOException: Not a file: hdfs://zeus:18004/user/hadoop/output6/MatrixA-Row1 at org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:195) at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:797) at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1142) at MatrixMultiplication.main(MatrixMultiplication.java:229) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.util.RunJar.main(RunJar.java:165) at org.apache.hadoop.mapred.JobShell.run(JobShell.java:54) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) at org.apache.hadoop.mapred.JobShell.main(JobShell.java:68)
Abhishek Agrawal SUNY- Buffalo (716-435-7122) On Sun 11/22/09 9:28 PM , Jason Venner jason.had...@gmail.com sent: > set the number of reduce tasks to 1. > > 2009/11/22 > Hi everybody, > The 10 different map-reducers store their > respective outputs in 10 > different files. This is the snap shot > had...@zeus:~/hadoop-0.19.1$ bin/hadoop dfs -ls output5 > Found 2 items > drwxr-xr-x - hadoop supergroup 0 2003-05-16 02:16 > /user/hadoop/output5/MatrixA-Row1 > drwxr-xr-x - hadoop supergroup 0 2003-05-16 02:16 > /user/hadoop/output5/MatrixA-Row2 > Now when I try to open any of these files I get an error message > had...@zeus:~/hadoop-0.19.1$ bin/hadoop dfs -cat > output5/MatrixA-Row1 > cat: Source must be a file. > had...@zeus:~/hadoop-0.19.1$ > But if I run > had...@zeus:~/hadoop-0.19.1$ bin/hadoop dfs -cat > output5/MatrixA-Row1/part-00000 > I get the correct output. I do not understand why I have to give > this extra > "part-00000". Now when I run a map reduce task to merge the outputs > of all the > files, I give the name of the directory output5 as the Input path. > But I get a > bug saying > java.io.IOException: Not a file: > hdfs://zeus:18004/user/hadoop/output5/MatrixA-Row1 > I cannot understand how to make the frame work read my files. > Alternatively I tried to avoid the map reduce approach for combining > files and do > it via a simple program, but I am unable to start. Can some one give > me some > sample implementation or something. > Any help is appreciated > Thank You > Abhishek Agrawal > SUNY- Buffalo > (716-435-7122) > On Sun 11/22/09 5:48 PM , sent: > > Hellow, > > If I write the output of the 10 tasks in 10 different files then > how do > > Igo about merging the output ? Is there some in built > functionality or do I > > haveto write some code for that ? > > > > Thank You > > > > Abhishek Agrawal > > > > SUNY- Buffalo > > (716-435-7122) > > > > On Sun 11/22/09 5:40 PM , Gang Luo lgpubli > > sent:> Hi. If the output path already exists, it seems > > you could not execute any> task with the same output path. I think > you can > > output the results of the> 10 tasks to 10 different paths, and > then do sth > > more (by the 11th task, for> example) to merge the 10 results into > 1 file. > > > > > > Gang Luo > > > --------- > > > Department of Computer Science > > > Duke University > > > (919)316-0993 > > > gang....@du> ke.edu [4] > > > > > > > > > ----- > > å��å&Acir > > c;§ï¿½Ã©ï¿½Â&r > > eg;ä»¶ ----> > > å��ä&A > > circ;»Â¶Ã¤ÂºÂºÃ > > ¯Â¼ï¿½ "aa...@buffa> lo.edu [5]" > > lo.edu [6]>�Ã&curre > > n;»¶äºº&Ati > > lde;¯Â¼ï¿½ common-u...@hadoop.apache.orgå�ï > > ¿½Ã©ï¿½ï¿½& > > Atilde;¦ï¿½Â¥Ã¦ï&iqu > > est;½ï¿½Ã¯Â¼ï&iques > > t;½ 2009/11/22> > > (å�¨æï&ique > > st;½Â¥) 5:25:55 > > ä¸�åï&iqu > > est;½ï¿½Ã¤Â¸Â&raq > > uo; > > é¢�ïÂ&frac > > 14;� Help in Hadoop> > > > Hello Everybody, > > > I have a doubt in a map reduce program and I > > would appreciate any> help. I run the program using the command > > bin/hadoop jar HomeWork.jar prg1> inputoutput. Ideally from within > prg1, I want to > > sequentially launch 10 map-> reducetasks. I want to store the > output of all > > these map reduce tasks in some> file.Currently I have kept the > input format and > > output format of the jobs as> TextInputFormat and TextOutputFormat > > respectively. Now I have the> followingquestions. > > > > > > 1. When I run more than 1 task from the same > > program, the output file of> all thetasks is same. The framework > does not > > allows the 2 map reduce task to> have thesame output file as > task 1. > > > > > > 2. Before the 2 task launches I also get this > > error > > > > Cannot initialize JVM Metrics with > > processName=JobTracker, sessionId= -> alreadyinitialized > > > > > > 3. When the 2 map reduce tasks writes its output > > to file> "output", wont theprevious content of > > this file get over written ?> > > > Thank You > > > > > > Abhishek Agrawal > > > > > > SUNY- Buffalo > > > (716-435-7122) > > > > > > > > > > > ___________________________________________________________ > > > 好ç� > > ;©è´ºåï&i > > quest;½Â¡&cce> > > dil;Â�你> > ; �Ã&ma > > cr;¼�é� > > ;®ç> > > ;®±è´ºà > > ;¥ï¿½Â¡Ã¥ï¿&frac1 > > 2;¨æ�°> > > ;ä¸�çÂ& > > ordm;¿ï¼� http://card.mail.cn.yahoo.com/ [7]> > > > > > > > > > > > > > -- > Pro Hadoop, a book to guide you from beginner to hadoop mastery, > http://www.amazon.com/dp/1430219424?tag=jewlerymall [8] > www.prohadoopbook.com [9] a community for Hadoop Professionals > > > Links: > ------ > [4] http://ke.edu > [5] http://lo.edu > [6] http://lo.edu > [7] http://card.mail.cn.yahoo.com/ > [8] http://www.amazon.com/dp/1430219424?tag=jewlerymall > [9] http://www.prohadoopbook.com > >