Did you do the clean work on all the datanodes? rm -Rf /path/to/my/hadoop/dfs/data
On 6/20/08, novice user <[EMAIL PROTECTED]> wrote: > > > Hi Mori Bellamy, > I did this twice. and still the same problem is persisting. I don't know > how to solve this issue. If any one know the answer, please let me know. > > Thanks > > Mori Bellamy wrote: > > > > That's bizarre. I'm not sure why your DFS would have magically gotten > > full. Whenever hadoop gives me trouble, i try the following sequence > > of commands > > > > stop-all.sh > > rm -Rf /path/to/my/hadoop/dfs/data > > hadoop namenode -format > > start-all.sh > > > > maybe you would get some luck if you ran that on all of the machines? > > (of course, don't run it if you don't want to lose all of that "data") > > On Jun 19, 2008, at 4:32 AM, novice user wrote: > > > >> > >> Hi Every one, > >> I am running a simple map-red application similar to k-means. But, > >> when I > >> ran it in on single machine, it went fine with out any issues. But, > >> when I > >> ran the same on a hadoop cluster of 9 machines. It fails saying > >> java.io.IOException: All datanodes are bad. Aborting... > >> > >> Here is more explanation about the problem: > >> I tried to upgrade my hadoop cluster to hadoop-17. During this > >> process, I > >> made a mistake of not installing hadoop on all machines. So, the > >> upgrade > >> failed. Nor I was able to roll back. So, I re-formatted the name node > >> afresh. and then hadoop installation was successful. > >> > >> Later, when I ran my map-reduce job, it ran successfully,but the > >> same job > >> with zero reduce tasks is failing with the error as: > >> java.io.IOException: All datanodes are bad. Aborting... > >> > >> When I looked into the data nodes, I figured out that file system is > >> 100% > >> full with different directories of name "subdir" in > >> hadoop-username/dfs/data/current directory. I am wondering where I > >> went > >> wrong. > >> Can some one please help me on this? > >> > >> The same job went fine on a single machine with same amount of input > >> data. > >> > >> Thanks > >> > >> > >> > >> -- > >> View this message in context: > >> > http://www.nabble.com/java.io.IOException%3A-All-datanodes-are-bad.-Aborting...-tp18006296p18006296.html > >> Sent from the Hadoop core-user mailing list archive at Nabble.com. > >> > > > > > > > > -- > View this message in context: > http://www.nabble.com/java.io.IOException%3A-All-datanodes-are-bad.-Aborting...-tp18006296p18022330.html > Sent from the Hadoop core-user mailing list archive at Nabble.com. > > -- 朱盛凯 Jash Zhu 复旦大学软件学院 Software School, Fudan University