Hey Humayun, Looks your hostname still not resoling properly. even though you configured hostnames as master, slave...etc, it is getting humayun as hostname. just edit /etc/HOSTNAME file with correct hostname what you are expecting here. To confirm whether it is resolving properly or not, you can just do below steps #hostname ............................//should get hostname here correctly ( ex: master) #hostname -i ..........................//should resolve correct IP here ... ( ex: master ip)
and make sure slave and slave1 sre pingable each other. Regards, Uma ________________________________________ From: Humayun kabir [humayun0...@gmail.com] Sent: Saturday, December 24, 2011 9:51 PM To: common-user@hadoop.apache.org Subject: Re: Hadoop configuration i've checked my log files. But i don't understand to why this error occurs. here i my logs files. please give me some suggestion. jobtracker.log < http://paste.ubuntu.com/781181/ > namenode.log < http://paste.ubuntu.com/781183/ > datanode.log(1st machine) < http://paste.ubuntu.com/781176/ > datanode.log(2nd machine) < http://paste.ubuntu.com/781195<goog_2054845717>/ > tasktracker.log(1st machine) < http://paste.ubuntu.com/781192/ > tasktracker.log(2nd machine) < http://paste.ubuntu.com/781197/ > On 24 December 2011 15:26, Joey Krabacher <jkrabac...@gmail.com> wrote: > have you checked your log files for any clues? > > --Joey > > On Sat, Dec 24, 2011 at 3:15 AM, Humayun kabir <humayun0...@gmail.com> > wrote: > > Hi Uma, > > > > Thank you very much for your tips. We tried it in 3 nodes in virtual box > as > > you suggested. But still we are facing problem. Here is our all > > configuration file to all nodes. please take a look and show us some ways > > to solve it. It was nice and it would be great if you help us in this > > regards. > > > > <http://core-site.xml>core-site.xml < http://pastebin.com/Twn5edrp > > > hdfs-site.xml < http://pastebin.com/k4hR4GE9 > > > mapred-site.xml < http://pastebin.com/gZuyHswS > > > > > /etc/hosts < http://pastebin.com/5s0yhgnj > > > > > output < http://paste.ubuntu.com/780807/ > > > > > > > Hope you will understand and extend your helping hand towards us. > > > > Have a nice day. > > > > Regards > > Humayun > > > > On 23 December 2011 17:31, Uma Maheswara Rao G <mahesw...@huawei.com> > wrote: > > > >> Hi Humayun , > >> > >> Lets assume you have JT, TT1, TT2, TT3 > >> > >> Now you should configure the \etc\hosts like below examle > >> > >> 10.18.xx.1 JT > >> > >> 10.18.xx.2 TT1 > >> > >> 10.18.xx.3 TT2 > >> > >> 10.18.xx.4 TT3 > >> > >> Configure the same set in all the machines, so that all task trackers > >> can talk each other with hostnames correctly. Also pls remove some > entries > >> from your files > >> > >> 127.0.0.1 localhost.localdomain localhost > >> > >> 127.0.1.1 humayun > >> > >> > >> > >> I have seen others already suggested many links for the regular > >> configuration items. Hope you might clear about them. > >> > >> hope it will help... > >> > >> Regards, > >> > >> Uma > >> > >> ________________________________ > >> > >> From: Humayun kabir [humayun0...@gmail.com] > >> Sent: Thursday, December 22, 2011 10:34 PM > >> To: common-user@hadoop.apache.org; Uma Maheswara Rao G > >> Subject: Re: Hadoop configuration > >> > >> Hello Uma, > >> > >> Thanks for your cordial and quick reply. It would be great if you > explain > >> what you suggested to do. Right now we are running on following > >> configuration. > >> > >> We are using hadoop on virtual box. when it is a single node then it > works > >> fine for big dataset larger than the default block size. but in case of > >> multinode cluster (2 nodes) we are facing some problems. We are able to > >> ping both "Master->Slave" and "Slave->Master". > >> Like when the input dataset is smaller than the default block size(64 > MB) > >> then it works fine. but when the input dataset is larger than the > default > >> block size then it shows ‘too much fetch failure’ in reduce state. > >> here is the output link > >> http://paste.ubuntu.com/707517/ > >> > >> this is our /etc/hosts file > >> > >> 192.168.60.147 humayun # Added by NetworkManager > >> 127.0.0.1 localhost.localdomain localhost > >> ::1 humayun localhost6.localdomain6 localhost6 > >> 127.0.1.1 humayun > >> > >> # The following lines are desirable for IPv6 capable hosts > >> ::1 localhost ip6-localhost ip6-loopback > >> fe00::0 ip6-localnet > >> ff00::0 ip6-mcastprefix > >> ff02::1 ip6-allnodes > >> ff02::2 ip6-allrouters > >> ff02::3 ip6-allhosts > >> > >> 192.168.60.1 master > >> 192.168.60.2 slave > >> > >> > >> Regards, > >> > >> -Humayun. > >> > >> > >> On 22 December 2011 15:47, Uma Maheswara Rao G <mahesw...@huawei.com > >> <mailto:mahesw...@huawei.com>> wrote: > >> Hey Humayun, > >> > >> To solve the too many fetch failures problem, you should configure host > >> mapping correctly. > >> Each tasktracker should be able to ping from each other. > >> > >> Regards, > >> Uma > >> ________________________________________ > >> From: Humayun kabir [humayun0...@gmail.com<mailto:humayun0...@gmail.com > >] > >> Sent: Thursday, December 22, 2011 2:54 PM > >> To: common-user@hadoop.apache.org<mailto:common-user@hadoop.apache.org> > >> Subject: Hadoop configuration > >> > >> someone please help me to configure hadoop such as core-site.xml, > >> hdfs-site.xml, mapred-site.xml etc. > >> please provide some example. it is badly needed. because i run in a 2 > node > >> cluster. when i run the wordcount example then it gives the result too > >> mutch fetch failure. > >> > >> >