Hi Uma, Thanks a lot. At last it is running without errors. Thank you very much for your suggestion.
On 26 December 2011 20:04, Uma Maheswara Rao G <mahesw...@huawei.com> wrote: > Hey Humayun, > Looks your hostname still not resoling properly. even though you > configured hostnames as master, slave...etc, it is getting humayun as > hostname. > just edit /etc/HOSTNAME file with correct hostname what you are expecting > here. > To confirm whether it is resolving properly or not, you can just do below > steps > #hostname > ............................//should get hostname here > correctly ( ex: master) > #hostname -i > ..........................//should resolve correct IP > here ... ( ex: master ip) > > > and make sure slave and slave1 sre pingable each other. > > Regards, > Uma > > ________________________________________ > From: Humayun kabir [humayun0...@gmail.com] > Sent: Saturday, December 24, 2011 9:51 PM > To: common-user@hadoop.apache.org > Subject: Re: Hadoop configuration > > i've checked my log files. But i don't understand to why this error occurs. > here i my logs files. please give me some suggestion. > > jobtracker.log < http://paste.ubuntu.com/781181/ > > > namenode.log < http://paste.ubuntu.com/781183/ > > > datanode.log(1st machine) < http://paste.ubuntu.com/781176/ > > > datanode.log(2nd machine) < http://paste.ubuntu.com/781195 > <goog_2054845717>/ > > > > tasktracker.log(1st machine) < http://paste.ubuntu.com/781192/ > > > tasktracker.log(2nd machine) < http://paste.ubuntu.com/781197/ > > > > > On 24 December 2011 15:26, Joey Krabacher <jkrabac...@gmail.com> wrote: > > > have you checked your log files for any clues? > > > > --Joey > > > > On Sat, Dec 24, 2011 at 3:15 AM, Humayun kabir <humayun0...@gmail.com> > > wrote: > > > Hi Uma, > > > > > > Thank you very much for your tips. We tried it in 3 nodes in virtual > box > > as > > > you suggested. But still we are facing problem. Here is our all > > > configuration file to all nodes. please take a look and show us some > ways > > > to solve it. It was nice and it would be great if you help us in this > > > regards. > > > > > > <http://core-site.xml>core-site.xml < http://pastebin.com/Twn5edrp > > > > hdfs-site.xml < http://pastebin.com/k4hR4GE9 > > > > mapred-site.xml < http://pastebin.com/gZuyHswS > > > > > > > /etc/hosts < http://pastebin.com/5s0yhgnj > > > > > > > output < http://paste.ubuntu.com/780807/ > > > > > > > > > > Hope you will understand and extend your helping hand towards us. > > > > > > Have a nice day. > > > > > > Regards > > > Humayun > > > > > > On 23 December 2011 17:31, Uma Maheswara Rao G <mahesw...@huawei.com> > > wrote: > > > > > >> Hi Humayun , > > >> > > >> Lets assume you have JT, TT1, TT2, TT3 > > >> > > >> Now you should configure the \etc\hosts like below examle > > >> > > >> 10.18.xx.1 JT > > >> > > >> 10.18.xx.2 TT1 > > >> > > >> 10.18.xx.3 TT2 > > >> > > >> 10.18.xx.4 TT3 > > >> > > >> Configure the same set in all the machines, so that all task > trackers > > >> can talk each other with hostnames correctly. Also pls remove some > > entries > > >> from your files > > >> > > >> 127.0.0.1 localhost.localdomain localhost > > >> > > >> 127.0.1.1 humayun > > >> > > >> > > >> > > >> I have seen others already suggested many links for the regular > > >> configuration items. Hope you might clear about them. > > >> > > >> hope it will help... > > >> > > >> Regards, > > >> > > >> Uma > > >> > > >> ________________________________ > > >> > > >> From: Humayun kabir [humayun0...@gmail.com] > > >> Sent: Thursday, December 22, 2011 10:34 PM > > >> To: common-user@hadoop.apache.org; Uma Maheswara Rao G > > >> Subject: Re: Hadoop configuration > > >> > > >> Hello Uma, > > >> > > >> Thanks for your cordial and quick reply. It would be great if you > > explain > > >> what you suggested to do. Right now we are running on following > > >> configuration. > > >> > > >> We are using hadoop on virtual box. when it is a single node then it > > works > > >> fine for big dataset larger than the default block size. but in case > of > > >> multinode cluster (2 nodes) we are facing some problems. We are able > to > > >> ping both "Master->Slave" and "Slave->Master". > > >> Like when the input dataset is smaller than the default block size(64 > > MB) > > >> then it works fine. but when the input dataset is larger than the > > default > > >> block size then it shows ‘too much fetch failure’ in reduce state. > > >> here is the output link > > >> http://paste.ubuntu.com/707517/ > > >> > > >> this is our /etc/hosts file > > >> > > >> 192.168.60.147 humayun # Added by NetworkManager > > >> 127.0.0.1 localhost.localdomain localhost > > >> ::1 humayun localhost6.localdomain6 localhost6 > > >> 127.0.1.1 humayun > > >> > > >> # The following lines are desirable for IPv6 capable hosts > > >> ::1 localhost ip6-localhost ip6-loopback > > >> fe00::0 ip6-localnet > > >> ff00::0 ip6-mcastprefix > > >> ff02::1 ip6-allnodes > > >> ff02::2 ip6-allrouters > > >> ff02::3 ip6-allhosts > > >> > > >> 192.168.60.1 master > > >> 192.168.60.2 slave > > >> > > >> > > >> Regards, > > >> > > >> -Humayun. > > >> > > >> > > >> On 22 December 2011 15:47, Uma Maheswara Rao G <mahesw...@huawei.com > > >> <mailto:mahesw...@huawei.com>> wrote: > > >> Hey Humayun, > > >> > > >> To solve the too many fetch failures problem, you should configure > host > > >> mapping correctly. > > >> Each tasktracker should be able to ping from each other. > > >> > > >> Regards, > > >> Uma > > >> ________________________________________ > > >> From: Humayun kabir [humayun0...@gmail.com<mailto: > humayun0...@gmail.com > > >] > > >> Sent: Thursday, December 22, 2011 2:54 PM > > >> To: common-user@hadoop.apache.org<mailto: > common-user@hadoop.apache.org> > > >> Subject: Hadoop configuration > > >> > > >> someone please help me to configure hadoop such as core-site.xml, > > >> hdfs-site.xml, mapred-site.xml etc. > > >> please provide some example. it is badly needed. because i run in a 2 > > node > > >> cluster. when i run the wordcount example then it gives the result too > > >> mutch fetch failure. > > >> > > >> > > >