Also if you are using a topology rack map, make sure you scripts responds correctly to every possible hostname or IP address as well.
On Tue, Jun 9, 2009 at 1:19 PM, John Martyniak<j...@avum.com> wrote: > It seems that this is the issue, as there several posts related to same > topic but with no resolution. > > I guess the thing of it is that it shouldn't use the hostname of the machine > at all. If I tell it the master is x and it has an IP Address of x.x.x.102 > that should be good enough. > > And if that isn't the case then I should be able to specify which network > adaptor to use as the ip address that it is going to lookup against, whether > it is by DNS or by /etc/hosts. > > Because I suspect the problem is that I have named the machine as > duey.xxxx.com but have told hadoop that machine is called duey-direct. > > Is there work around in 0.19.1? I am using this with Nutch so don't have an > option to upgrade at this time. > > -John > > > On Jun 9, 2009, at 11:59 AM, Steve Loughran wrote: > >> John Martyniak wrote: >>> >>> When I run either of those on either of the two machines, it is trying to >>> resolve against the DNS servers configured for the external addresses for >>> the box. >>> Here is the result >>> Server: xxx.xxx.xxx.69 >>> Address: xxx.xxx.xxx.69#53 >> >> OK. in an ideal world, each NIC has a different hostname. Now, that >> confuses code that assumes a host has exactly one hostname, not zero or two, >> and I'm not sure how well Hadoop handles the 2+ situation (I know it doesn't >> like 0, but hey, its a distributed application). With separate hostnames, >> you set hadoop up to work on the inner addresses, and give out the inner >> hostnames of the jobtracker and namenode. As a result, all traffic to the >> master nodes should be routed on the internal network > >