On Sunday 10 August 2008 08:02:52 Dhruba Borthakur wrote:
> In almost all hadoop configurations, all host names can be specified
> as IP address. So, in your hadoop-site.xml, please specify the IP
> address of the namenode (instead of its hostname).

Which is not necessary a solution for complicated situations => e.g. where a 
port is bound to a different IP address that is used to access it. Basically 
NAT ;)

I mention this, because one common environment for Hadoop, EC2 does exactly 
that ;)

Andreas

>
> -dhruba
>
> 2008/8/8 Lucas Nazário dos Santos <[EMAIL PROTECTED]>:
> > Thanks Andreas. I'll try it.
> >
> > On Fri, Aug 8, 2008 at 5:47 PM, Andreas Kostyrka 
<[EMAIL PROTECTED]>wrote:
> >> On Friday 08 August 2008 15:43:46 Lucas Nazário dos Santos wrote:
> >> > You are completely right. It's not safe at all. But this is what I
> >> > have
> >>
> >> for
> >>
> >> > now:
> >> > two computers distributed across the Internet. I would really
> >> > appreciate
> >>
> >> if
> >>
> >> > anyone could give me spark on how to configure the namenode's IP in a
> >> > datanode. As I could identify in log files, the datanode keeps trying
> >> > to connect
> >> > to the IP 10.1.1.5, which is the internal IP of the namenode. I just
> >>
> >> need a
> >>
> >> > way
> >> > to say to the datanode "Hey, could you instead connect to the IP
> >>
> >> 172.1.23.2
> >>
> >> > "?
> >>
> >> Your only bet is to set it up in a VPNed environment. That would make it
> >> securitywise okay too.
> >>
> >> Andreas
> >>
> >> > Lucas
> >> >
> >> > On Fri, Aug 8, 2008 at 10:25 AM, Lukáš Vlček <[EMAIL PROTECTED]>
> >>
> >> wrote:
> >> > > HI,
> >> > >
> >> > > I am not an expert on Hadoop configuration but is this safe? As far
> >> > > as
> >>
> >> I
> >>
> >> > > understand the IP address is public and connection to the datanode
> >> > > port is not secured. Am I correct?
> >> > >
> >> > > Lukas
> >> > >
> >> > > On Fri, Aug 8, 2008 at 8:35 AM, Lucas Nazário dos Santos <
> >> > >
> >> > > [EMAIL PROTECTED]> wrote:
> >> > > > Hello again,
> >> > > >
> >> > > > In fact I can get the cluster up and running with two nodes in
> >> > > > different LANs. The problem appears when executing a job.
> >> > > >
> >> > > > As you can see in the piece of log bellow, the datanode tries to
> >> > >
> >> > > comunicate
> >> > >
> >> > > > with the namenode using the IP 10.1.1.5. The issue is that the
> >>
> >> datanode
> >>
> >> > > > should be using a valid IP, and not 10.1.1.5.
> >> > > >
> >> > > > Is there a way of manually configuring the datanode with the
> >>
> >> namenode's
> >>
> >> > > IP,
> >> > >
> >> > > > so I can change from 10.1.1.5 to, say 189.11.131.172?
> >> > > >
> >> > > > Thanks,
> >> > > > Lucas
> >> > > >
> >> > > >
> >> > > > 2008-08-08 02:34:23,335 INFO org.apache.hadoop.mapred.TaskTracker:
> >> > > > TaskTracker up at: localhost/127.0.0.1:60394
> >> > > > 2008-08-08 02:34:23,335 INFO org.apache.hadoop.mapred.TaskTracker:
> >> > >
> >> > > Starting
> >> > >
> >> > > > tracker tracker_localhost:localhost/127.0.0.1:60394
> >> > > > 2008-08-08 02:34:23,589 INFO org.apache.hadoop.mapred.TaskTracker:
> >> > >
> >> > > Starting
> >> > >
> >> > > > thread: Map-events fetcher for all reduce tasks on
> >> > > > tracker_localhost:localhost/127.0.0.1:60394
> >> > > > 2008-08-08 03:06:43,239 INFO org.apache.hadoop.mapred.TaskTracker:
> >> > > > LaunchTaskAction: task_200808080234_0001_m_000000_0
> >> > > > 2008-08-08 03:07:43,989 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 1 time(s).
> >> > > > 2008-08-08 03:08:44,999 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 2 time(s).
> >> > > > 2008-08-08 03:09:45,999 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 3 time(s).
> >> > > > 2008-08-08 03:10:47,009 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 4 time(s).
> >> > > > 2008-08-08 03:11:48,009 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 5 time(s).
> >> > > > 2008-08-08 03:12:49,026 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 6 time(s).
> >> > > > 2008-08-08 03:13:50,036 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 7 time(s).
> >> > > > 2008-08-08 03:14:51,046 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 8 time(s).
> >> > > > 2008-08-08 03:15:52,056 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 9 time(s).
> >> > > > 2008-08-08 03:16:53,066 INFO org.apache.hadoop.ipc.Client:
> >> > > > Retrying
> >> > >
> >> > > connect
> >> > >
> >> > > > to server: /10.1.1.5:9000. Already tried 10 time(s).
> >> > > > 2008-08-08 03:17:54,077 WARN org.apache.hadoop.mapred.TaskTracker:
> >> > > > Error initializing task_200808080234_0001_m_000000_0:
> >> > > > java.net.SocketTimeoutException
> >> > > >    at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:109)
> >> > > >    at
> >>
> >> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:174)
> >>
> >> > > >    at org.apache.hadoop.ipc.Client.getConnection(Client.java:623)
> >> > > >    at org.apache.hadoop.ipc.Client.call(Client.java:546)
> >> > > >    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:212)
> >> > > >    at org.apache.hadoop.dfs.$Proxy5.getProtocolVersion(Unknown
> >>
> >> Source)
> >>
> >> > > >    at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:313)
> >> > > >    at
> >> > >
> >> > > org.apache.hadoop.dfs.DFSClient.createRPCNamenode(DFSClient.java:102
> >> > >)
> >> > >
> >> > > >    at org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:178)
> >> > > >    at
> >>
> >> org.apache.hadoop.dfs.DistributedFileSystem.initialize(DistributedFileSy
> >>s
> >>
> >> > >tem.java:68)
> >> > >
> >> > > >    at
> >>
> >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1280)
> >>
> >> > > >    at
> >> > > > org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:56) at
> >> > > > org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1291) at
> >> > > > org.apache.hadoop.fs.FileSystem.get(FileSystem.java:203) at
> >> > > > org.apache.hadoop.fs.FileSystem.getNamed(FileSystem.java:152) at
> >>
> >> org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:670)
> >>
> >> > > >    at
> >>
> >> org.apache.hadoop.mapred.TaskTracker.startNewTask(TaskTracker.java:1274
> >>
> >> > > >) at
> >>
> >> org.apache.hadoop.mapred.TaskTracker.offerService(TaskTracker.java:915)
> >>
> >> > > >    at
> >> > > > org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java:1310) at
> >>
> >> org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:2251)
> >>
> >> > > > On Fri, Aug 8, 2008 at 12:16 AM, Lucas Nazário dos Santos <
> >> > > >
> >> > > > [EMAIL PROTECTED]> wrote:
> >> > > > > Hello,
> >> > > > >
> >> > > > > Can someone point me out what are the extra tasks that need to
> >> > > > > be
> >> > > >
> >> > > > performed
> >> > > >
> >> > > > > in order to set up a cluster where nodes are spread over the
> >> > > > > Internet,
> >> > >
> >> > > in
> >> > >
> >> > > > > different LANs?
> >> > > > >
> >> > > > > Do I need to free any datanode/namenode ports? How do I get the
> >> > >
> >> > > datanodes
> >> > >
> >> > > > > to know the valid namenode IP, and not something like 10.1.1.1?
> >> > > > >
> >> > > > > Any help is appreciate.
> >> > > > >
> >> > > > > Lucas
> >> > >
> >> > > --
> >> > > http://blog.lukas-vlcek.com/


Attachment: signature.asc
Description: This is a digitally signed message part.

Reply via email to