In almost all hadoop configurations, all host names can be specified
as IP address. So, in your hadoop-site.xml, please specify the IP
address of the namenode (instead of its hostname).

-dhruba

2008/8/8 Lucas Nazário dos Santos <[EMAIL PROTECTED]>:
> Thanks Andreas. I'll try it.
>
>
> On Fri, Aug 8, 2008 at 5:47 PM, Andreas Kostyrka <[EMAIL PROTECTED]>wrote:
>
>> On Friday 08 August 2008 15:43:46 Lucas Nazário dos Santos wrote:
>> > You are completely right. It's not safe at all. But this is what I have
>> for
>> > now:
>> > two computers distributed across the Internet. I would really appreciate
>> if
>> > anyone could give me spark on how to configure the namenode's IP in a
>> > datanode. As I could identify in log files, the datanode keeps trying to
>> > connect
>> > to the IP 10.1.1.5, which is the internal IP of the namenode. I just
>> need a
>> > way
>> > to say to the datanode "Hey, could you instead connect to the IP
>> 172.1.23.2
>> > "?
>>
>> Your only bet is to set it up in a VPNed environment. That would make it
>> securitywise okay too.
>>
>> Andreas
>>
>> >
>> > Lucas
>> >
>> > On Fri, Aug 8, 2008 at 10:25 AM, Lukáš Vlček <[EMAIL PROTECTED]>
>> wrote:
>> > > HI,
>> > >
>> > > I am not an expert on Hadoop configuration but is this safe? As far as
>> I
>> > > understand the IP address is public and connection to the datanode port
>> > > is not secured. Am I correct?
>> > >
>> > > Lukas
>> > >
>> > > On Fri, Aug 8, 2008 at 8:35 AM, Lucas Nazário dos Santos <
>> > >
>> > > [EMAIL PROTECTED]> wrote:
>> > > > Hello again,
>> > > >
>> > > > In fact I can get the cluster up and running with two nodes in
>> > > > different LANs. The problem appears when executing a job.
>> > > >
>> > > > As you can see in the piece of log bellow, the datanode tries to
>> > >
>> > > comunicate
>> > >
>> > > > with the namenode using the IP 10.1.1.5. The issue is that the
>> datanode
>> > > > should be using a valid IP, and not 10.1.1.5.
>> > > >
>> > > > Is there a way of manually configuring the datanode with the
>> namenode's
>> > >
>> > > IP,
>> > >
>> > > > so I can change from 10.1.1.5 to, say 189.11.131.172?
>> > > >
>> > > > Thanks,
>> > > > Lucas
>> > > >
>> > > >
>> > > > 2008-08-08 02:34:23,335 INFO org.apache.hadoop.mapred.TaskTracker:
>> > > > TaskTracker up at: localhost/127.0.0.1:60394
>> > > > 2008-08-08 02:34:23,335 INFO org.apache.hadoop.mapred.TaskTracker:
>> > >
>> > > Starting
>> > >
>> > > > tracker tracker_localhost:localhost/127.0.0.1:60394
>> > > > 2008-08-08 02:34:23,589 INFO org.apache.hadoop.mapred.TaskTracker:
>> > >
>> > > Starting
>> > >
>> > > > thread: Map-events fetcher for all reduce tasks on
>> > > > tracker_localhost:localhost/127.0.0.1:60394
>> > > > 2008-08-08 03:06:43,239 INFO org.apache.hadoop.mapred.TaskTracker:
>> > > > LaunchTaskAction: task_200808080234_0001_m_000000_0
>> > > > 2008-08-08 03:07:43,989 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 1 time(s).
>> > > > 2008-08-08 03:08:44,999 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 2 time(s).
>> > > > 2008-08-08 03:09:45,999 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 3 time(s).
>> > > > 2008-08-08 03:10:47,009 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 4 time(s).
>> > > > 2008-08-08 03:11:48,009 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 5 time(s).
>> > > > 2008-08-08 03:12:49,026 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 6 time(s).
>> > > > 2008-08-08 03:13:50,036 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 7 time(s).
>> > > > 2008-08-08 03:14:51,046 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 8 time(s).
>> > > > 2008-08-08 03:15:52,056 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 9 time(s).
>> > > > 2008-08-08 03:16:53,066 INFO org.apache.hadoop.ipc.Client: Retrying
>> > >
>> > > connect
>> > >
>> > > > to server: /10.1.1.5:9000. Already tried 10 time(s).
>> > > > 2008-08-08 03:17:54,077 WARN org.apache.hadoop.mapred.TaskTracker:
>> > > > Error initializing task_200808080234_0001_m_000000_0:
>> > > > java.net.SocketTimeoutException
>> > > >    at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:109)
>> > > >    at
>> > > >
>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:174)
>> > > >    at org.apache.hadoop.ipc.Client.getConnection(Client.java:623)
>> > > >    at org.apache.hadoop.ipc.Client.call(Client.java:546)
>> > > >    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:212)
>> > > >    at org.apache.hadoop.dfs.$Proxy5.getProtocolVersion(Unknown
>> Source)
>> > > >    at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:313)
>> > > >    at
>> > >
>> > > org.apache.hadoop.dfs.DFSClient.createRPCNamenode(DFSClient.java:102)
>> > >
>> > > >    at org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:178)
>> > > >    at
>> > >
>> > >
>> org.apache.hadoop.dfs.DistributedFileSystem.initialize(DistributedFileSys
>> > >tem.java:68)
>> > >
>> > > >    at
>> > > >
>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1280)
>> > > >    at org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:56)
>> > > >    at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1291)
>> > > >    at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:203)
>> > > >    at org.apache.hadoop.fs.FileSystem.getNamed(FileSystem.java:152)
>> > > >    at
>> > > >
>> org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:670)
>> > > >    at
>> > > >
>> org.apache.hadoop.mapred.TaskTracker.startNewTask(TaskTracker.java:1274
>> > > >) at
>> > > >
>> org.apache.hadoop.mapred.TaskTracker.offerService(TaskTracker.java:915)
>> > > >    at org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java:1310)
>> > > >    at
>> org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:2251)
>> > > >
>> > > >
>> > > >
>> > > > On Fri, Aug 8, 2008 at 12:16 AM, Lucas Nazário dos Santos <
>> > > >
>> > > > [EMAIL PROTECTED]> wrote:
>> > > > > Hello,
>> > > > >
>> > > > > Can someone point me out what are the extra tasks that need to be
>> > > >
>> > > > performed
>> > > >
>> > > > > in order to set up a cluster where nodes are spread over the
>> > > > > Internet,
>> > >
>> > > in
>> > >
>> > > > > different LANs?
>> > > > >
>> > > > > Do I need to free any datanode/namenode ports? How do I get the
>> > >
>> > > datanodes
>> > >
>> > > > > to know the valid namenode IP, and not something like 10.1.1.1?
>> > > > >
>> > > > > Any help is appreciate.
>> > > > >
>> > > > > Lucas
>> > >
>> > > --
>> > > http://blog.lukas-vlcek.com/
>>
>>
>>
>

Reply via email to