A while back I was playing with building a cluster connected via
OpenVPN - it was working well for the most part.  The complication I
personally had was with making the OpenVPN nodes recognize eachother
and connect at boot time (I was using custom EC2 AMIs and wanted a way
to have external NAT'ed systems communicate with the cluster).

The trick is creating the proper UDP port forwarding rules to have the
nodes initially create the VPN.  After that, the Hadoop nodes can all
use the VPN IP's to intercommunicate.

-- Jim R. Wilson (jimbojw)

On Sun, Aug 10, 2008 at 4:29 PM, Andreas Kostyrka <[EMAIL PROTECTED]> wrote:
> On Sunday 10 August 2008 08:02:52 Dhruba Borthakur wrote:
>> In almost all hadoop configurations, all host names can be specified
>> as IP address. So, in your hadoop-site.xml, please specify the IP
>> address of the namenode (instead of its hostname).
>
> Which is not necessary a solution for complicated situations => e.g. where a
> port is bound to a different IP address that is used to access it. Basically
> NAT ;)
>
> I mention this, because one common environment for Hadoop, EC2 does exactly
> that ;)
>
> Andreas
>
>>
>> -dhruba
>>
>> 2008/8/8 Lucas Nazário dos Santos <[EMAIL PROTECTED]>:
>> > Thanks Andreas. I'll try it.
>> >
>> > On Fri, Aug 8, 2008 at 5:47 PM, Andreas Kostyrka
> <[EMAIL PROTECTED]>wrote:
>> >> On Friday 08 August 2008 15:43:46 Lucas Nazário dos Santos wrote:
>> >> > You are completely right. It's not safe at all. But this is what I
>> >> > have
>> >>
>> >> for
>> >>
>> >> > now:
>> >> > two computers distributed across the Internet. I would really
>> >> > appreciate
>> >>
>> >> if
>> >>
>> >> > anyone could give me spark on how to configure the namenode's IP in a
>> >> > datanode. As I could identify in log files, the datanode keeps trying
>> >> > to connect
>> >> > to the IP 10.1.1.5, which is the internal IP of the namenode. I just
>> >>
>> >> need a
>> >>
>> >> > way
>> >> > to say to the datanode "Hey, could you instead connect to the IP
>> >>
>> >> 172.1.23.2
>> >>
>> >> > "?
>> >>
>> >> Your only bet is to set it up in a VPNed environment. That would make it
>> >> securitywise okay too.
>> >>
>> >> Andreas
>> >>
>> >> > Lucas
>> >> >
>> >> > On Fri, Aug 8, 2008 at 10:25 AM, Lukáš Vlček <[EMAIL PROTECTED]>
>> >>
>> >> wrote:
>> >> > > HI,
>> >> > >
>> >> > > I am not an expert on Hadoop configuration but is this safe? As far
>> >> > > as
>> >>
>> >> I
>> >>
>> >> > > understand the IP address is public and connection to the datanode
>> >> > > port is not secured. Am I correct?
>> >> > >
>> >> > > Lukas
>> >> > >
>> >> > > On Fri, Aug 8, 2008 at 8:35 AM, Lucas Nazário dos Santos <
>> >> > >
>> >> > > [EMAIL PROTECTED]> wrote:
>> >> > > > Hello again,
>> >> > > >
>> >> > > > In fact I can get the cluster up and running with two nodes in
>> >> > > > different LANs. The problem appears when executing a job.
>> >> > > >
>> >> > > > As you can see in the piece of log bellow, the datanode tries to
>> >> > >
>> >> > > comunicate
>> >> > >
>> >> > > > with the namenode using the IP 10.1.1.5. The issue is that the
>> >>
>> >> datanode
>> >>
>> >> > > > should be using a valid IP, and not 10.1.1.5.
>> >> > > >
>> >> > > > Is there a way of manually configuring the datanode with the
>> >>
>> >> namenode's
>> >>
>> >> > > IP,
>> >> > >
>> >> > > > so I can change from 10.1.1.5 to, say 189.11.131.172?
>> >> > > >
>> >> > > > Thanks,
>> >> > > > Lucas
>> >> > > >
>> >> > > >
>> >> > > > 2008-08-08 02:34:23,335 INFO org.apache.hadoop.mapred.TaskTracker:
>> >> > > > TaskTracker up at: localhost/127.0.0.1:60394
>> >> > > > 2008-08-08 02:34:23,335 INFO org.apache.hadoop.mapred.TaskTracker:
>> >> > >
>> >> > > Starting
>> >> > >
>> >> > > > tracker tracker_localhost:localhost/127.0.0.1:60394
>> >> > > > 2008-08-08 02:34:23,589 INFO org.apache.hadoop.mapred.TaskTracker:
>> >> > >
>> >> > > Starting
>> >> > >
>> >> > > > thread: Map-events fetcher for all reduce tasks on
>> >> > > > tracker_localhost:localhost/127.0.0.1:60394
>> >> > > > 2008-08-08 03:06:43,239 INFO org.apache.hadoop.mapred.TaskTracker:
>> >> > > > LaunchTaskAction: task_200808080234_0001_m_000000_0
>> >> > > > 2008-08-08 03:07:43,989 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 1 time(s).
>> >> > > > 2008-08-08 03:08:44,999 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 2 time(s).
>> >> > > > 2008-08-08 03:09:45,999 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 3 time(s).
>> >> > > > 2008-08-08 03:10:47,009 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 4 time(s).
>> >> > > > 2008-08-08 03:11:48,009 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 5 time(s).
>> >> > > > 2008-08-08 03:12:49,026 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 6 time(s).
>> >> > > > 2008-08-08 03:13:50,036 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 7 time(s).
>> >> > > > 2008-08-08 03:14:51,046 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 8 time(s).
>> >> > > > 2008-08-08 03:15:52,056 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 9 time(s).
>> >> > > > 2008-08-08 03:16:53,066 INFO org.apache.hadoop.ipc.Client:
>> >> > > > Retrying
>> >> > >
>> >> > > connect
>> >> > >
>> >> > > > to server: /10.1.1.5:9000. Already tried 10 time(s).
>> >> > > > 2008-08-08 03:17:54,077 WARN org.apache.hadoop.mapred.TaskTracker:
>> >> > > > Error initializing task_200808080234_0001_m_000000_0:
>> >> > > > java.net.SocketTimeoutException
>> >> > > >    at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:109)
>> >> > > >    at
>> >>
>> >> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:174)
>> >>
>> >> > > >    at org.apache.hadoop.ipc.Client.getConnection(Client.java:623)
>> >> > > >    at org.apache.hadoop.ipc.Client.call(Client.java:546)
>> >> > > >    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:212)
>> >> > > >    at org.apache.hadoop.dfs.$Proxy5.getProtocolVersion(Unknown
>> >>
>> >> Source)
>> >>
>> >> > > >    at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:313)
>> >> > > >    at
>> >> > >
>> >> > > org.apache.hadoop.dfs.DFSClient.createRPCNamenode(DFSClient.java:102
>> >> > >)
>> >> > >
>> >> > > >    at org.apache.hadoop.dfs.DFSClient.<init>(DFSClient.java:178)
>> >> > > >    at
>> >>
>> >> org.apache.hadoop.dfs.DistributedFileSystem.initialize(DistributedFileSy
>> >>s
>> >>
>> >> > >tem.java:68)
>> >> > >
>> >> > > >    at
>> >>
>> >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1280)
>> >>
>> >> > > >    at
>> >> > > > org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:56) at
>> >> > > > org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1291) at
>> >> > > > org.apache.hadoop.fs.FileSystem.get(FileSystem.java:203) at
>> >> > > > org.apache.hadoop.fs.FileSystem.getNamed(FileSystem.java:152) at
>> >>
>> >> org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:670)
>> >>
>> >> > > >    at
>> >>
>> >> org.apache.hadoop.mapred.TaskTracker.startNewTask(TaskTracker.java:1274
>> >>
>> >> > > >) at
>> >>
>> >> org.apache.hadoop.mapred.TaskTracker.offerService(TaskTracker.java:915)
>> >>
>> >> > > >    at
>> >> > > > org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java:1310) at
>> >>
>> >> org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:2251)
>> >>
>> >> > > > On Fri, Aug 8, 2008 at 12:16 AM, Lucas Nazário dos Santos <
>> >> > > >
>> >> > > > [EMAIL PROTECTED]> wrote:
>> >> > > > > Hello,
>> >> > > > >
>> >> > > > > Can someone point me out what are the extra tasks that need to
>> >> > > > > be
>> >> > > >
>> >> > > > performed
>> >> > > >
>> >> > > > > in order to set up a cluster where nodes are spread over the
>> >> > > > > Internet,
>> >> > >
>> >> > > in
>> >> > >
>> >> > > > > different LANs?
>> >> > > > >
>> >> > > > > Do I need to free any datanode/namenode ports? How do I get the
>> >> > >
>> >> > > datanodes
>> >> > >
>> >> > > > > to know the valid namenode IP, and not something like 10.1.1.1?
>> >> > > > >
>> >> > > > > Any help is appreciate.
>> >> > > > >
>> >> > > > > Lucas
>> >> > >
>> >> > > --
>> >> > > http://blog.lukas-vlcek.com/
>
>
>

Reply via email to