Yes, this is the suggested configuration.  Hadoop relies on password-less
SSH to be able to start tasks on slave machines.  You can find instructions
on creating/transferring the SSH keys here:

http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Multi-Node_Cluster%29

On Wed, Apr 23, 2008 at 4:39 AM, Sridhar Raman <[EMAIL PROTECTED]>
wrote:

> Ok, what about the issue regarding the users?  Do all the machines need to
> be under the same user?
>
> On Wed, Apr 23, 2008 at 12:43 PM, Harish Mallipeddi <
> [EMAIL PROTECTED]> wrote:
>
> > On Wed, Apr 23, 2008 at 3:03 PM, Sridhar Raman <[EMAIL PROTECTED]>
> > wrote:
> >
> > > After trying out Hadoop in a single machine, I decided to run a
> > MapReduce
> > > across multiple machines.  This is the approach I followed:
> > > 1 Master
> > > 1 Slave
> > >
> > > (A doubt here:  Can my Master also be used to execute the Map/Reduce
> > > functions?)
> > >
> >
> > If you add the master node to the list of slaves (conf/slaves), then the
> > master node run will also run a TaskTracker.
> >
> >
> > >
> > > To do this, I set up the masters and slaves files in the conf
> directory.
> > > Following the instructions in this page -
> > > http://hadoop.apache.org/core/docs/current/cluster_setup.html, I had
> set
> > > up
> > > sshd in both the machines, and was able to ssh from one to the other.
> > >
> > > I tried to run bin/start-dfs.sh.  Unfortunately, this asked for a
> > password
> > > for [EMAIL PROTECTED], while in slave, there was only user2.  While in
> master,
> > > user1 was the logged on user.  How do I resolve this?  Should the user
> > > accounts be present in all the machines?  Or can I specify this
> > somewhere?
> > >
> >
> >
> >
> > --
> > Harish Mallipeddi
> > circos.com : poundbang.in/blog/
> >
>

Reply via email to