Yes, this is the suggested configuration. Hadoop relies on password-less SSH to be able to start tasks on slave machines. You can find instructions on creating/transferring the SSH keys here:
http://www.michael-noll.com/wiki/Running_Hadoop_On_Ubuntu_Linux_%28Multi-Node_Cluster%29 On Wed, Apr 23, 2008 at 4:39 AM, Sridhar Raman <[EMAIL PROTECTED]> wrote: > Ok, what about the issue regarding the users? Do all the machines need to > be under the same user? > > On Wed, Apr 23, 2008 at 12:43 PM, Harish Mallipeddi < > [EMAIL PROTECTED]> wrote: > > > On Wed, Apr 23, 2008 at 3:03 PM, Sridhar Raman <[EMAIL PROTECTED]> > > wrote: > > > > > After trying out Hadoop in a single machine, I decided to run a > > MapReduce > > > across multiple machines. This is the approach I followed: > > > 1 Master > > > 1 Slave > > > > > > (A doubt here: Can my Master also be used to execute the Map/Reduce > > > functions?) > > > > > > > If you add the master node to the list of slaves (conf/slaves), then the > > master node run will also run a TaskTracker. > > > > > > > > > > To do this, I set up the masters and slaves files in the conf > directory. > > > Following the instructions in this page - > > > http://hadoop.apache.org/core/docs/current/cluster_setup.html, I had > set > > > up > > > sshd in both the machines, and was able to ssh from one to the other. > > > > > > I tried to run bin/start-dfs.sh. Unfortunately, this asked for a > > password > > > for [EMAIL PROTECTED], while in slave, there was only user2. While in > master, > > > user1 was the logged on user. How do I resolve this? Should the user > > > accounts be present in all the machines? Or can I specify this > > somewhere? > > > > > > > > > > > -- > > Harish Mallipeddi > > circos.com : poundbang.in/blog/ > > >