Hello, I tried adding "usern...@hostname" for eachentry in slaves file.
My slave file have 2 data nodes.it looks like below localhost utdhado...@xxx.xx.xx.229 utdhad...@xxx.xx.xx.251 error what I get when i start dfs is as below: starting namenode, logging to /home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-utdhadoop1-namenode-opencirrus-992.hpl.hp.com.out xxx.xx.xx.229: starting datanode, logging to /home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-utdhadoop1-datanode-opencirrus-992.hpl.hp.com.out *xxx.xx.xx.251: bash: line 0: cd: /home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/..: No such file or directory xxx.xx.xx.251: bash: /home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/hadoop-daemon.sh: No such file or directory *localhost: datanode running as process 25814. Stop it first. xxx.xx.xx.229: starting secondarynamenode, logging to /home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-utdhadoop1-secondarynamenode-opencirrus-992.hpl.hp.com.out localhost: secondarynamenode running as process 25959. Stop it first. Basically it looks for "* /home/utdhadoop1/Hadoop/** hadoop-0.18.3/bin/hadoop-**daemon.sh" but instead it should look for "/home/utdhadoop/Hadoop/...." as xxx.xx.xx.251 has username as utdhadoop* . Any inputs?? Thanks Pankil On Wed, May 20, 2009 at 6:30 PM, Todd Lipcon <t...@cloudera.com> wrote: > On Wed, May 20, 2009 at 4:14 PM, Alex Loddengaard <a...@cloudera.com> > wrote: > > > First of all, if you can get all machines to have the same user, that > would > > greatly simplify things. > > > > If, for whatever reason, you absolutely can't get the same user on all > > machines, then you could do either of the following: > > > > 1) Change the *-all.sh scripts to read from a slaves file that has two > > fields: a host and a user > > > To add to what Alex said, you should actually already be able to do this > with the existing scripts by simply using the format "usern...@hostname" > for > each entry in the slaves file. > > -Todd >