Hello,

I tried adding "usern...@hostname" for eachentry in slaves file.

My slave file have 2 data nodes.it looks like below

localhost
utdhado...@xxx.xx.xx.229
utdhad...@xxx.xx.xx.251


error what I get when i start dfs is as below:

starting namenode, logging to
/home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-utdhadoop1-namenode-opencirrus-992.hpl.hp.com.out
xxx.xx.xx.229: starting datanode, logging to
/home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-utdhadoop1-datanode-opencirrus-992.hpl.hp.com.out
*xxx.xx.xx.251: bash: line 0: cd:
/home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/..: No such file or directory
xxx.xx.xx.251: bash:
/home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/hadoop-daemon.sh: No such file or
directory
*localhost: datanode running as process 25814. Stop it first.
xxx.xx.xx.229: starting secondarynamenode, logging to
/home/utdhadoop1/Hadoop/hadoop-0.18.3/bin/../logs/hadoop-utdhadoop1-secondarynamenode-opencirrus-992.hpl.hp.com.out
localhost: secondarynamenode running as process 25959. Stop it first.



Basically it looks for "* /home/utdhadoop1/Hadoop/**
hadoop-0.18.3/bin/hadoop-**daemon.sh"
but instead it should look for "/home/utdhadoop/Hadoop/...." as
xxx.xx.xx.251 has username as utdhadoop* .

Any inputs??

Thanks
Pankil

On Wed, May 20, 2009 at 6:30 PM, Todd Lipcon <t...@cloudera.com> wrote:

> On Wed, May 20, 2009 at 4:14 PM, Alex Loddengaard <a...@cloudera.com>
> wrote:
>
> > First of all, if you can get all machines to have the same user, that
> would
> > greatly simplify things.
> >
> > If, for whatever reason, you absolutely can't get the same user on all
> > machines, then you could do either of the following:
> >
> > 1) Change the *-all.sh scripts to read from a slaves file that has two
> > fields: a host and a user
>
>
> To add to what Alex said, you should actually already be able to do this
> with the existing scripts by simply using the format "usern...@hostname"
> for
> each entry in the slaves file.
>
> -Todd
>

Reply via email to