I modified sbin/hadoop-daemon.sh, where HADOOP_PID_DIR is set. It works!
 Everything looks fine now.

Seems direct command "hdfs namenode" gives a better sense of control  :)

Thanks a lot.

在 2013年4月18日星期四,Harsh J 写道:

> Yes you can but if you want the scripts to work, you should have them
> use a different PID directory (I think its called HADOOP_PID_DIR)
> every time you invoke them.
>
> I instead prefer to start the daemons up via their direct command such
> as "hdfs namenode" and so and move them to the background, with a
> redirect for logging.
>
> On Thu, Apr 18, 2013 at 2:34 PM, Lixiang Ao <aolixi...@gmail.com> wrote:
> > Hi all,
> >
> > Can I run mutiple HDFS instances, that is, n seperate namenodes and n
> > datanodes, on a single machine?
> >
> > I've modified core-site.xml and hdfs-site.xml to avoid port and file
> > conflicting between HDFSes, but when I started the second HDFS, I got the
> > errors:
> >
> > Starting namenodes on [localhost]
> > localhost: namenode running as process 20544. Stop it first.
> > localhost: datanode running as process 20786. Stop it first.
> > Starting secondary namenodes [0.0.0.0]
> > 0.0.0.0: secondarynamenode running as process 21074. Stop it first.
> >
> > Is there a way to solve this?
> > Thank you in advance,
> >
> > Lixiang Ao
>
>
>
> --
> Harsh J
>

Reply via email to