Glad you got this working... can you explain your use case a little?   I'm
trying to understand why you might want to do that.


On Thu, Apr 18, 2013 at 11:29 AM, Lixiang Ao <aolixi...@gmail.com> wrote:

> I modified sbin/hadoop-daemon.sh, where HADOOP_PID_DIR is set. It works!
>  Everything looks fine now.
>
> Seems direct command "hdfs namenode" gives a better sense of control  :)
>
> Thanks a lot.
>
> 在 2013年4月18日星期四,Harsh J 写道:
>
> Yes you can but if you want the scripts to work, you should have them
>> use a different PID directory (I think its called HADOOP_PID_DIR)
>> every time you invoke them.
>>
>> I instead prefer to start the daemons up via their direct command such
>> as "hdfs namenode" and so and move them to the background, with a
>> redirect for logging.
>>
>> On Thu, Apr 18, 2013 at 2:34 PM, Lixiang Ao <aolixi...@gmail.com> wrote:
>> > Hi all,
>> >
>> > Can I run mutiple HDFS instances, that is, n seperate namenodes and n
>> > datanodes, on a single machine?
>> >
>> > I've modified core-site.xml and hdfs-site.xml to avoid port and file
>> > conflicting between HDFSes, but when I started the second HDFS, I got
>> the
>> > errors:
>> >
>> > Starting namenodes on [localhost]
>> > localhost: namenode running as process 20544. Stop it first.
>> > localhost: datanode running as process 20786. Stop it first.
>> > Starting secondary namenodes [0.0.0.0]
>> > 0.0.0.0: secondarynamenode running as process 21074. Stop it first.
>> >
>> > Is there a way to solve this?
>> > Thank you in advance,
>> >
>> > Lixiang Ao
>>
>>
>>
>> --
>> Harsh J
>>
>

Reply via email to