Actually I'm trying to do something like combining multiple namenodes so
that they present themselves to clients as a single namespace, implementing
basic namenode functionalities.

在 2013年4月18日星期四,Chris Embree 写道:

> Glad you got this working... can you explain your use case a little?   I'm
> trying to understand why you might want to do that.
>
>
> On Thu, Apr 18, 2013 at 11:29 AM, Lixiang Ao 
> <aolixi...@gmail.com<javascript:_e({}, 'cvml', 'aolixi...@gmail.com');>
> > wrote:
>
>> I modified sbin/hadoop-daemon.sh, where HADOOP_PID_DIR is set. It works!
>>  Everything looks fine now.
>>
>> Seems direct command "hdfs namenode" gives a better sense of control  :)
>>
>> Thanks a lot.
>>
>> 在 2013年4月18日星期四,Harsh J 写道:
>>
>> Yes you can but if you want the scripts to work, you should have them
>>> use a different PID directory (I think its called HADOOP_PID_DIR)
>>> every time you invoke them.
>>>
>>> I instead prefer to start the daemons up via their direct command such
>>> as "hdfs namenode" and so and move them to the background, with a
>>> redirect for logging.
>>>
>>> On Thu, Apr 18, 2013 at 2:34 PM, Lixiang Ao <aolixi...@gmail.com> wrote:
>>> > Hi all,
>>> >
>>> > Can I run mutiple HDFS instances, that is, n seperate namenodes and n
>>> > datanodes, on a single machine?
>>> >
>>> > I've modified core-site.xml and hdfs-site.xml to avoid port and file
>>> > conflicting between HDFSes, but when I started the second HDFS, I got
>>> the
>>> > errors:
>>> >
>>> > Starting namenodes on [localhost]
>>> > localhost: namenode running as process 20544. Stop it first.
>>> > localhost: datanode running as process 20786. Stop it first.
>>> > Starting secondary namenodes [0.0.0.0]
>>> > 0.0.0.0: secondarynamenode running as process 21074. Stop it first.
>>> >
>>> > Is there a way to solve this?
>>> > Thank you in advance,
>>> >
>>> > Lixiang Ao
>>>
>>>
>>>
>>> --
>>> Harsh J
>>>
>>
>

Reply via email to