Are you trying to implement something like namespace federation, that's a
part of Hadoop 2.0 -
http://hadoop.apache.org/docs/r2.0.3-alpha/hadoop-project-dist/hadoop-hdfs/Federation.html


On Thu, Apr 18, 2013 at 10:02 PM, Lixiang Ao <aolixi...@gmail.com> wrote:

> Actually I'm trying to do something like combining multiple namenodes so
> that they present themselves to clients as a single namespace, implementing
> basic namenode functionalities.
>
> 在 2013年4月18日星期四,Chris Embree 写道:
>
> Glad you got this working... can you explain your use case a little?   I'm
>> trying to understand why you might want to do that.
>>
>>
>> On Thu, Apr 18, 2013 at 11:29 AM, Lixiang Ao <aolixi...@gmail.com> wrote:
>>
>>> I modified sbin/hadoop-daemon.sh, where HADOOP_PID_DIR is set. It works!
>>>  Everything looks fine now.
>>>
>>> Seems direct command "hdfs namenode" gives a better sense of control  :)
>>>
>>> Thanks a lot.
>>>
>>> 在 2013年4月18日星期四,Harsh J 写道:
>>>
>>> Yes you can but if you want the scripts to work, you should have them
>>>> use a different PID directory (I think its called HADOOP_PID_DIR)
>>>> every time you invoke them.
>>>>
>>>> I instead prefer to start the daemons up via their direct command such
>>>> as "hdfs namenode" and so and move them to the background, with a
>>>> redirect for logging.
>>>>
>>>> On Thu, Apr 18, 2013 at 2:34 PM, Lixiang Ao <aolixi...@gmail.com>
>>>> wrote:
>>>> > Hi all,
>>>> >
>>>> > Can I run mutiple HDFS instances, that is, n seperate namenodes and n
>>>> > datanodes, on a single machine?
>>>> >
>>>> > I've modified core-site.xml and hdfs-site.xml to avoid port and file
>>>> > conflicting between HDFSes, but when I started the second HDFS, I got
>>>> the
>>>> > errors:
>>>> >
>>>> > Starting namenodes on [localhost]
>>>> > localhost: namenode running as process 20544. Stop it first.
>>>> > localhost: datanode running as process 20786. Stop it first.
>>>> > Starting secondary namenodes [0.0.0.0]
>>>> > 0.0.0.0: secondarynamenode running as process 21074. Stop it first.
>>>> >
>>>> > Is there a way to solve this?
>>>> > Thank you in advance,
>>>> >
>>>> > Lixiang Ao
>>>>
>>>>
>>>>
>>>> --
>>>> Harsh J
>>>>
>>>
>>

Reply via email to