Hi, Arv, Actually, several days ago, I deployed a system which is similar with your requirements
In our cluster environment, since I have to run modified hadoop, we invoked two namenodes, two jobtrackers, two trackers on each node, and as you mentioned, two datanodes in single host, What you have to do is that you should ensure there is no conflicts on the port of datanodes occupy(modify the configuration files), and for assigning separate disks , just set the different root directory of your HDFS, I guess that you also have to run two namenodes just like we have done Cheers Nan On Wed, Sep 15, 2010 at 11:50 PM, Arv Mistry <a...@kindsight.net> wrote: > Hi, > > Is it possible to run multiple data nodes on a single machine? I > currently have a machine with multiple disks and enough disk capacity > for replication across them. I don't need redundancy at the machine > level but would like to be able to handle a single disk failure. > > So I was thinking if I can run multiple DataNodes on a single machine > each assigned a separate disk that would give me the protection I need > against disk failure. > > Can anyone give me any insights in to how I would setup multiple > DataNodes to run on a single machine? Thanks in advance, > > Cheers Arv >