It wouldn't be too much of a stretch to use Lustre directly... although it isn't trivial either. You'd need to implement the 'FileSystem' interface for Lustre, define a URI scheme (e.g. lfs://) etc. Please take a take a look at the KFS/ S3 implementations.

Arun

On Aug 21, 2008, at 9:59 AM, Joel Welling wrote:

Hi folks;
 I'm new to Hadoop, and I'm trying to set it up on a cluster for which
almost all the disk is mounted via the Lustre filesystem.  That
filesystem is visible to all the nodes, so I don't actually need HDFS to
implement a shared filesystem.  (I know the philosophical reasons why
people say local disks are better for Hadoop, but that's not the
situation I've got). My system is failing, and I think it's because the
different nodes are tripping over each other when they try to run HDFS
out of the same directory tree.
Is there a way to turn off HDFS and just let Lustre do the distributed
filesystem?  I've seen discussion threads about Hadoop with NFS which
said something like 'just specify a local filesystem and everything will
be fine', but I don't know how to do that.  I'm using Hadoop 0.17.2.

Thanks, I hope;
-Joel Welling
[EMAIL PROTECTED]


Reply via email to