> 2. Could you set up symlinks from the local filesystem, so point every 
> node at a local dir
>   /tmp/hadoop
> with each node pointing to a different subdir in the big filesystem?

Yes, I could do that!  Do I need to do it for the log directories as
well, or can they be shared?

-Joel

On Fri, 2008-08-22 at 15:48 +0100, Steve Loughran wrote:
> Joel Welling wrote:
> > Thanks, Steve and Arun.  I'll definitely try to write something based on
> > the KFS interface.  I think that for our applications putting the mapper
> > on the right rack is not going to be that useful.  A lot of our
> > calculations are going to be disordered stuff based on 3D spatial
> > relationships like nearest-neighbor finding, so things will be in a
> > random access pattern most of the time.
> > 
> > Is there a way to set up the configuration for HDFS so that different
> > datanodes keep their data in different directories?  That would be a big
> > help in the short term.
> 
> yes, but you'd have to push out a different config to each datanode.
> 
> 1. I have some stuff that could help there, but its not ready for 
> production use yet [1].
> 
> 2. Could you set up symlinks from the local filesystem, so point every 
> node at a local dir
>   /tmp/hadoop
> with each node pointing to a different subdir in the big filesystem?
> 
> 
> [1] 
> http://people.apache.org/~stevel/slides/deploying_hadoop_with_smartfrog.pdf

Reply via email to