Also, it only provides access to a local chunk of a file which isn't very
useful.

On Mon, Apr 11, 2011 at 5:32 PM, Edward Capriolo <edlinuxg...@gmail.com>wrote:

> On Mon, Apr 11, 2011 at 7:05 PM, Jason Rutherglen
> <jason.rutherg...@gmail.com> wrote:
> > Yes you can however it will require customization of HDFS.  Take a
> > look at HDFS-347 specifically the HDFS-347-branch-20-append.txt patch.
> >  I have been altering it for use with HBASE-3529.  Note that the patch
> > noted is for the -append branch which is mainly for HBase.
> >
> > On Mon, Apr 11, 2011 at 3:57 PM, Benson Margulies <bimargul...@gmail.com>
> wrote:
> >> We have some very large files that we access via memory mapping in
> >> Java. Someone's asked us about how to make this conveniently
> >> deployable in Hadoop. If we tell them to put the files into hdfs, can
> >> we obtain a File for the underlying file on any given node?
> >>
> >
>
> This features it not yet part of hadoop so doing this is not "convenient".
>

Reply via email to