@Robin East :  Thank you for keeping me updated. I was on 1.0.3 when I had
tried append last time and it was not working despite of the fact that API
had it. I tried it with 1.1.2 and it seems to work fine.

@Manickam : Apologies for the incorrect info. Latest stable release(1.1.2)
supports append. But, you should consider whatever Harsh has said.

Warm Regards,
Tariq
cloudfront.blogspot.com


On Fri, Jul 5, 2013 at 4:24 PM, Harsh J <ha...@cloudera.com> wrote:

> If it is 1k new records at the "end of the file" then you may extract
> them out and append the existing file in HDFS. I'd recommend using
> HDFS from Apache Hadoop 2.x for this purpose.
>
> On Fri, Jul 5, 2013 at 4:22 PM, Manickam P <manicka...@outlook.com> wrote:
> > Hi,
> >
> > Let me explain the question clearly. I have a file which has one million
> > records and i moved into my hadoop cluster.
> > After one month i got a new file which has same one million plus 1000 new
> > records added in end of the file.
> > Here i just want to move the 1000 records alone into HDFS instead of
> > overwriting the entire file.
> >
> > Can i use HBase for this scenario? i don't have clear idea about HBase.
> Just
> > asking.
> >
> >
> >
> >
> > Thanks,
> > Manickam P
> >
> >
> >> From: ha...@cloudera.com
> >> Date: Fri, 5 Jul 2013 16:13:16 +0530
> >
> >> Subject: Re: How to update a file which is in HDFS
> >> To: user@hadoop.apache.org
> >
> >>
> >> The answer to the "delta" part is more that HDFS does not presently
> >> support random writes. You cannot alter a closed file for anything
> >> other than appending at the end, which I doubt will help you if you
> >> are also receiving updates (it isn't clear from your question what
> >> this added data really is).
> >>
> >> HBase sounds like something that may solve your requirement though,
> >> depending on how much of your read/write load is random. You could
> >> consider it.
> >>
> >> P.s. HBase too doesn't use the append() APIs today (and doesn't need
> >> it either). AFAIK, only Flume's making use of it, if you allow it to.
> >>
> >> On Thu, Jul 4, 2013 at 5:17 PM, Mohammad Tariq <donta...@gmail.com>
> wrote:
> >> > Hello Manickam,
> >> >
> >> > Append is currently not possible.
> >> >
> >> > Warm Regards,
> >> > Tariq
> >> > cloudfront.blogspot.com
> >> >
> >> >
> >> > On Thu, Jul 4, 2013 at 4:40 PM, Manickam P <manicka...@outlook.com>
> >> > wrote:
> >> >>
> >> >> Hi,
> >> >>
> >> >> I have moved my input file into the HDFS location in the cluster
> setup.
> >> >> Now i got a new set of file which has some new records along with the
> >> >> old
> >> >> one.
> >> >> I want to move the delta part alone into HDFS because it will take
> more
> >> >> time to move the file from my local to HDFS location.
> >> >> Is it possible or do i need to move the entire file into HDFS again?
> >> >>
> >> >>
> >> >>
> >> >> Thanks,
> >> >> Manickam P
> >> >
> >> >
> >>
> >>
> >>
> >> --
> >> Harsh J
>
>
>
> --
> Harsh J
>

Reply via email to