The append in 1.x is very broken. You'll run into very weird states and we officially do not support it (we even call out in the config as broken). I wouldn't recommend using it even if a simple test appears to work.
On Sat, Jul 6, 2013 at 6:27 AM, Mohammad Tariq <donta...@gmail.com> wrote: > @Robin East : Thank you for keeping me updated. I was on 1.0.3 when I had > tried append last time and it was not working despite of the fact that API > had it. I tried it with 1.1.2 and it seems to work fine. > > @Manickam : Apologies for the incorrect info. Latest stable release(1.1.2) > supports append. But, you should consider whatever Harsh has said. > > Warm Regards, > Tariq > cloudfront.blogspot.com > > > On Fri, Jul 5, 2013 at 4:24 PM, Harsh J <ha...@cloudera.com> wrote: >> >> If it is 1k new records at the "end of the file" then you may extract >> them out and append the existing file in HDFS. I'd recommend using >> HDFS from Apache Hadoop 2.x for this purpose. >> >> On Fri, Jul 5, 2013 at 4:22 PM, Manickam P <manicka...@outlook.com> wrote: >> > Hi, >> > >> > Let me explain the question clearly. I have a file which has one million >> > records and i moved into my hadoop cluster. >> > After one month i got a new file which has same one million plus 1000 >> > new >> > records added in end of the file. >> > Here i just want to move the 1000 records alone into HDFS instead of >> > overwriting the entire file. >> > >> > Can i use HBase for this scenario? i don't have clear idea about HBase. >> > Just >> > asking. >> > >> > >> > >> > >> > Thanks, >> > Manickam P >> > >> > >> >> From: ha...@cloudera.com >> >> Date: Fri, 5 Jul 2013 16:13:16 +0530 >> > >> >> Subject: Re: How to update a file which is in HDFS >> >> To: user@hadoop.apache.org >> > >> >> >> >> The answer to the "delta" part is more that HDFS does not presently >> >> support random writes. You cannot alter a closed file for anything >> >> other than appending at the end, which I doubt will help you if you >> >> are also receiving updates (it isn't clear from your question what >> >> this added data really is). >> >> >> >> HBase sounds like something that may solve your requirement though, >> >> depending on how much of your read/write load is random. You could >> >> consider it. >> >> >> >> P.s. HBase too doesn't use the append() APIs today (and doesn't need >> >> it either). AFAIK, only Flume's making use of it, if you allow it to. >> >> >> >> On Thu, Jul 4, 2013 at 5:17 PM, Mohammad Tariq <donta...@gmail.com> >> >> wrote: >> >> > Hello Manickam, >> >> > >> >> > Append is currently not possible. >> >> > >> >> > Warm Regards, >> >> > Tariq >> >> > cloudfront.blogspot.com >> >> > >> >> > >> >> > On Thu, Jul 4, 2013 at 4:40 PM, Manickam P <manicka...@outlook.com> >> >> > wrote: >> >> >> >> >> >> Hi, >> >> >> >> >> >> I have moved my input file into the HDFS location in the cluster >> >> >> setup. >> >> >> Now i got a new set of file which has some new records along with >> >> >> the >> >> >> old >> >> >> one. >> >> >> I want to move the delta part alone into HDFS because it will take >> >> >> more >> >> >> time to move the file from my local to HDFS location. >> >> >> Is it possible or do i need to move the entire file into HDFS again? >> >> >> >> >> >> >> >> >> >> >> >> Thanks, >> >> >> Manickam P >> >> > >> >> > >> >> >> >> >> >> >> >> -- >> >> Harsh J >> >> >> >> -- >> Harsh J > > -- Harsh J