On Thu, Sep 2, 2010 at 3:34 AM, Gagandeep Singh
<[email protected]> wrote:
> Hi Daniel
>
> I have downloaded hadoop-0.20.2+320.tar.gz from this location
> http://archive.cloudera.com/cdh/3/


That looks right, yes.

> And also changed the *dfs.support.append* flag to *true* in your *
> hdfs-site.xml* as mentioned here
> http://wiki.apache.org/hadoop/Hbase/HdfsSyncSupport.
>

That sounds right too.  As Ted suggests, you put it in to all configs
(though I believe it enabled by default on that branch -- in the UI
you'd see a warning if it was NOT enabled).

> But data loss is still happening. Am I using the right version?
> Is there any other settings that I need to make so that data gets flushed to
> HDFS.
>

It looks like you are doing the right thing.  Can we see master log please?

Thanks,
St.Ack


> Thanks,
> Gagan
>
>
>
> On Thu, Aug 26, 2010 at 11:57 PM, Jean-Daniel Cryans 
> <[email protected]>wrote:
>
>> That, or use CDH3b2.
>>
>> J-D
>>
>> On Thu, Aug 26, 2010 at 11:22 AM, Gagandeep Singh
>> <[email protected]> wrote:
>> > Thanks Daniel
>> >
>> > It means I have to checkout the code from branch and build it on my local
>> > machine.
>> >
>> > Gagan
>> >
>> >
>> > On Thu, Aug 26, 2010 at 9:51 PM, Jean-Daniel Cryans <[email protected]
>> >wrote:
>> >
>> >> Then I would expect some form of dataloss yes, because stock hadoop
>> >> 0.20 doesn't have any form of fsync so HBase doesn't know whether the
>> >> data made it to the datanodes when appending to the WAL. Please use
>> >> the 0.20-append hadoop branch with HBase 0.89 or cloudera's CDH3b2.
>> >>
>> >> J-D
>> >>
>> >> On Thu, Aug 26, 2010 at 7:22 AM, Gagandeep Singh
>> >> <[email protected]> wrote:
>> >> > HBase - 0.20.5
>> >> > Hadoop - 0.20.2
>> >> >
>> >> > Thanks,
>> >> > Gagan
>> >> >
>> >> >
>> >> >
>> >> > On Thu, Aug 26, 2010 at 7:11 PM, Jean-Daniel Cryans <
>> [email protected]
>> >> >wrote:
>> >> >
>> >> >> Hadoop and HBase version?
>> >> >>
>> >> >> J-D
>> >> >>
>> >> >> On Aug 26, 2010 5:36 AM, "Gagandeep Singh" <
>> [email protected]>
>> >> >> wrote:
>> >> >>
>> >> >> Hi Group,
>> >> >>
>> >> >> I am checking HBase/HDFS fail over. I am inserting 1M records from my
>> >> HBase
>> >> >> client application. I am clubbing my Put operation such that 10
>> records
>> >> get
>> >> >> added into the List<Put> and then I call the table.put(). I have not
>> >> >> modified the default setting of Put operation which means all data is
>> >> >> written in WAL and in case of server failure my data should not be
>> lost.
>> >> >>
>> >> >> But I noticed somewhat strange behavior, while adding records if I
>> kill
>> >> my
>> >> >> Region Server then my application waits till the time region data is
>> >> moved
>> >> >> to another region. But I noticed while doing so all my data is lost
>> and
>> >> my
>> >> >> table is emptied.
>> >> >>
>> >> >> Could you help me understand the behavior. Is there some kind of
>> Cache
>> >> also
>> >> >> involved while writing because of which my data is lost.
>> >> >>
>> >> >>
>> >> >> Thanks,
>> >> >> Gagan
>> >> >>
>> >> >
>> >>
>> >
>>
>

Reply via email to