Ok.  At the very least, I'll put in a big warning in the book.




On 1/12/12 1:05 AM, "Stack" <st...@duboce.net> wrote:

>On Wed, Jan 11, 2012 at 3:39 PM, lars hofhansl <lhofha...@yahoo.com>
>wrote:
>> Shouldn't we deprecate them?
>>
>
>Yes.
>
>But what to replace them with?  I think they haven't been deprecated
>up to this because we've not done the work to put in place an
>alternative.  We could just drop the functionality (after
>deprecating).  I'd be fine w/ that.
>
>St.Ack
>
>> They are only guaranteed to work for the duration of a region server
>>operation.
>> If there are splits/etc the lock would be silently gone.
>>
>> -- Lars
>>
>>
>>
>> ________________________________
>>  From: Jean-Daniel Cryans <jdcry...@apache.org>
>> To: dev@hbase.apache.org
>> Sent: Wednesday, January 11, 2012 2:36 PM
>> Subject: Re: rowlock advice for book
>>
>> Row locks aren't deprecated officially but it's really that type of
>> feature that you should use at your own risk / you might not enjoy
>> life after that.
>>
>> J-D
>>
>> On Wed, Jan 11, 2012 at 1:08 PM, Doug Meil
>> <doug.m...@explorysmedical.com> wrote:
>>>
>>> Hey dev-list, regarding this...
>>>
>>> re:  "Be careful using hbase row locks.  They are (unofficially -- we
>>>need
>>> to make it official) deprecated."
>>>
>>>
>>>
>>>
>>> ... is this the official advice?  Should I update the book with this?
>>>
>>>
>>>
>>>
>>> On 1/3/12 4:37 PM, "Stack" <st...@duboce.net> wrote:
>>>
>>>>On Tue, Jan 3, 2012 at 12:38 PM, Joe Stein
>>>><charmal...@allthingshadoop.com> wrote:
>>>>> when the event happened so if we see something from November 3rd
>>>>>today
>>>>>then
>>>>> we will only keep it for 4 more months (and for events that we see
>>>>>today
>>>>> those stay for 6 months) .  so it sounds like this might be a viable
>>>>>option
>>>>> and when we set the timestamp in our checkAndPut we make the
>>>>>timestamp
>>>>>be
>>>>> the value that represents it as November 3rd, right? cool
>>>>>
>>>>
>>>>This should be fine.
>>>>
>>>>You might want to protect against future dates.
>>>>
>>>>> well what i was thinking is that my client code would know to use the
>>>>> november table and put the data in the november table (it is all just
>>>>> strings) but I am leaning now towards the TTL option (need to futz
>>>>>with
>>>>>it
>>>>> all more though).  One issue/concern with TTL is when all of a
>>>>>sudden we
>>>>> want to keep things for only 4 months or maybe 8 months and then
>>>>>having
>>>>>to
>>>>> re-TTL trillions of rows =8^( (which is nagging thought in the back
>>>>>of
>>>>>my
>>>>> head about ttls, requirements change)....
>>>>
>>>>This schema attribute is kept at the table level, not per row.  You'll
>>>>have to change the table schema which in 0.90.x hbase means offlining
>>>>table (in 0.92 hbase, there is an online schema edit but needs to be
>>>>enabled and can be problematic in the face of splitting.... more on
>>>>this later).
>>>>
>>>>> That makes sense why a narrow long schema works well then, got it (I
>>>>>am
>>>>>use
>>>>> to Cassandra and do lots of wide column range slices on those columns
>>>>>this
>>>>> is like inverting everything up on myself but the row locks and
>>>>>checkAndPut
>>>>> (and co-processors) hit so many of my uses cases (as Cassandra still
>>>>>does
>>>>> also)
>>>>>
>>>>
>>>>Be careful using hbase row locks.  They are (unofficially -- we need
>>>>to make it official) deprecated.  You can lock yourself out of a
>>>>regionserver if all incoming handlers end up waiting on a particular
>>>>row lock to clear.  Check back in this mailing list for other rowlock
>>>>downsides.
>>>>
>>>>You can column range slices in hbase if you use filters (if you need
>>>>to).
>>>>
>>>>checkAndPut shouldn't care if row is wide or not?
>>>>
>>>>
>>>>> right now I am on 0.90.4 but right now I am going back and forth in
>>>>> changing our hadoop cluster, HBase is the primary driver for that so
>>>>>I
>>>>>am
>>>>> currently wrestling on the decision with upgrading from existing
>>>>>cluster
>>>>> CDH2 to CDH3 or going with MapR ...
>>>>
>>>>Go to CDH3 if you are on CDH2.  Does CDH2 have a working sync?
>>>>(CDH3u3 when it arrives has some nice perf improvements).
>>>>
>>>>> my preference is to run my own version
>>>>> of HBase (like I do with Kafka and Cassandra) I feel I can do this
>>>>>though I
>>>>> am not comfortable with running my own Hadoop build (already
>>>>>overloaded
>>>>> with things).  0.92 is exciting for co-processors too and it is cool
>>>>>system
>>>>> to hack on, maybe I will pull from trunk build and test it out some
>>>>>too.
>>>>>
>>>>
>>>>Don't do hbase trunk.  Do tip of 0.92 branch if you want to hack.
>>>>HBase Trunk is different from 0.92 already and will get even more
>>>>"differenter"; it'll be hard to help you if you are pulling from trunk
>>>>
>>>>St.Ack
>>>>
>>>
>>>
>


Reply via email to