We're all volunteers here so we don't always have the time to fully understand and plan others' schemas.
In general your questions seemed to be worried about a lot of things that may or may not matter depending on the specifics of your implementation. Without knowing those specifics it is hard to be super definitive. You seem to be very worried about the cost of compactions and retention. Is that because you're having issues now? Short answers: q1: Unless you have a good reason for splitting up into two tables, I'd keep as one. Pros: Easier to understand/better matches intellectual understanding/allows checkAndPuts across both families/data is colocated (server, not disk) on retrieval if you want to work with both groups simultaneously using get, MR, etc. Con: There will be some extra merge/flush activity if the two columns grow at substantially different rates. q2: 365*1000 regions is problematic (if that is what you're suggesting). Even with HFilev2 and partially loaded multi-level indexes, there is still quite a bit of overhead per region. I pointed you at the Jive thing in part since hashing that value as a bucket seems a lot more reasonable. Additional Random idea: if you know retention policy on insert and your data is immutable post insertion, consider shifting the insert timestamp and maintain a single ttl. Would require more client side code but would allow configurable ttls while utilizing existing HBase infrastructure. q3: Sounds like you're prematurely optimizing here. Maybe others would disagree. I'd use ttl until you find that isn't performant enough. The tension between flexibility and speed is clear here. I'd say you either need to pick specific ttls and optimize for that scenario via region pruning (e.g. separate tables for each ttl type) or you need to use a more general approach that leverages the per value ttl and compaction methodology. There is enough operational work managing an HBase/HDFS cluster without having to worry about specialized region management. Jacques On Wed, Oct 3, 2012 at 11:31 AM, Karthikeyan Muthukumarasamy < mkarthik.h...@gmail.com> wrote: > Hi Jacques, > Thanks for the response! > Yes, I have seen the video before. It suggets usage of TTL based retention > implementation. In their usecase, Jive has a fixed retention say 3 months > and so they can pre-create regions for so many buckets, their bucket id is > DAY_OF_YEAR%retention_in_days. But, in our usecase, the retention period is > configurable, so pre-creationg regions based on retention will not work. > Thats why we went for MMDD based buckets which is immune to retention > period changes. > Now that you know that Ive gone through that video from Jive, I would > request you to re-read my specific questions and share your suggestions. > Thanks & Regards > MK > > > > On Wed, Oct 3, 2012 at 11:51 PM, Jacques <whs...@gmail.com> wrote: > > > I would suggest you watch this video: > > > > > http://www.cloudera.com/resource/video-hbasecon-2012-real-performance-gains-with-real-time-data/ > > > > The jive guys solved a lot of the problems you're talking about and > discuss > > it in that case study. > > > > > > > > On Wed, Oct 3, 2012 at 6:27 AM, Karthikeyan Muthukumarasamy < > > mkarthik.h...@gmail.com> wrote: > > > > > Hi, > > > Our usecase is as follows: > > > We have time series data continuously flowing into the system and has > to > > be > > > stored in HBase. > > > Subscriber Mobile Number (a.k.a MSISDN) is the primary identifier based > > on > > > which data is stored and later retrieved. > > > There are two sets of parameters that get stored in every record in > > HBase, > > > lets call them group1 and group2. The number of records that would have > > > group1 parameters would be approx. 6 per day and the same for group2 > > > parameters is approx. 1 per 3 days (their cardinality is different). > > > > > > Typically, the retention policy for group1 parameters is 3 months and > for > > > group2 parameters is 1 year. The read-pattern is as follows: An online > > > query would ask for records matching an MSISDN for a given date range, > > and > > > the system needs to respond with all available data (both from group1 > and > > > group2) satifying the MSISDN and data range filters. > > > > > > Question1: > > > Alternative1: Create a single table with G1 and G2 as two column > > families. > > > Alternative2: Create two tables one for each group > > > Which is the better alternative and what are the pros and cons? > > > > > > > > > Question2: > > > To achieve max. distribution during write and reasonable complexity > > during > > > read, we decided on the following row key design: > > > <last 3 digits of MSISDN>,<MMDD>,<full MSISDN> > > > We will manually pre-split regions for the table based on the <last 3 > > > digits of MSISDN>,<MMDD> part of row key > > > So there are 1000 (from 3 digits of MSISDN) * 365 (from MMDD) buckets > > that > > > would translate to as many regions > > > In this case, when retention is configured as < 1 year, the design > looks > > > optimal > > > When retention is configured > 1 year, one region might store data for > > more > > > than 1 day (feb 1 of 2012 and also feb 1 of 2013), which means more > data > > is > > > to be handled by hbase during compactions and read. > > > An alternative Key design, which does not have the above disadvantage > is: > > > <last 3 digits of MSISDN>,<YYYYMMDD>,<full MSISDN> > > > this way, in one region, there will be only 1 days data at any point, > > > regardless of retention > > > What are other pros & cons of the two key designs? > > > > > > Question3: > > > In our usecase, delete happens only based on retention policy, where > one > > > days full data has to be deleted when rention period is crossed (for > eg, > > if > > > retention is 30 days, on Apr 1 all the data for Mar 1 is deleted) > > > What is the most optimal way to implement this retention policy? > > > Alternative 1: TTL for column famil is configured and we leave it to > > HBase > > > to delete data during major compaction, but we are not sure of the cost > > of > > > this major compaction happening in all regions at same time > > > Alternative 2: Through key design logic mentioned before, if we ensure > > data > > > for one day goes into one set of regions, can we use HBase APIs like > > > HFileArchiver to programatically archive and drop regions? > > > > > > Thanks & Regards > > > MK > > > > > >