Looking at source of IncreasingToUpperBoundRegionSplitPolicy, I don't see
other parameters being used.

FYI

On Sun, Aug 28, 2016 at 5:58 PM, yeshwanth kumar <yeshwant...@gmail.com>
wrote:

> Hi Ted,
>
> thanks for the reply,
>
> i couldn't find the hbase.increasing.policy.initial.size in hbase conf,
> we haven't changed that value.
>
> so that means intial regionsize should be 2 GB, but the region size is
> 2.5TB
> i can manually split the regions, but trying to figure out the root cause.
> any other conf properties causing this behavior?
>
> please let me know,
>
> Thanks,
> Yeshwanth
>
>
>
> On Fri, Aug 26, 2016 at 5:41 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
> > From IncreasingToUpperBoundRegionSplitPolicy#configureForRegion():
> >
> >     initialSize = conf.getLong("hbase.increasing.policy.initial.size",
> > -1);
> >
> > ...
> >
> >     if (initialSize <= 0) {
> >
> >       initialSize = 2 * conf.getLong(HConstants.
> > HREGION_MEMSTORE_FLUSH_SIZE,
> >
> >                                      HTableDescriptor.
> > DEFAULT_MEMSTORE_FLUSH_SIZE);
> >
> > If you haven't changed the value for
> > "hbase.increasing.policy.initial.size", the last two lines should have
> > been
> > executed.
> >
> > initialSize would be 2GB in that case according to the config you listed.
> >
> >
> > FYI
> >
> > On Fri, Aug 26, 2016 at 3:23 PM, yeshwanth kumar <yeshwant...@gmail.com>
> > wrote:
> >
> > > Hi we are using  CDH 5.7 HBase 1.2
> > >
> > > we are doing a performance testing over HBase through regular Load,
> which
> > > has 4 Region Servers.
> > >
> > > Input Data is compressed binary files around 2TB, which we process and
> > > write as Key-Value pairs to HBase.
> > > the output data size in  HBase is almost 4 times around 8TB, because we
> > are
> > > writing as text.
> > > this process is a Map-Reduce Job,
> > >
> > > when we are doing the load, we observed there's a lot of GC happening
> on
> > > Region Server's ,so we changed couple of  parameters to decrease the GC
> > > time.
> > >
> > > we increased the flush size to 128MB to 1 GB and compactionThreshold to
> > 50
> > > and  regionserver.maxlogs to 42
> > > following are the configuration we changed from default.
> > >
> > >
> > > hbase.hregion.memstore.flush.size = 1 GB
> > > hbase.hstore.max.filesize=10GB
> > > hbase.hregion.preclose.flush.size= 50 MB
> > >
> > > hbase.hstore.compactionThreshold=50
> > > hbase.regionserver.maxlogs=42
> > >
> > > after the load, we observed that HBase table has only 4 regions with
> each
> > > of size around 2.5 TB
> > >
> > > i am trying to understand, what configuration parameter caused this
> > issue.
> > >
> > > i was going through this article
> > > http://hortonworks.com/blog/apache-hbase-region-splitting-and-merging/
> > >
> > > Region split policy in our HBase is
> > > org.apache.hadoop.hbase.regionserver.IncreasingToUpperBoundRegionSp
> > > litPolicy
> > > according to Region Split policy, Region Server should create regions
> > when
> > > the region size limit is exceeded.
> > > can some one explain me the root cause.
> > >
> > >
> > > Thanks,
> > > Yeshwanth
> > >
> >
>

Reply via email to