On Tue, Aug 8, 2017 at 12:01 PM, Yu Li <car...@gmail.com> wrote:
...

> Maybe we could open several umbrellas in JIRA to follow up the
> implement-able topics and do some prioritizing? Thanks.
>

Agree.

Let me start up the DISCUSS suggested by Ashish; it just came up again on
the tail of an issue....

St.Ack




> Best Regards,
> Yu
>
> On 8 August 2017 at 10:54, ashish singhi <ashish.sin...@huawei.com> wrote:
>
> > Great write up, Stack. Covering everything what we all discussed.
> > It was very nice meeting you all and hope we can continue this HBaseCon
> > Asia.
> >
> > Regards,
> > Ashish
> >
> > From: saint....@gmail.com [mailto:saint....@gmail.com] On Behalf Of
> Stack
> > Sent: 08 August 2017 00:07
> > To: HBase Dev List <dev@hbase.apache.org>
> > Subject: Notes from dev meetup in Shenzhen, August 5th, 2017
> >
> > At fancy Huawei headquarters, 10:00-12:00AM or so (with nice coffee and
> > fancy little cake squares provided about half way through the session).
> >
> > For list of attendees, see picture at end of this email.
> >
> > Discussion was mostly in Chinese with about 25% in English plus some
> > gracious sideline translation so the below is patchy. Hopefully you get
> the
> > gist.
> >
> > For client-side scanner going against hfiles directly; is there a means
> of
> > being able to pass the permissions from hbase to hdfs?
> >
> > Issues w/ the hbase 99th percentile were brought up. "DynamoDB can do
> > 10ms". How to do better?
> >
> > SSD is not enough.
> >
> > GC messes us up.
> >
> > Will the Distributed Log Replay come back to help improve MTTR? We could
> > redo on new ProcedureV2 basis. ZK timeout is the biggest issue. Do as we
> > used to and just rely on the regionserver heartbeating...
> >
> > Read replica helps w/ MTTR.
> >
> > Ratis incubator project to do a quorum based hbase?
> >
> > Digression on licensing issues around fb wangle and folly.
> >
> > Redo of hbase but quorum based would be another project altogether.
> >
> > Decided to go around the table to talk about concerns and what people are
> > working on.
> >
> > Jieshan wondered what could be done to improve OLAP over hbase.
> >
> > Client side scanner was brought up again as means of skipping RS overhead
> > and doing better OLAP.
> >
> > Have HBase compact to parquet files. Query parquet and hbase.
> >
> > At Huawei, they are using 1.0 hbase. Most problems are assignment. They
> > have .5M regions. RIT is a killer. Double assignment issues. And RIT.
> They
> > run their own services. Suggested they upgrade to get fixes at least.
> Then
> > 2.0.
> >
> > Will HBase federate like HDFS? Can Master handle load at large scale? It
> > needs to do federation too?
> >
> > Anyone using Bulk loaded replication? (Yes, it just works so no one talks
> > about it...)
> >
> > Request that fixes be backported to all active branches, not just most
> > current.
> >
> > Andrew was good at backporting... not all RMs are.
> >
> > Too many branches. What should we do?
> >
> > Proliferation of branches makes for too much work.
> >
> > Need to cleanup bugs in 1.3. Make it stable release now.
> >
> > Lets do more active EOL'ing of branches. 1.1?.
> >
> > Hubert asked if we can have clusters where RS are differently capable?
> > i.e. several generations of HW all running in the same cluster.
> >
> > What if fat server goes down.
> >
> > Balancer could take of it all. RS Capacity. Balancer can take it into
> > account.
> > Regionserver labels like YARN labels. Characteristics.
> >
> > Or run it all in docker when heterogeneous cluster. The K8 talk from day
> > before was mentioned; we should all look at being able to deploy in k8
> and
> > docker.
> >
> > Lets put out kubernetes blog...(Doing).
> >
> > Alibaba looking at HBase as native YARN app.
> >
> > i/o is hard even when containers.
> >
> > Use autoscaler of K8 when heavy user.
> >
> > Limit i/o use w/ CP. Throttle.
> >
> > Spark and client-side scanner came up again.
> >
> > Snapshot input format in spark.
> >
> > HBase federation came up again. jd.com<http://jd.com> talking of 3k to
> 4k
> > nodes in a cluster. Millions of regions. Region assignment is messing
> them
> > up.
> >
> > Maybe federation is good idea? Argument that it is too much operational
> > conplexity. Can we fix master load w/ splittable meta, etc?
> >
> > Was brought up that even w/ 100s of RS there is scale issue, nvm
> thousands.
> >
> > Alibaba talked about disaster recovery. Described issue where HDFS has
> > fencing problem during an upgrade. There was no active NN. All RS went
> down.
> > ZK is another POF. If ZK is not available. Operators were being asked how
> > much longer the cluster was going to be down but they could not answer
> the
> > question. No indicators from HBase on how much longer it will be down or
> > how many WALs its processed and how many more to go. Operator unable to
> > tell his org how long it would be before it all came back on line. Should
> > say how many regions are online and how many more to do.
> >
> > Alibaba use SQL to lower cost. HBase API is low-level. Row-key
> > construction is tricky. New users make common mistakes. If you don't do
> > schema right, high-performance is difficult.
> >
> > Alibaba are using a subset of Phoenix... simple sql only; throws
> > exceptions if user tries to do joins, etc.., anything but basic ops.
> >
> > HareQL is using hive for meta store.  Don't have data typing in hbase.
> >
> > HareQL could perhaps contribute some piece... or a module in hbase to
> > sql... From phoenix?
> >
> > Secondary index.
> >
> > Client is complicated in phoenix. Was suggested thin client just does
> > parse... and then offload to server for optimization and execution.
> >
> > Then secondary index. Need transaction engine. Consistency of secondary
> > index.
> >
> > We adjourned.
> >
> > Your dodgy secretary,
> > St.Ack
> > P.S. Please add to this base set of notes if I missed anything.
> >
> >
> >
> >
>

Reply via email to