[
https://issues.apache.org/jira/browse/HBASE-17340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Jean-Marc Spaggiari resolved HBASE-17340.
-----------------------------------------
Resolution: Invalid
> Workload management with Hbase
> -------------------------------
>
> Key: HBASE-17340
> URL: https://issues.apache.org/jira/browse/HBASE-17340
> Project: HBase
> Issue Type: Brainstorming
> Components: Client
> Reporter: vincent eiger
>
> Hi guys, I am project manager in a bank where we work on a POC. We try to
> test the first step of our Datalake with:
> - a first layer for landing area : store files without transformation in HDFS
> (nowdays 6 files)
> -a second layer in Hbase where we upload with one spark job a unique Hbase
> table.
> We know that with capacity scheduler in Yarn we could guarantee resources
> consumption (memory, CPU) for spark jobs.
> My question is :
> - how i could guarantee resources for my Hbase client : I mean one
> application which request my table with Hbase native client and at the same
> time some other application, batch run spark jobs... I want ensure all the
> time at least 30% of capactity to my application (which request my Hbase
> table) in order to guarantee for instance 50 request at the same time with
> less than 200 ms per user.
> Is Slider the good answer ?
> Thanks a lot for answers,
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)