We dont want to invest into another DB like Dynamo, Cassandra and Already
are in the Hadoop Stack. Managing another DB would be a pain. Why HBase
over RDMS, is because we call HBase via Spark Streaming to lookup the keys.
Manish
On Mon, Aug 29, 2016 at 1:47 PM, Dima Spivak
Hey Manish,
Just to ask the naive question, why use HBase if the data fits into such a
small table?
On Sunday, August 28, 2016, Manish Maheshwari wrote:
> Hi,
>
> We have a scenario where HBase is used like a Key Value Database to map
> Keys to Regions. We have over 5
Hi,
We have a scenario where HBase is used like a Key Value Database to map
Keys to Regions. We have over 5 Million Keys, but the table size is less
than 7 GB. The read volume is pretty high - About 50x of the put/delete
volume. This causes hot spotting on the Data Node and the region is not
Ok, thanks for the link Ted
On Aug 29, 2016 9:54 AM, "Ted Yu" wrote:
> For hortonworks product(s), consider raising question on
> https://community.hortonworks.com
>
> FYI
>
> On Sun, Aug 28, 2016 at 6:45 PM, spats wrote:
>
> > Regarding hbase
Looking at source of IncreasingToUpperBoundRegionSplitPolicy, I don't see
other parameters being used.
FYI
On Sun, Aug 28, 2016 at 5:58 PM, yeshwanth kumar
wrote:
> Hi Ted,
>
> thanks for the reply,
>
> i couldn't find the hbase.increasing.policy.initial.size in hbase
For hortonworks product(s), consider raising question on
https://community.hortonworks.com
FYI
On Sun, Aug 28, 2016 at 6:45 PM, spats wrote:
> Regarding hbase connector by hortonworks
> https://github.com/hortonworks-spark/shc, it would be great if someone can
> answer
Regarding hbase connector by hortonworks
https://github.com/hortonworks-spark/shc, it would be great if someone can
answer these
1. What versions of Hbase & Spark expected? I could not run examples
provided using spark 1.6.0 & hbase 1.2.0
2. I get error when i run example provided here , any
Hi Ted,
thanks for the reply,
i couldn't find the hbase.increasing.policy.initial.size in hbase conf,
we haven't changed that value.
so that means intial regionsize should be 2 GB, but the region size is 2.5TB
i can manually split the regions, but trying to figure out the root cause.
any other
And what kind of performance do you see vs. what you expect to see? How big
is your cluster in production/how much total data will you be storing in
production?
On Sunday, August 28, 2016, Manjeet Singh
wrote:
> Hi
> I performed this testing on 2 node cluster where
Hi
I performed this testing on 2 node cluster where its i7 core processor with
16 gb ram 8 core on each node.
I have very frequent get put operation on hbase using spark streaming and
sql where we r aggregate data on spark group and saving it to hbase
Can you give us more specifics about what
10 matches
Mail list logo