Storing the value on hdfs and using reference to the hdfs location in key value 
is an option. 

> On Oct 19, 2016, at 6:49 PM, big data <[email protected]> wrote:
> 
> actually, there is only one huge value in the hbase cell which large 
> than 100M, maybe it's not a good idea to store such huge value in hbase.
> 
> Any suggestions to store this huge objects?
> 
> 
>> 在 16/10/19 下午8:16, 吴国泉wgq 写道:
>> hi biodata:
>> 
>>         you can try “ scan.setbatch()”  or other filter to limit the number 
>> of column returned.
>> 
>>        This is because: There is a very large row in your table,when you try 
>> to retrieve it, OOM will happen.
>> 
>>         As I can see, There is no other method to solve this problem. 
>> Default,every read operation, hbase will retrieve an entire row if you don’t 
>> limit  the col。
>> 
>>         Don’t insert very large row.  you can delete it .  But don’t forget  
>> trigger the major compaction after you delete the row.
>> 
>> 
>> 
>> 吴国泉   wgq.wu
>> Post: DBA  Hbase
>> Email: [email protected]<http://qunar.com>
>> Tel: 13051697997
>> Adr: 中国电子大厦17层
>> 
>> 
>> 
>> 在 2016年10月19日,下午7:43,big data 
>> <[email protected]<mailto:[email protected]>> 写道:
>> 
>> I've adjusted the jvm xmx in hbase-env.xml, now in hbase shell, count
>> runs well.
>> 
>> But java client still crashes because :
>> 
>> Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol
>> message was too large.  May be malicious.  Use
>> CodedInputStream.setSizeLimit() to increase the size limit.
>> 
>> I've browsed apache's jira, but still don't know how to call
>> CodedInputStream.setSizeLimit()  in client side.
>> 
>> 
>> 在 16/10/19 下午6:38, Jean-Marc Spaggiari 写道:
>> Interesting. Can you bump the client heap size? How much do you have for
>> the client?
>> 
>> JMS
>> 
>> 2016-10-19 3:50 GMT-04:00 big data 
>> <[email protected]<mailto:[email protected]>>:
>> 
>> Dear all,
>> 
>> I've a hbase table, one row has a huge keyvalue, about 100M size.
>> 
>> When I execute count table in hbase shell, hbase crash to bash, and
>> display error like this:
>> 
>> hbase(main):005:0> count 'table', CACHE=>10000
>> java.lang.OutOfMemoryError: Java heap space
>> Dumping heap to java_pid17735.hprof ...
>> Unable to create java_pid17735.hprof: 权限不够
>> #
>> # java.lang.OutOfMemoryError: Java heap space
>> # -XX:OnOutOfMemoryError="kill -9 %p"
>> #   Executing /bin/sh -c "kill -9 17735"...
>> 已杀死
>> 
>> Meanwhile, my java client execute get operation from this table, it
>> hangs for a long time.
>> 
>> How can I adjust some parameters to support huge keyvalue?
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> 安全提示:本邮件非QUNAR内部邮件,请注意保护个人及公司信息安全,如有索取帐号密码等可疑情况请向 secteam发送邮件
> 

Reply via email to