how to transfer app from hbase0.90 to hbase 0.94

2012-12-20 Thread tgh
Hi I used hbase0.90 and hadoop0.20 and use hbaseAPI to store some data, and it works , and now I turn to use hbase 0.94 and hadoop1.1.1, and there is no ERROR in hbase logs, it seems OK, but , I try to run my program , ERROR arise, I want to know when we transfer f

Re: recommended nodes

2012-12-20 Thread Adrien Mogenet
Maybe you should give a little more information about your RAID controller (write back / write through ?) and the underlying filesystem (ext3 ? blocksize ?). Very interesting benchmark and discussion by the way :-) On Thu, Dec 20, 2012 at 11:07 PM, Jean-Marc Spaggiari < jean-m...@spaggiari.org>

Re: recommended nodes

2012-12-20 Thread Jean-Marc Spaggiari
I did the test with a 2GB file... So read and write were spread over the 2 drives for RAID0. Those test were to give an overall idea of the performances vs CPU usage etc. and you might need to adjust them based on the way it's configured on your system. I don't know how RAID0 is managing small fi

Re: recommended nodes

2012-12-20 Thread Varun Sharma
Hmm, I thought that RAID0 simply stripes across all disks. So if you got 4 disks - an HFile block for example could get striped across 4 disks. So to read that block, you would need all 4 of them to seek so that you could read all 4 stripes for that HFile block. This could make things as slow as th

Re: Ihbase creating secondary indices

2012-12-20 Thread Stack
That project has not had a commit in over two years. HBase has had a few since then so the project has probably gone stale. St.Ack On Thu, Dec 20, 2012 at 6:04 AM, Shengjie Min wrote: > Hi, > > I know Coprocessor 2ndary indices is still being developed. Has anybody > tried using Ihbase to crea

Re: recommended nodes

2012-12-20 Thread Jean-Marc Spaggiari
Hi Varun, The hard drivers I used are now used on the hadoop/hbase cluster, but they was clear and formated for the tests I did. The computer where I run those tests was one of the region servers. It was re-installed to be very clear, and it's now running a datanode and a RS. Regarding RAID, I th

Re: recommended nodes

2012-12-20 Thread Varun Sharma
Hi Jean, Very interesting benchmark - how are these numbers arrived at ? Is this on a real hbase cluster ? To me, it felt kind of counter intuitive that RAID0 beats JBOD on random seeks because with RAID0 all disks need to seek at the same time and the performance should basically be as bad as the

Re: Is it necessary to set MD5 on rowkey?

2012-12-20 Thread Michael Segel
Nick, Yes there is an implied definition of the term 'salting' which to those with a CS or Software Engineering background will take to heart. However it goes beyond this definition. Per Lars and Alex, they are talking about bucketing the data. Again this is not a good idea. As you point ou

Re: Customizing hbase shell's table formatting

2012-12-20 Thread Adam Phelps
Thanks for the info. I see that info now that I look at the current version's scan command, its just not in the 0.90.6 we're currently on but should be updating before long. This sounds like the formatting is for column only, our rowkeys themselves are serialized, is there no way to specify a for

Re: Can Hbase run on NFS cluster

2012-12-20 Thread Andrew Purtell
I don't think this makes much sense. All of your cluster talks to one filer. Surely it will be the bottleneck of anything but a trivial test. Just use the filer. On Wed, Dec 19, 2012 at 10:04 PM, Ling Kun wrote: > Dear all, >I am doing some test on Hbase running on a NFS cluster. Does anyone

Re: Many scanner opening

2012-12-20 Thread lars hofhansl
Cool. You probably made it less likely that your scanners will scan the same HFile in parallel. -- Lars From: Eugeny Morozov To: user@hbase.apache.org; lars hofhansl Sent: Thursday, December 20, 2012 2:32 AM Subject: Re: Many scanner opening Lars, Cool

Re: Is it necessary to set MD5 on rowkey?

2012-12-20 Thread Nick Dimiduk
I think there's some hair-splitting going on here. The term "salting," by strict definition [0] from the cryptographic context, means the introduction of randomness to produce a one-way encoding of a value. The technique for rowkey design described here does not include the introduction of said ran

Re: HBASE - select distinct query against the rowkey

2012-12-20 Thread Michael Segel
There is no concept of transaction in the NoSQL world. At least not in HBase. All writes are atomic. Note that you *could* hold a lock, however, not really a good idea for a client to hold a lock. Don't know if its really a problem though... HTH -Mike On Dec 20, 2012, at 10:08 AM, Shengji

Re: HBASE - select distinct query against the rowkey

2012-12-20 Thread Shengjie Min
Thanks Michael, >Not sure why you have timestamp in the key... assuming that message id would be incremented therefore rows would be in time order anyways. I will need to do query like give me the message from timestamp1 to timestamp2. >You will want to use a separate table. That's what I though

Re: HBASE - select distinct query against the rowkey

2012-12-20 Thread Michael Segel
Not sure why you have timestamp in the key... assuming that message id would be incremented therefore rows would be in time order anyways. But to answer your question... You will want to use a separate table. In both instances you will end up doing a full table scan, however the number of row

Re: HBase Wire Protocol

2012-12-20 Thread Matan Amir
Thanks again St.Ack. Will do. I have to say that the comments alone in asynchbase are worth the price of admission! Very entertaining :) On Wed, Dec 19, 2012 at 11:31 PM, Stack wrote: > On Wed, Dec 19, 2012 at 8:53 PM, Matan Amir wrote: > > > Thanks St.Ack, > > > > Funny enough, since we us

HBASE - select distinct query against the rowkey

2012-12-20 Thread Shengjie Min
I have a hbase table called "users", rowkey consists of three parts: 1. userid 2. messageid 3. timestamp rowkey looks like: ${userid}_${messageid}_${timestamp} Given I can hash the userid and make the length of the field fixed, is there anyway I can do a query like SQL query: select di

Ihbase creating secondary indices

2012-12-20 Thread Shengjie Min
Hi, I know Coprocessor 2ndary indices is still being developed. Has anybody tried using Ihbase to create 2ndary indices: https://issues.apache.org/jira/browse/HBASE-2037. https://github.com/ykulbak/ihbase -- All the best, Shengjie Min

Re: Is it necessary to set MD5 on rowkey?

2012-12-20 Thread Michael Segel
Lars, Ok... he's talking about buckets. So when you have N buckets, what is the least number of get()s do you need to fetch the single row? (Hint: The answer is N) How many scans? (N again) Do you disagree? On Dec 19, 2012, at 8:06 PM, lars hofhansl wrote: > Mike, please think about wha

Re: Many scanner opening

2012-12-20 Thread Michael Segel
I'd suggest looking in to a schema design change. On Dec 18, 2012, at 2:01 AM, Eugeny Morozov wrote: > Hello! > > We faced an issue recently that the more map tasks are completed, the > longer it takes to complete one more map task. > > In our architecture we have two scanners to read the tabl

help! code for hbase 0.90 ries ERROR in hbase 0.94

2012-12-20 Thread tgh
Meanwhile, I use hbase0.90 and hadoop1.1.1 And I want to know if we use these two version together, How to configuration these two, I wander if there some ERROR in my configuration, -configuration files in hbase/conf--- # See http://wiki.apache.org/hadoop/GangliaMetric

答复: code for hbase 0.90 ries ERROR in hbase 0.94

2012-12-20 Thread tgh
Thank you for your reply -code- import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable;

Re: code for hbase 0.90 ries ERROR in hbase 0.94

2012-12-20 Thread Lichao (andylee)
Hi guanhua: From the error , Your Code Configuration class shoule be import org.apache.hadoop.conf.Configuration ,not org.apache.commons.configuration.Configuration. Please check it. -邮件原件- 发件人: tgh [mailto:guanhua.t...@ia.ac.cn] 发送时间: 2012年12月20日 18:17 收件人: user@hbase.apache.org

答复: code for hbase 0.90 ries ERROR in hbase 0.94

2012-12-20 Thread Lichao (andylee)
Hi guanhua: From the error , Your Code Configuration class shoule be import org.apache.hadoop.conf.Configuration ,not org.apache.commons.configuration.Configuration. Please check it. -邮件原件- 发件人: tgh [mailto:guanhua.t...@ia.ac.cn] 发送时间: 2012年12月20日 18:17 收件人: user@hbase.apache.or

Re: Many scanner opening

2012-12-20 Thread Eugeny Morozov
Lars, Cool stuff! Thanks a lot! I'm not sure I can apply the patch, cause we're using CDH-4.1.1, but increasing size of internal scanner does the trick - decreased number of scanners. At least temporarily it's good enough. Thanks! On Wed, Dec 19, 2012 at 6:23 AM, lars hofhansl wrote: > You mig

code for hbase 0.90 ries ERROR in hbase 0.94

2012-12-20 Thread tgh
Hi I used hbase0.90,and now turn to use hbase 0.94 , and my program is ERROR, Could you help me --- my code --- Configuration hbase_config = new Configuration(); hbase_config.set("hbase.zookeeper.quorum", sms_zook

Re: CF still contains data after deletion

2012-12-20 Thread Roger Miller
Thans. So I guess we have to wait until Cloudera includes a new HBase version. Von: anil gupta An: user@hbase.apache.org; Roger Miller Gesendet: 9:48 Donnerstag, 20.Dezember 2012 Betreff: Re: CF still contains data after deletion Hi Roger, I think you are

Re: CF still contains data after deletion

2012-12-20 Thread anil gupta
Hi Roger, I think you are hitting: https://issues.apache.org/jira/browse/HBASE-6564 The above jira was fixed in HBase0.94.1 and later releases. CDH4.1.2 has HBase0.92.1 so it doesn't contains that fix. HTH, Anil Gupta On Thu, Dec 20, 2012 at 12:43 AM, Roger Miller wrote: > Hello, > > we are cu

CF still contains data after deletion

2012-12-20 Thread Roger Miller
Hello, we are currently using Cloudera CDH4.1.2 and have the following problem: Assume we have a column family in HBase that already contains data that we want to get rid of. In past versions of HBase it was possible to delete this CF and re-create it and it contained no data afterwards. Now in