Hi
I used hbase0.90 and hadoop0.20 and use hbaseAPI to store some data,
and it works ,
and now I turn to use hbase 0.94 and hadoop1.1.1, and there is no
ERROR in hbase logs, it seems OK,
but , I try to run my program , ERROR arise,
I want to know when we transfer f
Maybe you should give a little more information about your RAID controller
(write back / write through ?) and the underlying filesystem (ext3 ?
blocksize ?).
Very interesting benchmark and discussion by the way :-)
On Thu, Dec 20, 2012 at 11:07 PM, Jean-Marc Spaggiari <
jean-m...@spaggiari.org>
I did the test with a 2GB file... So read and write were spread over the 2
drives for RAID0.
Those test were to give an overall idea of the performances vs CPU usage
etc. and you might need to adjust them based on the way it's configured on
your system.
I don't know how RAID0 is managing small fi
Hmm, I thought that RAID0 simply stripes across all disks. So if you got 4
disks - an HFile block for example could get striped across 4 disks. So to
read that block, you would need all 4 of them to seek so that you could
read all 4 stripes for that HFile block. This could make things as slow as
th
That project has not had a commit in over two years. HBase has had a few
since then so the project has probably gone stale.
St.Ack
On Thu, Dec 20, 2012 at 6:04 AM, Shengjie Min wrote:
> Hi,
>
> I know Coprocessor 2ndary indices is still being developed. Has anybody
> tried using Ihbase to crea
Hi Varun,
The hard drivers I used are now used on the hadoop/hbase cluster, but they
was clear and formated for the tests I did. The computer where I run those
tests was one of the region servers. It was re-installed to be very clear,
and it's now running a datanode and a RS.
Regarding RAID, I th
Hi Jean,
Very interesting benchmark - how are these numbers arrived at ? Is this on
a real hbase cluster ? To me, it felt kind of counter intuitive that RAID0
beats JBOD on random seeks because with RAID0 all disks need to seek at the
same time and the performance should basically be as bad as the
Nick,
Yes there is an implied definition of the term 'salting' which to those with a
CS or Software Engineering background will take to heart.
However it goes beyond this definition.
Per Lars and Alex, they are talking about bucketing the data. Again this is
not a good idea.
As you point ou
Thanks for the info. I see that info now that I look at the current
version's scan command, its just not in the 0.90.6 we're currently on
but should be updating before long.
This sounds like the formatting is for column only, our rowkeys
themselves are serialized, is there no way to specify a for
I don't think this makes much sense. All of your cluster talks to one
filer. Surely it will be the bottleneck of anything but a trivial test.
Just use the filer.
On Wed, Dec 19, 2012 at 10:04 PM, Ling Kun wrote:
> Dear all,
>I am doing some test on Hbase running on a NFS cluster. Does anyone
Cool.
You probably made it less likely that your scanners will scan the same HFile in
parallel.
-- Lars
From: Eugeny Morozov
To: user@hbase.apache.org; lars hofhansl
Sent: Thursday, December 20, 2012 2:32 AM
Subject: Re: Many scanner opening
Lars,
Cool
I think there's some hair-splitting going on here. The term "salting," by
strict definition [0] from the cryptographic context, means the
introduction of randomness to produce a one-way encoding of a value. The
technique for rowkey design described here does not include the
introduction of said ran
There is no concept of transaction in the NoSQL world. At least not in HBase.
All writes are atomic. Note that you *could* hold a lock, however, not really a
good idea for a client to hold a lock.
Don't know if its really a problem though...
HTH
-Mike
On Dec 20, 2012, at 10:08 AM, Shengji
Thanks Michael,
>Not sure why you have timestamp in the key... assuming that message id
would be incremented therefore rows would be in time order anyways.
I will need to do query like give me the message from timestamp1 to
timestamp2.
>You will want to use a separate table.
That's what I though
Not sure why you have timestamp in the key... assuming that message id would be
incremented therefore rows would be in time order anyways.
But to answer your question...
You will want to use a separate table.
In both instances you will end up doing a full table scan, however the number
of row
Thanks again St.Ack. Will do.
I have to say that the comments alone in asynchbase are worth the price of
admission! Very entertaining :)
On Wed, Dec 19, 2012 at 11:31 PM, Stack wrote:
> On Wed, Dec 19, 2012 at 8:53 PM, Matan Amir wrote:
>
> > Thanks St.Ack,
> >
> > Funny enough, since we us
I have a hbase table called "users", rowkey consists of three parts:
1. userid
2. messageid
3. timestamp
rowkey looks like: ${userid}_${messageid}_${timestamp}
Given I can hash the userid and make the length of the field fixed, is
there anyway I can do a query like SQL query:
select di
Hi,
I know Coprocessor 2ndary indices is still being developed. Has anybody
tried using Ihbase to create 2ndary indices:
https://issues.apache.org/jira/browse/HBASE-2037.
https://github.com/ykulbak/ihbase
--
All the best,
Shengjie Min
Lars,
Ok... he's talking about buckets.
So when you have N buckets, what is the least number of get()s do you need to
fetch the single row?
(Hint: The answer is N)
How many scans? (N again)
Do you disagree?
On Dec 19, 2012, at 8:06 PM, lars hofhansl wrote:
> Mike, please think about wha
I'd suggest looking in to a schema design change.
On Dec 18, 2012, at 2:01 AM, Eugeny Morozov wrote:
> Hello!
>
> We faced an issue recently that the more map tasks are completed, the
> longer it takes to complete one more map task.
>
> In our architecture we have two scanners to read the tabl
Meanwhile, I use hbase0.90 and hadoop1.1.1
And I want to know if we use these two version together,
How to configuration these two,
I wander if there some ERROR in my configuration,
-configuration files in hbase/conf---
# See http://wiki.apache.org/hadoop/GangliaMetric
Thank you for your reply
-code-
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.HTable;
Hi guanhua:
From the error , Your Code Configuration class shoule be import
org.apache.hadoop.conf.Configuration ,not
org.apache.commons.configuration.Configuration.
Please check it.
-邮件原件-
发件人: tgh [mailto:guanhua.t...@ia.ac.cn]
发送时间: 2012年12月20日 18:17
收件人: user@hbase.apache.org
Hi guanhua:
From the error , Your Code Configuration class shoule be import
org.apache.hadoop.conf.Configuration ,not
org.apache.commons.configuration.Configuration.
Please check it.
-邮件原件-
发件人: tgh [mailto:guanhua.t...@ia.ac.cn]
发送时间: 2012年12月20日 18:17
收件人: user@hbase.apache.or
Lars,
Cool stuff! Thanks a lot! I'm not sure I can apply the patch, cause we're
using CDH-4.1.1, but increasing size of internal scanner does the trick -
decreased number of scanners.
At least temporarily it's good enough.
Thanks!
On Wed, Dec 19, 2012 at 6:23 AM, lars hofhansl wrote:
> You mig
Hi
I used hbase0.90,and now turn to use hbase 0.94 , and my program is
ERROR,
Could you help me
--- my code ---
Configuration hbase_config = new Configuration();
hbase_config.set("hbase.zookeeper.quorum",
sms_zook
Thans. So I guess we have to wait until Cloudera includes a new HBase version.
Von: anil gupta
An: user@hbase.apache.org; Roger Miller
Gesendet: 9:48 Donnerstag, 20.Dezember 2012
Betreff: Re: CF still contains data after deletion
Hi Roger,
I think you are
Hi Roger,
I think you are hitting: https://issues.apache.org/jira/browse/HBASE-6564
The above jira was fixed in HBase0.94.1 and later releases. CDH4.1.2 has
HBase0.92.1 so it doesn't contains that fix.
HTH,
Anil Gupta
On Thu, Dec 20, 2012 at 12:43 AM, Roger Miller wrote:
> Hello,
>
> we are cu
Hello,
we are currently using Cloudera CDH4.1.2 and have the following problem:
Assume we have a column family in HBase that already contains data that we want
to get rid of. In past versions of HBase it was possible to delete this CF and
re-create it and it contained no data afterwards. Now in
29 matches
Mail list logo