RE: Coprocessor / threading model

2013-01-15 Thread Anoop Sam John
Thanks Andrew. A detailed and useful reply Nothing more needed to explain the anti pattern.. :) -Anoop- From: Andrew Purtell [apurt...@apache.org] Sent: Wednesday, January 16, 2013 12:50 AM To: user@hbase.apache.org Subject: Re: Coprocessor / threadin

Bloom filter based scanner/filter

2013-01-15 Thread David G. Boney
I am building a data cube on top of HBase. All access to the data is by map/reduce jobs. I want to build a scanner where its first matching criteria is based on the set intersection of bloom filters, followed by additional matching criteria specified in the current filter architecture. First, I

Re: Coprocessor / threading model

2013-01-15 Thread Wei Tan
Thanks Andrew for your detailed clarification. Now I understand that in general, the system is subject to CAP theorem. You want good consistency AND latency, then partition tolerance needs to be sacrificed: this is the "local index" approach, i.e., colocate index and data and avoid RPC. Otherwi

Re: User mailing list

2013-01-15 Thread Jean-Marc Spaggiari
Yep, the same way ;) https://www.google.com/search?q=unsubscribe+hbase+issues 2013/1/15, Renato Marroquín Mogrovejo : > Thanks for the quick replay. I am all confused now. It says: > > iss...@hbase.apache.org > > I don't remember subscribing to this list, anyways how do I > unsubscribe? the same

Re: Constructing rowkeys and HBASE-7221

2013-01-15 Thread Doug Meil
Hi there, well, this request for input fell like a thud. :-) But I think perhaps it has to do with the fact that I sent it to the dev-list instead of the user-list, as people that are actively writing HBase itself (devs) need less help with such keybuilding utilities. So one last request for fe

Re: User mailing list

2013-01-15 Thread Renato Marroquín Mogrovejo
Thanks for the quick replay. I am all confused now. It says: iss...@hbase.apache.org I don't remember subscribing to this list, anyways how do I unsubscribe? the same way as the dev one? Thanks again! Renato M. 2013/1/15 Jean-Marc Spaggiari : > Renato, can you check the "To:" from the JIRA yo

Re: User mailing list

2013-01-15 Thread Jean-Marc Spaggiari
Renato, can you check the "To:" from the JIRA your are receiving? Which address is it? What are those JIRAs? Are you "watching" them? JM 2013/1/15, Jean-Daniel Cryans : > I checked a few emails and as far as I can see it's going to issues@ not > user@ > > J-D > > On Tue, Jan 15, 2013 at 11:48 AM,

Re: User mailing list

2013-01-15 Thread Jean-Daniel Cryans
I checked a few emails and as far as I can see it's going to issues@ not user@ J-D On Tue, Jan 15, 2013 at 11:48 AM, Renato Marroquín Mogrovejo wrote: > Hi all, > > I am an HBase user but not a developer. The thing is that I keep on > getting all the JIRA emails from HBase + the user's emails, i

User mailing list

2013-01-15 Thread Renato Marroquín Mogrovejo
Hi all, I am an HBase user but not a developer. The thing is that I keep on getting all the JIRA emails from HBase + the user's emails, isn't that supposed to be only for the dev list? Could anyone with the necessary superpowers take the JIRA emails out of the users list please? Thanks in advance!

Re: Coprocessor / threading model

2013-01-15 Thread Andrew Purtell
HTable is a blocking interface. When a client issues a put, for example, we do not want to return until we can confirm the store has been durably persisted. For client convenience many additional details of remote region invocation are hidden, for example META table lookups for relocated regions, r

Re: Coprocessor / threading model

2013-01-15 Thread Varun Sharma
You should look at the jstack - I think HTablePool is the reason for the large number of threads. Note that HTablePool is a reusable pool HTable(s) and each HTable consists of an ExecutorService containing 1 thread by default. Are you closing the HTable you obtain from HTablePool - if you are not c

Re: Coprocessor / threading model

2013-01-15 Thread Wei Tan
Andrew, could you explain more, why doing cross-table operation is an anti-pattern of using CP? Durability might be an issue, as far as I understand. Thanks, Best Regards, Wei From: Andrew Purtell To: "user@hbase.apache.org" , Date: 01/12/2013 09:39 PM Subject:Re: Coprocess

Re: Maximizing throughput

2013-01-15 Thread Andrew Purtell
Thanks Bryan, really appreciate you letting us know the outcome. I'm sure it will be useful to others. On Tue, Jan 15, 2013 at 9:28 AM, Bryan Keller wrote: > I'll follow up on this in case it is useful to anyone. It seems I was > network I/O limited. The switch I was using was in managed mode wh

Re: Maximizing throughput

2013-01-15 Thread Bryan Keller
I'll follow up on this in case it is useful to anyone. It seems I was network I/O limited. The switch I was using was in managed mode which decreased throughput to 1gbit/sec within the switch, not just on the wire. So with replication set to 2, throughput was about half of the theoretical max on

Re: Hbase hbck -repair slow

2013-01-15 Thread Lashing
Thanks Kevin , it looks like this option is only available in 0.90.7 "Kevin O'dell" 於 2013/1/15 下午11:56 寫道: > If you have only lost your META, I would just use: > > hbase hbck -fixMeta -fixAssignments > > This will rebuild your META > > On Tue, Jan 15, 2013 at 10:53 AM, Lashing wrote: > >>

Re: Hbase tables disappear

2013-01-15 Thread Ibrahim Yakti
The upload finished and sqoop confirmed it was done successfully. I use sqoop to create the table and schema. I've tried the same command and same data on Amazon EMR instance and it worked without any problem. If it's in schema why the count changes when you run the count command a few minutes la

Re: Hbase tables disappear

2013-01-15 Thread Stack
On Tue, Jan 15, 2013 at 1:07 AM, Ibrahim Yakti wrote: > The root directory is set to /var/lib/hbase/ > > The count issue is still exist: > > *Sqoop: * > 13/01/15 08:55:23 INFO mapreduce.ImportJobBase: Retrieved 1754285 records. > > *MySQL:* > +--+ > | count(1) | > +--+ > | 175430

How to upgrade HBase from 0.90.5 to 0.94

2013-01-15 Thread 幻风
Hi, everybody, I want to upgrade a HBase cluster from 0.90.5 to 0.94. The cluster has about 50 nodes and the hadoop version is 0.20.2. Is there an easy way to do it? Shutdown is acceptable and what I really care is the safety of the data. Hope for your suggestions. Thanks, Mickey

Re: Hbase hbck -repair slow

2013-01-15 Thread Kevin O'dell
If you have only lost your META, I would just use: hbase hbck -fixMeta -fixAssignments This will rebuild your META On Tue, Jan 15, 2013 at 10:53 AM, Lashing wrote: > Kevin > Thanks for the update. Actually you can think of the whole meta is > missing , or say deleted. > Can multiple hbck

Re: Hbase hbck -repair slow

2013-01-15 Thread Lashing
Kevin Thanks for the update. Actually you can think of the whole meta is missing , or say deleted. Can multiple hbck repair table run at the same time ? "Kevin O'dell" 於 2013/1/15 下午11:20 寫道: > Lashing, > > Without knowning the exact errors you are seeing it would be hard to make > reco

Re: Slow start of HBase operations with YCSB, possibly because of zookeeper ?

2013-01-15 Thread Akshay Singh
Thanks Samar. You are right YCSB writes data to a single table 'usertable', but I see very slow operations (in order of 1-2 operations/second) even for read/update workload and not only for inserts. So, the region is already split in to multiple RS before I start my transaction workload. And k

Re: Hbase hbck -repair slow

2013-01-15 Thread Kevin O'dell
Lashing, Without knowning the exact errors you are seeing it would be hard to make recommendations. If it is just META that is messed up you can use: hbase hbck -fixMeta -fixAssignments I think(not sure) in that version you can do hbase hbck -repair This will let you target a specific tabl

Hbase hbck -repair slow

2013-01-15 Thread Lashing
We recently suffered meta corrupt issue. We tried to use hbck -repair to fix But the performance is not fast (about 1 region for 1 second). Are there any better way to speed up the process? Btw, My Hbase is 0.90.6. Thanks.

Re: How to upgrade HBase from 0.90.5 to 0.94

2013-01-15 Thread Ted
You can upgrade to 0.92.2 first, followed by upgrade to 0.94.4 Please consider upgrading Hadoop to 1.0.4 as well Thanks On Jan 15, 2013, at 4:47 AM, 幻风 wrote: > Hi, everybody, > > I want to upgrade a HBase cluster from 0.90.5 to 0.94. The cluster has > about 50 nodes and the hadoop version is

Re: RegionServer level co-processor

2013-01-15 Thread Jean-Marc Spaggiari
Why not running a MR job? It will run a task per region. Le 15 janv. 2013 06:38, "Bing Jiang" a écrit : > In my view, if you want to realize Jubtaus, you can use Endpoint. Which > is different from RegionServerObserver. > A client will launch a task about a table, and it will execute an Endpoin

Re: RegionServer level co-processor

2013-01-15 Thread Bing Jiang
In my view, if you want to realize Jubtaus, you can use Endpoint. Which is different from RegionServerObserver. A client will launch a task about a table, and it will execute an Endpoint in each region that belongs to that table. 2013/1/15 蔡斌 > > > > HiWe try to build an online machine le

RegionServer level co-processor

2013-01-15 Thread 蔡斌
HiWe try to build an online machine learning system, like Jubatus(http://www-conf.slac.stanford.edu/xldb2012/talks/xldb2012_wed_LT09_HMakino.pdf), based on HBase. Our training data will be stored in HBase, and algorithms will scan these data to build model. We need a region server level

Re: Hbase tables disappear

2013-01-15 Thread Ibrahim Yakti
The root directory is set to /var/lib/hbase/ The count issue is still exist: *Sqoop: * 13/01/15 08:55:23 INFO mapreduce.ImportJobBase: Retrieved 1754285 records. *MySQL:* +--+ | count(1) | +--+ | 1754303 | +--+ *HBase:* 104210 row(s) in 11.4980 seconds running count c

Re: Slow start of HBase operations with YCSB, possibly because of zookeeper ?

2013-01-15 Thread samar kumar
YCSB would be writing all data to one table.. So initially when the table is small or just created all the writes would go to one RS.. As the table grows the Region is split into different RS. The would allow parallel writes, if the keys are random and could possibly make the writes faster. Samar

Re: Hbase tables disappear

2013-01-15 Thread Ibrahim Yakti
HBase Version: 0.92.1-cdh4.1.2, rUnknown, Thu Nov 1 18:03:50 PDT 2012 Sqoop Version: 1.4.1-cdh4.1.2 Stack: yes, it seems that hbase default root directory is /tmp working to fix this now. I am writing the data for the first time into HBase from MySQL, not overwriting it. -- Ibrahim On Mon,