Thanks Andrew. A detailed and useful reply Nothing more needed to explain
the anti pattern.. :)
-Anoop-
From: Andrew Purtell [apurt...@apache.org]
Sent: Wednesday, January 16, 2013 12:50 AM
To: user@hbase.apache.org
Subject: Re: Coprocessor / threadin
I am building a data cube on top of HBase. All access to the data is by
map/reduce jobs. I want to build a scanner where its first matching criteria is
based on the set intersection of bloom filters, followed by additional matching
criteria specified in the current filter architecture. First, I
Thanks Andrew for your detailed clarification.
Now I understand that in general, the system is subject to CAP theorem.
You want good consistency AND latency, then partition tolerance needs to
be sacrificed: this is the "local index" approach, i.e., colocate index
and data and avoid RPC.
Otherwi
Yep, the same way ;)
https://www.google.com/search?q=unsubscribe+hbase+issues
2013/1/15, Renato Marroquín Mogrovejo :
> Thanks for the quick replay. I am all confused now. It says:
>
> iss...@hbase.apache.org
>
> I don't remember subscribing to this list, anyways how do I
> unsubscribe? the same
Hi there, well, this request for input fell like a thud. :-)
But I think perhaps it has to do with the fact that I sent it to the
dev-list instead of the user-list, as people that are actively writing
HBase itself (devs) need less help with such keybuilding utilities.
So one last request for fe
Thanks for the quick replay. I am all confused now. It says:
iss...@hbase.apache.org
I don't remember subscribing to this list, anyways how do I
unsubscribe? the same way as the dev one?
Thanks again!
Renato M.
2013/1/15 Jean-Marc Spaggiari :
> Renato, can you check the "To:" from the JIRA yo
Renato, can you check the "To:" from the JIRA your are receiving?
Which address is it? What are those JIRAs? Are you "watching" them?
JM
2013/1/15, Jean-Daniel Cryans :
> I checked a few emails and as far as I can see it's going to issues@ not
> user@
>
> J-D
>
> On Tue, Jan 15, 2013 at 11:48 AM,
I checked a few emails and as far as I can see it's going to issues@ not user@
J-D
On Tue, Jan 15, 2013 at 11:48 AM, Renato Marroquín Mogrovejo
wrote:
> Hi all,
>
> I am an HBase user but not a developer. The thing is that I keep on
> getting all the JIRA emails from HBase + the user's emails, i
Hi all,
I am an HBase user but not a developer. The thing is that I keep on
getting all the JIRA emails from HBase + the user's emails, isn't that
supposed to be only for the dev list?
Could anyone with the necessary superpowers take the JIRA emails out
of the users list please?
Thanks in advance!
HTable is a blocking interface. When a client issues a put, for example, we
do not want to return until we can confirm the store has been durably
persisted. For client convenience many additional details of remote region
invocation are hidden, for example META table lookups for relocated
regions, r
You should look at the jstack - I think HTablePool is the reason for the
large number of threads. Note that HTablePool is a reusable pool HTable(s)
and each HTable consists of an ExecutorService containing 1 thread by
default. Are you closing the HTable you obtain from HTablePool - if you are
not c
Andrew, could you explain more, why doing cross-table operation is an
anti-pattern of using CP?
Durability might be an issue, as far as I understand. Thanks,
Best Regards,
Wei
From: Andrew Purtell
To: "user@hbase.apache.org" ,
Date: 01/12/2013 09:39 PM
Subject:Re: Coprocess
Thanks Bryan, really appreciate you letting us know the outcome. I'm sure
it will be useful to others.
On Tue, Jan 15, 2013 at 9:28 AM, Bryan Keller wrote:
> I'll follow up on this in case it is useful to anyone. It seems I was
> network I/O limited. The switch I was using was in managed mode wh
I'll follow up on this in case it is useful to anyone. It seems I was network
I/O limited. The switch I was using was in managed mode which decreased
throughput to 1gbit/sec within the switch, not just on the wire. So with
replication set to 2, throughput was about half of the theoretical max on
Thanks Kevin , it looks like this option is only available in 0.90.7
"Kevin O'dell" 於 2013/1/15 下午11:56 寫道:
> If you have only lost your META, I would just use:
>
> hbase hbck -fixMeta -fixAssignments
>
> This will rebuild your META
>
> On Tue, Jan 15, 2013 at 10:53 AM, Lashing wrote:
>
>>
The upload finished and sqoop confirmed it was done successfully. I use
sqoop to create the table and schema.
I've tried the same command and same data on Amazon EMR instance and it
worked without any problem.
If it's in schema why the count changes when you run the count command a
few minutes la
On Tue, Jan 15, 2013 at 1:07 AM, Ibrahim Yakti wrote:
> The root directory is set to /var/lib/hbase/
>
> The count issue is still exist:
>
> *Sqoop: *
> 13/01/15 08:55:23 INFO mapreduce.ImportJobBase: Retrieved 1754285 records.
>
> *MySQL:*
> +--+
> | count(1) |
> +--+
> | 175430
Hi, everybody,
I want to upgrade a HBase cluster from 0.90.5 to 0.94. The cluster has
about 50 nodes and the hadoop version is 0.20.2.
Is there an easy way to do it? Shutdown is acceptable and what I really
care is the safety of the data.
Hope for your suggestions.
Thanks,
Mickey
If you have only lost your META, I would just use:
hbase hbck -fixMeta -fixAssignments
This will rebuild your META
On Tue, Jan 15, 2013 at 10:53 AM, Lashing wrote:
> Kevin
> Thanks for the update. Actually you can think of the whole meta is
> missing , or say deleted.
> Can multiple hbck
Kevin
Thanks for the update. Actually you can think of the whole meta is
missing , or say deleted.
Can multiple hbck repair table run at the same time ?
"Kevin O'dell" 於 2013/1/15 下午11:20 寫道:
> Lashing,
>
> Without knowning the exact errors you are seeing it would be hard to make
> reco
Thanks Samar.
You are right YCSB writes data to a single table 'usertable', but I see very
slow operations (in order of 1-2 operations/second) even for read/update
workload and not only for inserts. So, the region is already split in to
multiple RS before I start my transaction workload.
And k
Lashing,
Without knowning the exact errors you are seeing it would be hard to make
recommendations. If it is just META that is messed up you can use:
hbase hbck -fixMeta -fixAssignments
I think(not sure) in that version you can do
hbase hbck -repair
This will let you target a specific tabl
We recently suffered meta corrupt issue. We tried to use hbck -repair to fix
But the performance is not fast (about 1 region for 1 second). Are there any
better way to speed up the process?
Btw, My Hbase is 0.90.6. Thanks.
You can upgrade to 0.92.2 first, followed by upgrade to 0.94.4
Please consider upgrading Hadoop to 1.0.4 as well
Thanks
On Jan 15, 2013, at 4:47 AM, 幻风 wrote:
> Hi, everybody,
>
> I want to upgrade a HBase cluster from 0.90.5 to 0.94. The cluster has
> about 50 nodes and the hadoop version is
Why not running a MR job? It will run a task per region.
Le 15 janv. 2013 06:38, "Bing Jiang" a écrit :
> In my view, if you want to realize Jubtaus, you can use Endpoint. Which
> is different from RegionServerObserver.
> A client will launch a task about a table, and it will execute an Endpoin
In my view, if you want to realize Jubtaus, you can use Endpoint. Which
is different from RegionServerObserver.
A client will launch a task about a table, and it will execute an Endpoint
in each region that belongs to that table.
2013/1/15 蔡斌
>
>
>
> HiWe try to build an online machine le
HiWe try to build an online machine learning system, like
Jubatus(http://www-conf.slac.stanford.edu/xldb2012/talks/xldb2012_wed_LT09_HMakino.pdf),
based on HBase. Our training data will be stored in HBase, and algorithms will
scan these data to build model. We need a region server level
The root directory is set to /var/lib/hbase/
The count issue is still exist:
*Sqoop: *
13/01/15 08:55:23 INFO mapreduce.ImportJobBase: Retrieved 1754285 records.
*MySQL:*
+--+
| count(1) |
+--+
| 1754303 |
+--+
*HBase:*
104210 row(s) in 11.4980 seconds
running count c
YCSB would be writing all data to one table.. So initially when the table
is small or just created all the writes would go to one RS.. As the table
grows the Region is split into different RS. The would allow parallel
writes, if the keys are random and could possibly make the writes faster.
Samar
HBase Version: 0.92.1-cdh4.1.2, rUnknown, Thu Nov 1 18:03:50 PDT 2012
Sqoop Version: 1.4.1-cdh4.1.2
Stack: yes, it seems that hbase default root directory is /tmp working to
fix this now.
I am writing the data for the first time into HBase from MySQL, not
overwriting it.
--
Ibrahim
On Mon,
30 matches
Mail list logo