Hi Rita,
HBase Bulk Loader is a viable solution for loading such huge data set. Even
if your import file has a separator other than tab you can use ImportTsv as
long as the separator is single character. If in case you want to put in
your business logic while writing the data to HBase then you can
Well, if the file that you have contains TSV, you can directly use the
ImportTSV utility of HBase to do a bulk load.
More details about that can be found here :
http://hbase.apache.org/book/ops_mgt.html#importtsv
The other option for you is to run a MR job on the file that you have, to
generate t
Oh... I just meant you need to have your hands on a RegionScanner :)
As long as you only scan forward it should work.
- Original Message -
From: Tom Brown
To: user@hbase.apache.org; lars hofhansl
Cc:
Sent: Friday, August 3, 2012 5:47 PM
Subject: Re: Need to fast-forward a scanner insi
So I understand I'll need to upgrade to 0.94 (which won't be a problem
because the releases are binary-compatible). I see that the
RegionScanner interface contains the new method "reseek(byte[] row)".
I have a reference to a RegionScanner in my coprocessor because I'm
using: getEnvironment().getR
Good!
Submitted initial patch of fuzzy row key filter at
https://issues.apache.org/jira/browse/HBASE-6509. You can just copy the
filter class and include it in your code and use it in your setup as any
other custom filter (no need to patch HBase).
Please let me know if you try it out (or post you
Hi guys, I've been trying to setup HBase for OpenTSDB for a few days now and
am completely stuck. I've gotten .92 running on a virtual machine but I am
completely unable to deploy it to a real machine.
Firstly, I've been following this guide:
http://opentsdb.net/setup-hbase.html
Here's what I've
We recently added a new API for that:
RegionScanner.reseek(...). See HBASE-5520. 0.94+ only, unfortunately.
So it depends specifically on where you hook this up. If you do it at
RegionObserver.postScannerOpen you can reseek forward at any time.
-- Lars
- Original Message -
From: Tom
https://issues.apache.org/jira/browse/HDFS-197 has a workaround (see
last comment)
On Fri, Aug 3, 2012 at 1:33 PM, Mohit Anchlia wrote:
> I ran test from cygwin but it fails here. Could someone help me with how to
> go about fixing this issue?
>
> java.io.IOException: Expecting a line not the end
I ran test from cygwin but it fails here. Could someone help me with how to
go about fixing this issue?
java.io.IOException: Expecting a line not the end of stream
at org.apache.hadoop.fs.DF.parseExecResult(DF.java:117)
at org.apache.hadoop.util.Shell.runCommand(Shell.java:237)
at org.apache.hadoo
I have a custom coprocessor that aggregates a selection of records
from the table based various criteria. For efficiency, I would like to
make it skip a bunch of records. For example, if I don't need any
"" records and I encounter "", I would like to tell it to
skip everything until "AA
2012/8/3, Jean-Daniel Cryans :
> On Fri, Aug 3, 2012 at 8:15 AM, Jean-Marc Spaggiari
> wrote:
>> Me again ;)
>>
>> I did some more investigation.
>
> It would really help to see the region server log although the fsck
> output might be enough.
I looked under evey directory and only one is contain
Hi Mohit:
You might need to install Cygwin if the tool has dependency on Linux
command like bash.
Best Regards,
Jerry
On Friday, August 3, 2012, N Keywal wrote:
> Hi Mohit,
>
> For simple cases, it works for me for hbase 0.94 at least. But I'm not
> sure it works for all features. I've never t
On Fri, Aug 3, 2012 at 8:15 AM, Jean-Marc Spaggiari
wrote:
> Me again ;)
>
> I did some more investigation.
It would really help to see the region server log although the fsck
output might be enough.
BTW you'll find 0.94.1 RC1 here:
http://people.apache.org/~larsh/hbase-0.94.1-rc1/
>
> And I fo
hbck should be able to take care of it now.
J-D
On Fri, Aug 3, 2012 at 2:21 AM, holger.lewin wrote:
>
> I just checked out hbase-0.92.1 and noticed that /bin/add_table.rb has been
> deleted. (CHANGES.txt: "HBASE-2460 add_table.rb deletes any tables for
> which the target table name is a prefix"
Me again ;)
I did some more investigation.
And I found that:
http://pastebin.com/Bedm6Ldy
Seems that no region is serving my logs. That's strange because all my
servers are up and fsck is telling me that FS is clean.
Can I just delete those files? What's the impact of such delete? I
don't real
Here us the complete log. And seems it's every 30 seconds and not
every 20 seconds...
http://pastebin.com/gMiURnnj
2012/8/3, Jean-Marc Spaggiari :
> Hi,
>
> I'm using HBase 0.94.0.
>
> I stopped the cluster for some maintenance, and I'm have some troubles
> to restart it.
>
> I'm getting one line
Hi,
I'm using HBase 0.94.0.
I stopped the cluster for some maintenance, and I'm have some troubles
to restart it.
I'm getting one line every about
Start Time Description State Status
Fri Aug 03 08:59:54 EDT 2012Doing distributed log split in
[hdfs://node3:9000/hbase/.logs/latitud
Hi,
What does your schema look like?
Would it make sense to changing the key to user_id '|' timestamp and then use
the session_id in the column name?
On Aug 2, 2012, at 7:23 AM, Christian Schäfer wrote:
> OK,
>
> at first I will try the scans.
>
> If that's too slow I will have to upgr
Hi Matt,
sure I got this in mind as an last option (at least on a limited subset of
data).
Due to our estimation of some billions rows a week a selective filtering needs
to take place at the server side.
But I agree that one could do fine filtering stuff on the client side on a
handy data sub
Hi Alex,
thanks a lot for the hint about setting the timestamp of the put.
I didn't know that this would be possible but that's solving the problem (first
test was successful).
So I'm really glad that I don't need to apply a filter to extract the time and
so on for every row.
Nevertheless I wou
I just checked out hbase-0.92.1 and noticed that /bin/add_table.rb has been
deleted. (CHANGES.txt: "HBASE-2460 add_table.rb deletes any tables for
which the target table name is a prefix"). I wonder if theres a replacement
or fixed version of it somewhere?
Thanks,
Holger
--
View this message in
The problem resolved. it caused by the zookeeper data corrupted. so I
modified zookeeper data dir to another directory on hbase-site.xml and
restart hbase again.
hbase.zookeeper.property.dataDir
/home/zhouhh/myhadoop/zk
Property from ZooKeeper's config zoo.cfg.
22 matches
Mail list logo