How can we verify that the data(tables) is distributed across the cluster??
Is there a way to confirm it that the data is distributed across all the
nodes in the cluster.?
On Thu, Sep 27, 2012 at 12:26 PM, Venkateswara Rao Dokku
dvrao@gmail.com wrote:
Hi,
I am completely new to Hbase.
Check http://hbase.apache.org/book/arch.bulk.load.html
Best Regards,
Sonal
Crux: Reporting for HBase https://github.com/sonalgoyal/crux
Nube Technologies http://www.nubetech.co
http://in.linkedin.com/in/sonalgoyal
On Thu, Sep 27, 2012 at 12:45 PM, Venkateswara Rao Dokku
dvrao@gmail.com
You can use mapreduce. We have an utility called ImportTsv tool that allows
you to bulk load data from a flat file? Is this your use case?
Pls refer to http://hbase.apache.org/book.html#arch.bulk.load
Regards
Ram
-Original Message-
From: Venkateswara Rao Dokku
Hi,
I would like to direct you to the reference guide, but I must acknowledge
that, well, it's a reference guide, hence not really easy for a plain new
start.
You should have a look at Lars' blog (and may be buy his book), and
especially this entry:
I can see that HMaster is not started on the data-node machine when the
start scripts in hadoop hbase ran on the hadoop-namenode. My doubt is
that,Shall we have to start that master on the hadoop-datanode1 too or the
hadoop-datanode1 will access the Hmaster that is running on the
hadoop-namenode
On Thu, Sep 27, 2012 at 1:09 PM, Venkateswara Rao Dokku dvrao@gmail.com
wrote:
I can see that HMaster is not started on the data-node machine when the
start scripts in hadoop hbase ran on the hadoop-namenode. My doubt is
that,Shall we have to start that master on the hadoop-datanode1 too
You should launch the master only once, on whatever machine you like. Then
you will be able to access it from any other machine.
Please have a look at the blog I mentioned in my previous mail.
On Thu, Sep 27, 2012 at 9:39 AM, Venkateswara Rao Dokku dvrao@gmail.com
wrote:
I can see that
Hello,
Corollary, what is the better way to migrate data from a 0.90 cluster to a
0.92 cluser ?
Hbase 0.90 = Client 0.90 = stdout | stdin = client 0.92 = Hbase 0.92
All the data must tansit on a single host where compute the 2 clients.
It may be paralalize with mutiple version working with
Hello Mohit,
It should be /hbase/hbase/table/SESSIONID_TIMELINE..Apologies for the
typo. For rest of the things, I feel Ramkrishna sir has provided a good and
proper explanation. Please let us know if you still have any doubt or
question.
Ramkrishna.S.Vasudevan : You are welcome sir. It's my
You don't have to migrate the data when you upgrade, it's done on the fly.
But it seems you want to do something more complex? A kind of realtime
replication between two clusters in two different versions?
On Thu, Sep 27, 2012 at 9:56 AM, Damien Hardy dha...@viadeoteam.com wrote:
Hello,
Dan, see inlined.
On Thu, Sep 27, 2012 at 5:30 AM, Dan Han dannahan2...@gmail.com wrote:
Hi, Eugeny ,
Thanks for your response. I answered your questions inline in Blue.
And I'd like to give an example to describe my problem.
Let's think about two data schemas for the same dataset.
The
Actually, I have an old cluster on on prod with 0.90.3 version installed
manually and I am working on a CDH4 new cluster deployed full automatic
with puppet.
While migration is not reversible (according to the pointer given by
Jean-Daniel) I would like to keep he old cluster safe by side to be
Anyone had a rough measurement of random read/write perf and throughput?
Assume a typical machines/workload, the region server has 5GB for memtable,
further assume each key (20 bytes) has 100 bytes value (for simplicity,
just one cf, one column).
Further assume workload is against a single region
Hi
That is not needed, infact it has been fixed in the latest trunk version as
part of HBASE-6327.
We can back port the issue I feel. Thanks for bringing this into notice.
Regards
Ram
-Original Message-
From: jlei liu [mailto:liulei...@gmail.com]
Sent: Thursday, September 27, 2012
I started the Hmaster on the hadoop-namenode. But I was not able to access
it from the hadoop-datanode. Could you please help me solving this problem
by sharing what are the possibilities for this to happen.
On Thu, Sep 27, 2012 at 1:21 PM, n keywal nkey...@gmail.com wrote:
You should launch
Rao,
Can you make sure your region server is actually running? You can use jps
command to see Java processes, or a ps ax |grep region.
Thanks,
Stas
On Thu, Sep 27, 2012 at 12:25 PM, Venkateswara Rao Dokku
dvrao@gmail.com wrote:
When I try to scan the table that is created by
Yes, I can see the region server running. The output of the jps command is
given below
*Hadoop-namenode:*
* *HQuorumPeer
Main
HMaster
HRegionServer
SecondaryNameNode
Jps
NameNode
JobTracker
hadoop-datanode1:
TaskTracker
DataNode
Jps
Main
NC
HRegionServer
The complete error is given
I'm setting up HBase using CDH4.
https://ccp.cloudera.com/display/CDH4DOC/HBase+Installation#HBaseInstallation-DeployingHBaseinaDistributedCluster
I installed Zookeeper on my namenode, which is also my HBase master.
hbase-master now starts and runs. My understanding from the above guide is
that
Hello Bai Shen,
It is not a compulsion to run ZK on the same machine where NN is
running. You can run it anywhere and if this is the case you have to
specify the location of you ZK node through the hbase-site.xml file.
Infact, in real world scenarios people create a separate ZK cluster and
Thanks everyone for the input, it's helpful. I did remove the znode from
/hbase/table/SESSIONID_TIMELINE and after that I was able to list the
table. At that point I tried to do a put but when I did a put I got a
message NoRegionServer online. I looked in the logs and it says the Failed
to open
Hi Mohit,
We should not delete znode's manually which will cause inconsistencies like
region may be shown as online on master, but it wont be on region server.
That's put is failing in your case. Master restart will bring back your
cluster to normal state(recovery any failures in enable/disable).
I did restart entire cluster and still that didn't help. Looks like once I
get in this Race condition there is no way to come out of it?
On Thu, Sep 27, 2012 at 8:00 AM, rajesh babu chintaguntla
chrajeshbab...@gmail.com wrote:
Hi Mohit,
We should not delete znode's manually which will cause
What property do I set in hbase-site.xml? That's what I'm having trouble
finding.
Thanks.
On Thu, Sep 27, 2012 at 10:30 AM, Mohammad Tariq donta...@gmail.com wrote:
Hello Bai Shen,
It is not a compulsion to run ZK on the same machine where NN is
running. You can run it anywhere and if
NM. Turned out that I had screwed up the property setting. Everything is
working now.
Thanks.
On Thu, Sep 27, 2012 at 1:28 PM, Bai Shen baishen.li...@gmail.com wrote:
What property do I set in hbase-site.xml? That's what I'm having trouble
finding.
Thanks.
On Thu, Sep 27, 2012 at 10:30
Hi
Thanks for you help. I've been doing this in a pseudo-distributed
hbase-0.92.1 environment with one region server. I'm trying to scan a
table and see it timeout. I'm trying to recreate a scenario where the RS
is not responding (for instance due to NIC failure). So I've been
issuing a 'kill
Including dev mailing list. So I let it run, and after about 43 minutes
I finally got some exceptions (Sorry for the long paste)
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after
attempts=10, exceptions:
Thu Sep 27 14:59:29 EDT 2012,
Hello,
I am exploring HBase Lily and I have a few starter questions hoping
to get some help from users in this group who had tried that before:
(1) Do I need to post all the HBase table contents to Lily (treat Lily
as another DataStore) in order to enable the index and search
functionality? If
Thanks for your advice, Eugeny.
Best Wishes
Dan Han
On Thu, Sep 27, 2012 at 2:34 AM, Eugeny Morozov
emoro...@griddynamics.comwrote:
Dan, see inlined.
On Thu, Sep 27, 2012 at 5:30 AM, Dan Han dannahan2...@gmail.com wrote:
Hi, Eugeny ,
Thanks for your response. I answered your
Hi Ramkrishna,
I think relocating regions is based on the queries and queried data.
The relocation can scatter the regions involved in the query across region
servers
which might enable large queries get better load balance.
For small queries, distribution of regions can also impact the
I confirm, once I removed the localhost entry, HBase started working.
My hosts file now contains only:
x.x.x.1 Machine1
x.x.x..2Machine2
x.x.x.x.3 Machine3
x.x.x.N MachineN
-Original Message-
From: Artem Ervits [mailto:are9...@nyp.org]
Sent:
Lily is based on HBase.
--- On Thu, 9/27/12, Jason Huang jason.hu...@icare.com wrote:
From: Jason Huang jason.hu...@icare.com
Subject: HBase and Lily?
To: user@hbase.apache.org
Date: Thursday, September 27, 2012, 1:58 PM
Hello,
I am exploring HBase Lily and I have a few starter questions
Hi Mohith
Before restarting again just disable the compression means set it to default
and restart the cluster. This is just to ensure that the cluster is able to
come back from the enable/disable problem.
The Snappy problem could be different. I am suggesting this so that we
can isolate the
Hi Dan
Am not very sure whether my answer was infact relevant to your problem.
Any way I can try answering about the 'region being redundant'?
No two regions can be responsible for the same range of data in one table.
That is why if any region is not available that portion of data is not
Hi
Lily is an indexing solution for HBase. This indexing is purely happening at
the client side. If you see Lily will sit in between the client app and HBase.
The app need to insert/delete data via Lily only. Lily will write the user
data into HBase table. Also there is another index table
34 matches
Mail list logo