Hi,

I am trying to upload close to 4 GB of data on hbase, but I am not able to
do. Following are my observations:

1. Before failing number of rows inserted were about 800,000, and the time
taken for this upload was close to 6 hours, and this is just 1/10th of total
rows that I have to insert,. This way it would take me ages before I can
insert all the data into the table. And by the time I would be finished I
would have similar size ready for another insertion. Is there any better way
to do this?

2. I saw this happens whenever a datanode goes down. Doesn't hbase have
mechanism where it continues loading data to other datanodes, in case one
fails?

For inserting data in the table, I wrote a php script having database
connectivity using the REST [ https://issues.apache.org/jira/browse/HBASE-37]
Since the datafile is big, and on my first execution I faced a similar
situation even when none of my datanodes were down. I thought of breaking it
into smaller files and then processing them, but, it doesn't help.

Can you guys help me with this.

Thanks
-------------------------------------------------

hbase-hadoop-

Reply via email to