Hi, I am trying to upload close to 4 GB of data on hbase, but I am not able to do. Following are my observations:
1. Before failing number of rows inserted were about 800,000, and the time taken for this upload was close to 6 hours, and this is just 1/10th of total rows that I have to insert,. This way it would take me ages before I can insert all the data into the table. And by the time I would be finished I would have similar size ready for another insertion. Is there any better way to do this? 2. I saw this happens whenever a datanode goes down. Doesn't hbase have mechanism where it continues loading data to other datanodes, in case one fails? For inserting data in the table, I wrote a php script having database connectivity using the REST [ https://issues.apache.org/jira/browse/HBASE-37] Since the datafile is big, and on my first execution I faced a similar situation even when none of my datanodes were down. I thought of breaking it into smaller files and then processing them, but, it doesn't help. Can you guys help me with this. Thanks ------------------------------------------------- hbase-hadoop-
