I would really just like to insert 1,000,000 rows on 2 machines like so
[r...@localhost hadoop]# ./bin/hadoop org.apache.hadoop.hbase.PerformanceEvalua tion --rows=1000000 sequentialWrite 2 and then insert 1,000,000 rows on 3 machines [r...@localhost hadoop]# ./bin/hadoop org.apache.hadoop.hbase.PerformanceEvalua tion --rows=1000000 sequentialWrite 3 but when I ran the first test and then looked at the size afterwards, I had 3.6 gigs on one node and 0 gigs on the other node. I have dfs.replication set to 1(ie. I don't want it for this test right now). I am just trying to play with the system a bit here but not sure why it is very imbalanced on the writes??? Is there a block size I should change or something? Once I get done with this, I want to write a map job to take data from another db and stuff It into my small little cluster to test it all out as a small prototype. Thanks, Dean This message and any attachments are intended only for the use of the addressee and may contain information that is privileged and confidential. If the reader of the message is not the intended recipient or an authorized representative of the intended recipient, you are hereby notified that any dissemination of this communication is strictly prohibited. If you have received this communication in error, please notify us immediately by e-mail and delete the message and any attachments from your system.