I would really just like to insert 1,000,000 rows on 2 machines like so

 

[r...@localhost hadoop]# ./bin/hadoop
org.apache.hadoop.hbase.PerformanceEvalua

tion --rows=1000000 sequentialWrite 2

 

and then insert 1,000,000 rows on 3 machines

 

[r...@localhost hadoop]# ./bin/hadoop
org.apache.hadoop.hbase.PerformanceEvalua

tion --rows=1000000 sequentialWrite 3

 

but when I ran the first test and then looked at the size afterwards, I
had 3.6 gigs on one

node and 0 gigs on the other node.  I have dfs.replication set to 1(ie.
I don't want it for

this test right now).  I am just trying to play with the system a bit
here but not sure why

it is very imbalanced on the writes???  Is there a block size I should
change or something?

 

Once I get done with this, I want to write a map job to take data from
another db and stuff

It into my small little cluster to test it all out as a small prototype.

 

Thanks,

Dean


This message and any attachments are intended only for the use of the addressee 
and
may contain information that is privileged and confidential. If the reader of 
the 
message is not the intended recipient or an authorized representative of the
intended recipient, you are hereby notified that any dissemination of this
communication is strictly prohibited. If you have received this communication in
error, please notify us immediately by e-mail and delete the message and any
attachments from your system.

Reply via email to