I am new to HBASE and Phoenix world. 

I have designed and executed a map reduce job which writes around 2.4 billion 
cells (rows*columns) in HBASE via Phoenix in about 80min. I have reduced the 
"mapreduce.input.fileinputformat.split.maxsize" to 8MB to increase the number 
of mapper which helped me to increase throughput and reduce duration from 
140min to 80min. I have wide rows i.e each row has about ~27-30k columns and 
most of these ( 95%) are dynamic columns which are created during UPSERT 
operations.

I wanted guidance with if I can reduce the time duration of my job any further 
by tweaking any parameters which are default right now from either HBASE side  
or Phoenix side.

Thanks.

Reply via email to