What database is this ? Was hbase mentioned ?

On Monday, February 18, 2013, Mohammad Tariq wrote:

> Hello Masoud,
>
>           You can use the Bulk Load feature. You might find it more
> efficient than normal client APIs or using the TableOutputFormat.
>
> The bulk load feature uses a MapReduce job to output table data
> in HBase's internal data format, and then directly loads the
> generated StoreFiles into a running cluster. Using bulk load will use
> less CPU and network resources than simply using the HBase API.
>
> For a detailed info you can go here :
> http://hbase.apache.org/book/arch.bulk.load.html
>
> Warm Regards,
> Tariq
> https://mtariq.jux.com/
> cloudfront.blogspot.com
>
>
> On Mon, Feb 18, 2013 at 5:00 PM, Masoud 
> <mas...@agape.hanyang.ac.kr<javascript:_e({}, 'cvml', 
> 'mas...@agape.hanyang.ac.kr');>
> > wrote:
>
>>
>> Dear All,
>>
>> We are going to do our experiment of a scientific papers, ]
>> We must insert data in our database for later consideration, it almost
>> 300 tables each one has 2/000/000 records.
>> as you know It takes lots of time to do it with a single machine,
>> we are going to use our Hadoop cluster (32 machines) and divide 300
>> insertion tasks between them,
>> I need some hint to progress faster,
>> 1- as i know we dont need to Reduser, just Mapper in enough.
>> 2- so wee need just implement Mapper class with needed code.
>>
>> Please let me know if there is any point,
>>
>> Best Regards
>> Masoud
>>
>>
>>
>>
>

Reply via email to