Re: hbase importtsv

2014-05-01 Thread Amit Tewari
Make sure there are no primary key clash. HBase would over write the row if you upload data with same primary key. That's one reason you can possibly get less rows than what you uploaded Sent from my mobile device, please excuse the typos On May 1, 2014, at 3:34 PM, Kennedy, Sean C.

RE: hbase importtsv

2014-02-17 Thread Kennedy, Sean C.
Matouck, Thank you very much, I had success! Have a great day ... Sincerely, Sean From: Matouk IFTISSEN [mailto:matouk.iftis...@ysance.com] Sent: Friday, February 14, 2014 6:52 PM To: user@hive.apache.org Subject: Re: hbase importtsv hello, You can use bulkload in to pahses

Re: hbase importtsv

2014-02-15 Thread Manish
On Feb 14, 2014, at 3:51 PM, Matouk IFTISSEN matouk.iftis...@ysance.com wrote: hello, You can use bulkload in to pahses, in MapR distribution we use this: 2014-02-14 16:59 GMT+01:00 Kennedy, Sean C. sean.kenn...@merck.com: I am trying to load 1.4M records in a 7 column CSV file into

hbase importtsv

2014-02-14 Thread Kennedy, Sean C.
I am trying to load 1.4M records in a 7 column CSV file into hbase. Question 1: Is this feasible? Question 2: What type of tuning on hbase and or hdfs would be needed? I am using apache hbase 0.94.15 and apache hadoop 1.2.1 Here is my command string: /hd/hadoop/bin/hadoop jar

Re: hbase importtsv

2014-02-14 Thread Matouk IFTISSEN
hello, You can use bulkload in to pahses, in MapR distribution we use this: 1. first phase : map metadata to table hbase hadoop jar /opt/mapr/hbase/hbase-0.94.5/hbase-0.94.5-mapr.jar importtsv -Dimporttsv.separator=';' -Dimporttsv.bulk.output=folder_bulk_local