Looks like HBASE-9538 is for the same issue. On Sep 15, 2013, at 1:09 AM, "M. BagherEsmaeily" <mbesmae...@gmail.com> wrote:
> HBASE-9537 > > > On Sun, Sep 15, 2013 at 11:56 AM, Ted Yu <yuzhih...@gmail.com> wrote: > >> File a JIRA for the issue ? >> >> On Sep 14, 2013, at 11:10 PM, "M. BagherEsmaeily" <mbesmae...@gmail.com> >> wrote: >> >>> Hi rajesh, >>> >>> I use Hbase 0.94.11 and Hadoop 1.2.1. The file system of bulkload >>> output directory and hbase cluster are the same, too. >>> >>> I've also coded a MapReduce job using HFileOutputFormat. When I use >>> LoadIncrementalHFiles >>> to move the output of my job to HBase table, it still copies instead of >> cut. >>> >>> Thanks >>> >>> >>> On Sat, Sep 14, 2013 at 2:50 PM, rajesh babu Chintaguntla < >>> chrajeshbab...@gmail.com> wrote: >>> >>>> Hi BagherEsmaeily, >>>> >>>> which version of hbase you are using? Is the file system of >> bulkload >>>> output directory and hbase cluster are same? >>>> >>>> If you are using hbase older than 0.94.5 version, the Storefiles >>>> generated by importtsv are getting copied instead of moving even >> if >>>> the file system of bulkload output directory and hbase cluster are same. >>>> >>>> Its a bug and solved in 0.94.5 (HBASE-5498). >>>> >>>> Thanks. >>>> Rajeshbabu >>>> >>>> On Sat, Sep 14, 2013 at 12:01 PM, M. BagherEsmaeily < >> mbesmae...@gmail.com >>>>> wrote: >>>> >>>>> Hello, >>>>> >>>>> I was using HBase complete bulk load to transfer the output of >> ImportTsv >>>> to >>>>> a table in HBase, and I noticed that it copies the output instead of >>>>> cutting. This takes long time for my gigabytes of data. >>>>> >>>>> In HBase documentation ( >>>>> http://hbase.apache.org/book/ops_mgt.html#completebulkload) I read >> that >>>>> the >>>>> files would be moved not copied. Can anyone help me with this? >>>>> >>>>> Kind Regards >>