Hi rajesh, I use Hbase 0.94.11 and Hadoop 1.2.1. The file system of bulkload output directory and hbase cluster are the same, too.
I've also coded a MapReduce job using HFileOutputFormat. When I use LoadIncrementalHFiles to move the output of my job to HBase table, it still copies instead of cut. Thanks On Sat, Sep 14, 2013 at 2:50 PM, rajesh babu Chintaguntla < [email protected]> wrote: > Hi BagherEsmaeily, > > which version of hbase you are using? Is the file system of bulkload > output directory and hbase cluster are same? > > If you are using hbase older than 0.94.5 version, the Storefiles > generated by importtsv are getting copied instead of moving even if > the file system of bulkload output directory and hbase cluster are same. > > Its a bug and solved in 0.94.5 (HBASE-5498). > > Thanks. > Rajeshbabu > > On Sat, Sep 14, 2013 at 12:01 PM, M. BagherEsmaeily <[email protected] > >wrote: > > > Hello, > > > > I was using HBase complete bulk load to transfer the output of ImportTsv > to > > a table in HBase, and I noticed that it copies the output instead of > > cutting. This takes long time for my gigabytes of data. > > > > In HBase documentation ( > > http://hbase.apache.org/book/ops_mgt.html#completebulkload) I read that > > the > > files would be moved not copied. Can anyone help me with this? > > > > Kind Regards > > >
