copytable will start a MR job and will do the copy in parallele, which is good. But it's still going to do a lot of puts on the destination cluster which will trigger flushs and compactions. If it's easy for you to send your csv file there I think it will be more efficient, even if copyTbale can solve your issue.
JM 2015-09-01 15:01 GMT-04:00 Gaurav Agarwal <gaurav130...@gmail.com>: > In this will copyTable command of hbase is good to use or transfer the csv > file on the other side and bulkload from there which one is good according > to performance > > On Wed, Sep 2, 2015 at 12:23 AM, Jean-Marc Spaggiari < > jean-m...@spaggiari.org> wrote: > >> Hi Gaurav, >> >> bulk load bypass the WAL, that's correct. It's true for Phoenix, it's >> true for HBase (outside of Phoenix). >> >> If you have replication activated, you will have to bulkload the data >> into the 2 clusters. Transfert your csv files on the other side too and >> bulkload from there. >> >> JM >> >> 2015-09-01 14:51 GMT-04:00 Gaurav Agarwal <gaurav130...@gmail.com>: >> >>> Hello >>> >>> We are using phoenix Map reduce CSV uploader to load data into HBASe . I >>> read documentation on Phoenix site, it will only create HFLE no WAL logs >>> will be created.Please confirm understanding is correct or wrong >>> >>> We have to use HBASe replication across cluster for Master Master >>> scenario. Will the replication work in that scenario or do we need to use >>> Copy Table to replicate ? >>> >>> thanks >>> >> >> >