You can set the block replication in your hdfs-site.xml configuration file
by specifying the property dfs.replication with replication value.

Moreover once the remote table dump is completed, you can run the
replication for your individual or group of table by,

hadoop fs –setrep –w [int REPLICATION_FACTOR] -R
/path/to/your/directory/or/file

Cheers!


On Fri, Nov 28, 2014 at 1:44 PM, Li Li <[email protected]> wrote:

> thank you. How can I specify replication factor with this command?
> will hbase  -Ddfs.replication=1
> org.apache.hadoop.hbase.mapreduce.CopyTable work?
>
> On Fri, Nov 28, 2014 at 3:32 PM, Vineet Mishra <[email protected]>
> wrote:
> > Hi Li Li,
> >
> > You can copy the Hbase Tables Remotely to another machine with the
> > following commands,
> >
> > # create new tableOrig on destination cluster
> > dstCluster$ echo "create 'tableOrig', 'cf1', 'cf2'" | hbase shell
> > # on source cluster run copy table with destination ZK quorum specified
> > using --peer.adr
> > # WARNING: In older versions, you are not alerted about any typo in these
> > arguments!
> > srcCluster$ hbase org.apache.hadoop.hbase.mapreduce.CopyTable --peer.adr=
> > dstClusterZK:2181:/hbase tableOrig
> >
> > More information given in the section Remote HBase instance backup at
> >
> >
> http://blog.cloudera.com/blog/2012/06/online-hbase-backups-with-copytable-2/
> >
> > Hope that works for you pretty straightforwardly.
> >
> > Cheers!
> >
> >
> > On Fri, Nov 28, 2014 at 12:44 PM, Li Li <[email protected]> wrote:
> >
> >> I have a hbase cluster of version 0.98.5 with hadoop-1.2.1(no mapreduce)
> >> I want to copy all the tables to another cluster whose version is
> >> 0.98.1-cdh5.1.0 with 2.3.0-cdh5.1.0.
> >> And also I want specify the hdfs replication factor of the files in
> >> new cluster. is it possible?
> >>
>

Reply via email to