Hi Jon, 

No problem. We do have snapshots enabled on the target cluster, and we are 
using the default hfile archiver settings on both clusters.

Thanks,

Sean 


On Tuesday, 23 April, 2013 at 1:54 PM, Jonathan Hsieh wrote:

> Sean,
> 
> Thanks for finding this problem. Can you provide some more information so
> that we can try to duplicate and fix this problem?
> 
> Are snapshots on on the target cluster?
> What are the hfile archiver settings in your hbase-site.xml on both
> clusters?
> 
> Thanks,
> Jon.
> 
> 
> On Mon, Apr 22, 2013 at 4:47 PM, Sean MacDonald <[email protected] 
> (mailto:[email protected])> wrote:
> 
> > It looks like you can't export a snapshot to a running cluster or it will
> > start cleaning up files from the archive after a period of time. I have
> > turned off HBase on the destination cluster and the export is working as
> > expected now.
> > 
> > Sean
> > 
> > 
> > On Monday, 22 April, 2013 at 9:22 AM, Sean MacDonald wrote:
> > 
> > > Hello,
> > > 
> > > I am using HBase 0.94.6 on CDH 4.2 and trying to export a snapshot to
> > another cluster (also CDH 4.2), but this is failing repeatedly. The table I
> > am trying to export is approximately 4TB in size and has 10GB regions. Each
> > of the map jobs runs for about 6 minutes and appears to be running
> > properly, but then fails with a message like the following:
> > > 
> > > 2013-04-22 16:12:50,699 WARN org.apache.hadoop.hdfs.DFSClient:
> > DataStreamer Exception
> > org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException):
> > No lease on
> > /hbase/.archive/queries/533fcbb7858ef34b103a4f8804fa8719/d/651e974dafb64eefb9c49032aec4a35b
> > File does not exist. Holder DFSClient_NONMAPREDUCE_-192704511_1 does not
> > have any open files. at
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2396)
> > at
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2387)
> > at
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2183)
> > at
> > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:481)
> > at
> > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:297)
> > at
> > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtoc
> > ol
> > > $2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44080) at
> > 
> > 
> > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
> > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002) at
> > org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1695) at
> > org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1691) at
> > java.security.AccessController.doPrivileged(Native Method) at
> > javax.security.auth.Subject.doAs(Subject.java:396) at
> > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
> > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1689)
> > > 
> > > I was able to see the file that the LeaseExpiredException mentions on
> > the destination cluster before the exception happened (it is gone
> > afterwards).
> > > 
> > > Any help that could be provided in resolving this would be greatly
> > appreciated.
> > > 
> > > Thanks and have a great day,
> > > 
> > > Sean
> 
> 
> -- 
> // Jonathan Hsieh (shay)
> // Software Engineer, Cloudera
> // [email protected] (mailto:[email protected])



Reply via email to