Hi

Yes .. i had gone thro that
But those scenarios were something like the region server went down or data
node is down.

Here it is like only master got restarted.

Regards
Ram

****************************************************************************
***********
This e-mail and attachments contain confidential information from HUAWEI,
which is intended only for the person or entity whose address is listed
above. Any use of the information contained herein in any way (including,
but not limited to, total or partial disclosure, reproduction, or
dissemination) by persons other than the intended recipient's) is
prohibited. If you receive this e-mail in error, please notify the sender by
phone or email immediately and delete it!

-----Original Message-----
From: Ted Yu [mailto:[email protected]] 
Sent: Saturday, April 09, 2011 7:42 PM
To: [email protected]; [email protected]
Subject: Re: Killing and restarting of master caused
AlreadyBeingCreatedException from HLogs

Have you read the email thread entitled 'file is already being created by
NN_Recovery' on user mailing list ?

On Sat, Apr 9, 2011 at 7:06 AM, Ramkrishna S Vasudevan <
[email protected]> wrote:

> If we kill the HMaster and try restarting it.. the following exceptions
are
> logged
>
>
>
> plitting hlog 2 of 2:
> hdfs://
> 10.18.52.108:9000/hbase/.logs/linux108,60020,1302346754067/linux108%3
> A60020.1302350355407, length=1459
>
> 2011-04-09 18:02:56,017 INFO org.apache.hadoop.hbase.util.FSUtils:
> Recovering file
> hdfs://
> 10.18.52.108:9000/hbase/.logs/linux108,60020,1302346754067/linux108%3
> A60020.1302350355407
>
> 2011-04-09 18:02:56,037 ERROR
> com.huawei.isap.ump.ha.client.RPCRetryAndSwitchInvoker: Exception occured
> while connecting to server : /10.18.52.108:9000
>
> org.apache.hadoop.ipc.RemoteException:
> org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to
> create file
> /hbase/.logs/linux108,60020,1302346754067/linux108%3A60020.1302350355407
> for
> DFSClient_hb_m_linux108:60000_1302352358592 on client 10.18.52.108,
because
> this file is already being created by NN_Recovery on 10.18.52.108
>
>            at
>
>
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLeaseInternal(FSN
> amesystem.java:1453)
>
>            at
>
>
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSName
> system.java:1291)
>
>            at
>
>
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFile(FSNamesystem.
> java:1473)
>
>            at
> org.apache.hadoop.hdfs.server.namenode.NameNode.append(NameNode.java:628)
>
>            at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>
>            at
>
>
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39
> )
>
>            at
>
>
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl
> .java:25)
>
>            at java.lang.reflect.Method.invoke(Method.java:597)
>
>            at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:541)
>
>            at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1105)
>
>            at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1101)
>
>            at java.security.AccessController.doPrivileged(Native Method)
>
>            at javax.security.auth.Subject.doAs(Subject.java:396)
>
>            at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1099)
>
>
>
>            at org.apache.hadoop.ipc.Client.call(Client.java:942)
>
>            at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:239)
>
>            at $Proxy5.append(Unknown Source)
>
>            at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>
>            at
>
>
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39
> )
>
>            at
>
>
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl
> .java:25)
>
>            at java.lang.reflect.Method.invoke(Method.java:597)
>
>            at
>
>
com.huawei.isap.ump.ha.client.RPCRetryAndSwitchInvoker.invokeMethod(RPCRetry
> AndSwitchInvoker.java:157)
>
>            at
>
>
com.huawei.isap.ump.ha.client.RPCRetryAndSwitchInvoker.invokeMethod(RPCRetry
> AndSwitchInvoker.java:145)
>
>            at
>
>
com.huawei.isap.ump.ha.client.RPCRetryAndSwitchInvoker.invoke(RPCRetryAndSwi
> tchInvoker.java:54)
>
>            at $Proxy5.append(Unknown Source)
>
>            at org.apache.hadoop.hdfs.DFSClient.append(DFSClient.java:741)
>
>            at
>
>
org.apache.hadoop.hdfs.DistributedFileSystem.append(DistributedFileSystem.ja
> va:366)
>
>            at org.apache.hadoop.fs.FileSystem.append(FileSystem.java:665)
>
>            at
> org.apache.hadoop.hbase.util.FSUtils.recoverFileLease(FSUtils.java:634)
>
>            at
>
>
org.apache.hadoop.hbase.regionserver.wal.HLogSplitter.splitLog(HLogSplitter.
> java:261)
>
>            at
>
>
org.apache.hadoop.hbase.regionserver.wal.HLogSplitter.splitLog(HLogSplitter.
> java:188)
>
>            at
>
>
org.apache.hadoop.hbase.master.MasterFileSystem.splitLog(MasterFileSystem.ja
> va:196)
>
>            at
>
>
org.apache.hadoop.hbase.master.MasterFileSystem.splitLogAfterStartup(MasterF
> ileSystem.java:180)
>
>            at
>
>
org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:379
> )
>
>            at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:278)
>
>
>
> But the HMaster is starting correctly.  Here only I have only 2 datanodes
> and the replication factor is 2.
>
>
>
> Regards
>
> Ram
>
>
>
>
>
****************************************************************************
> ***********
> This e-mail and attachments contain confidential information from HUAWEI,
> which is intended only for the person or entity whose address is listed
> above. Any use of the information contained herein in any way (including,
> but not limited to, total or partial disclosure, reproduction, or
> dissemination) by persons other than the intended recipient's) is
> prohibited. If you receive this e-mail in error, please notify the sender
> by
> phone or email immediately and delete it!
>
>
>
>

Reply via email to