Hi,

Please also check for coalesced RDD. I encountered the same error while
writing a coalesced rdd/df to HDFS. If this is the case, please use
repartition instead.

Sent from OnePlus 3

Thanks & Regards,
Gaurav Kumar

On Oct 17, 2016 11:22 AM, "Zhang Jianfeng" <jzhang...@gmail.com> wrote:

Thanks Rakesh for your kind help. Actually during the job only one reducer
result file (for example part-r-2) had this error, other reducers worked
well.

Best Regards,
Jian Feng

2016-10-17 11:49 GMT+08:00 Rakesh Radhakrishnan <rake...@apache.org>:

> Hi Jian Feng,
>
> Could you please check your code and see any possibilities of simultaneous
> access to the same file. Mostly this situation happens when multiple
> clients tries to access the same file.
>
> Code Reference:- https://github.com/apache/hadoop/blob/branch-2.
> 2/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/
> hadoop/hdfs/server/namenode/FSNamesystem.java#L2737
>
> Best Regards,
> Rakesh
> Intel
>
> On Mon, Oct 17, 2016 at 7:16 AM, Zhang Jianfeng <jzhang...@gmail.com>
> wrote:
>
>> Hi ,
>>
>>     I hit an wired error. On our hadoop cluster (2.2.0), occasionally a
>> LeaseExpiredException is thrown.
>>
>> The stacktrace is as below:
>>
>>
>> *org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException):
>> No lease on /user/biadmin/analytic‐root/SX5XPWPPDPQH/.executions/.at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2737)*
>>
>> *at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFileInternal(FSNamesystem.java:2801)*
>>
>> *at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFile(FSNamesystem.java:2783)*
>>
>> *at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.com
>> <http://namenode.NameNodeRpcServer.com>plete(NameNodeRpcServer.java:611)*
>>
>> *at
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.complete(ClientNamenodeProtocolServerSideTranslatorPB.java:428)*
>>
>> *at
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59586)*
>>
>> *at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)*
>>
>> *at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)*
>>
>> *at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2048)*
>>
>> *at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044)*
>>
>> *at
>> java.security.AccessController.doPrivileged(AccessController.java:310)*
>>
>> *at javax.security.auth.Subject.do
>> <http://javax.security.auth.Subject.do>As(Subject.java:573)*
>>
>> *at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1502)*
>>
>> *at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2042)*
>>
>> *at org.apache.hadoop.ipc.Client.call(Client.java:1347)*
>>
>> *at org.apache.hadoop.ipc.Client.call(Client.java:1300)*
>>
>> *at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)*
>>
>> *at $Proxy7.complete(Unknown Source)*
>>
>> *at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)*
>>
>> *at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:60)*
>>
>> *at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:37)*
>>
>> at java.lang.reflect.Method.invoke(Method.java:611)
>>
>> at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMeth
>> od(RetryInvocationHandler.java:186)
>>
>> at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(Ret
>> ryInvocationHandler.java:102)
>>
>> at $Proxy7.complete(Unknown Source)
>>
>> at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTran
>> slatorPB.complete(ClientNamenodeProtocolTranslatorPB.java:371)
>>
>> at org.apache.hadoop.hdfs.DFSOutputStream.completeFile(DFSOutpu
>> tStream.java:1894)
>>
>> at org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream
>> .java:1881)
>>
>> at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(
>> FSDataOutputStream.java:71)
>>
>> at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputSt
>> ream.java:104)
>>
>> at java.io.FilterOutputStream.close(FilterOutputStream.java:154)
>>
>> Any help will be appreciated!
>>
>> --
>> Best Regards,
>> Jian Feng
>>
>
>


-- 
Best Regards,
Jian Feng

Reply via email to