Hi Stephan,

it's not a problem, but makes finding other errors on my NameNode
complicated as I have this error message every minute.
Can't we just delete the directory recursively?

Regards,
Max

2016-10-11 17:59 GMT+02:00 Stephan Ewen <se...@apache.org>:

> Hi!
>
> I think to some extend this is expected. There is some cleanup code that
> deletes files and then  issues parent directory remove requests. It relies
> on the fact that the parent directory is only removed if it is empty (after
> the last file was deleted).
>
> Is this a problem right now, or just a confusing behavior?
>
> Greetings,
> Stephan
>
>
> On Tue, Oct 11, 2016 at 5:25 PM, static-max <flasha...@googlemail.com>
> wrote:
>
>> Hi,
>>
>> I get many (multiple times per minute) errors in my Namenode HDFS logfile:
>>
>> 2016-10-11 17:17:07,596 INFO  ipc.Server (Server.java:logException(2401))
>> - IPC Server handler 295 on 8020, call 
>> org.apache.hadoop.hdfs.protocol.ClientProtocol.delete
>> from datanode1:34872 Call#2361 Retry#0
>> org.apache.hadoop.fs.PathIsNotEmptyDirectoryException: `/flink/recovery
>> is non empty': Directory is not empty
>>         at org.apache.hadoop.hdfs.server.namenode.FSDirDeleteOp.delete(
>> FSDirDeleteOp.java:89)
>>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(F
>> SNamesystem.java:3829)
>>         at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.del
>> ete(NameNodeRpcServer.java:1071)
>>         at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServ
>> erSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTr
>> anslatorPB.java:619)
>>         at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocol
>> Protos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNam
>> enodeProtocolProtos.java)
>>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcIn
>> voker.call(ProtobufRpcEngine.java:640)
>>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313)
>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309)
>>         at java.security.AccessController.doPrivileged(Native Method)
>>         at javax.security.auth.Subject.doAs(Subject.java:422)
>>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGro
>> upInformation.java:1724)
>>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307)
>>
>> That is the directory I configured for Jobmanager HA. I deleted it before
>> starting the YARN session but that did not help. The folder gets created by
>> Flink without problems.
>>
>> I'm using latest Flink Master (Commit: 6731ec1) and build it for Hadoop
>> 2.7.3.
>>
>> Any idea is highly appreciated. Thanks a lot!
>>
>
>

Reply via email to