[
https://issues.apache.org/jira/browse/FLINK-33288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17776041#comment-17776041
]
Xin Chen commented on FLINK-33288:
----------------------------------
This is a serious bug. Specifically, when adding the following configuration to
the namenodeļ¼
{code:java}
<property>
<name>dfs.namenode.fs-limits.max-directory-items</name>
<value>50</value>
</property>
{code}
Once the number of empty subdirectories remaining in 'high availability.
storageDir' exceeds 50, Flink can not continue submitting tasks, with an error
as follows:
{code:java}
Caused by:
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.protocol.FSLimitException$MaxDirectoryItemsExceededException):
The directory item limit of /flink/recovery/ is exceeded: limit=50 items=50
{code}
> Empty directory residue with appid name in HA(highly-available) related
> directory of hdfs, not cleaned
> ------------------------------------------------------------------------------------------------------
>
> Key: FLINK-33288
> URL: https://issues.apache.org/jira/browse/FLINK-33288
> Project: Flink
> Issue Type: Bug
> Components: Runtime / Configuration
> Affects Versions: 1.16.2, 1.17.1
> Reporter: Xin Chen
> Priority: Major
> Attachments: screenshot-1.png
>
>
> When I submitted a large number of tasks in Flink-on-Yarn mode and
> successfully executed, I unexpectedly found a large number of empty
> directories left in the directory related to 'high availability.storageDir'
> on hdfs, with appids as shown below. I believe this must be cleared! However,
> after verification in the environments of 1.16.2 and 1.17.1, it was proven
> that neither of them solved this problem.
> my flink-conf.yaml about 'high availability.storageDir':
> {code:java}
> high-availability.storageDir: hdfs://hdfsHACluster/flink/recovery
> {code}
> !screenshot-1.png!
--
This message was sent by Atlassian Jira
(v8.20.10#820010)