Abey Sam Alex created FLINK-4674:
------------------------------------
Summary: File Descriptors not being released after Completion of
Flink Job Run via Flink Web Portal
Key: FLINK-4674
URL: https://issues.apache.org/jira/browse/FLINK-4674
Project: Flink
Issue Type: Bug
Components: Client, JobManager, TaskManager
Affects Versions: 1.1.2, 1.1.1, 1.0.3, 1.0.2, 1.0.1, 1.1.0
Environment: RHEL6,7, UBUNTU
Reporter: Abey Sam Alex
File descriptors utilized by Flink Task Manager are not released even after
completion of job.
For releasing all file descriptors, we need to reboot the flink cluster. This
causes all Jobs to run succesfully until the OS limit is hit and post which Job
keeps failing -
Error on Flink -
java.io.IOException: Error opening the Input Split
file:/data/Temp/RUN10_1000.csv [84950,1699]: /data/Temp/RUN10_1000.csv (Too
many open files)
at
org.apache.flink.api.common.io.FileInputFormat.open(FileInputFormat.java:682)
at
org.apache.flink.api.common.io.DelimitedInputFormat.open(DelimitedInputFormat.java:411)
at
org.apache.flink.api.common.io.DelimitedInputFormat.open(DelimitedInputFormat.java:45)
at
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:147)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.FileNotFoundException: /data/Temp/RUN10_1000.csv (Too many
open files)
at java.io.FileInputStream.open0(Native Method)
at java.io.FileInputStream.open(FileInputStream.java:195)
at java.io.FileInputStream.<init>(FileInputStream.java:138)
at
org.apache.flink.core.fs.local.LocalDataInputStream.<init>(LocalDataInputStream.java:52)
at
org.apache.flink.core.fs.local.LocalFileSystem.open(LocalFileSystem.java:143)
at
org.apache.flink.api.common.io.FileInputFormat$InputSplitOpenThread.run(FileInputFormat.java:842)
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)