[ 
https://issues.apache.org/jira/browse/FLINK-13590?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16901802#comment-16901802
 ] 

shuwenjun edited comment on FLINK-13590 at 8/7/19 10:02 AM:
------------------------------------------------------------

Hi Yang,

You are right, it's right that the files are created when launching a new yarn 
container. And it is better optimized to override taskmanager-conf.yaml. Now, 
it could cause a more serious problem, that it will produce hundreds of 
thousands of small files, because the RM keep retrying to apply for 
taskmanager(container) when the source is enough. And the RM always don't get 
some new correct taskmanagers.

 

Thank you for your comment.

 

There are some logs.
{code:java}
//代码占位符
2019-08-07 16:54:52,729 INFO 
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy - 
Opening proxy : hadoopxxx
2019-08-07 16:54:58,171 INFO org.apache.flink.yarn.YarnResourceManager - 
Requesting new TaskExecutor container with resources <memory:8192, vCores:2>. 
Number pending requests 1.
2019-08-07 16:54:58,171 INFO org.apache.flink.yarn.YarnResourceManager - 
Requesting new TaskExecutor container with resources <memory:8192, vCores:2>. 
Number pending requests 2.
2019-08-07 16:54:58,171 INFO org.apache.flink.yarn.YarnResourceManager - 
Requesting new TaskExecutor container with resources <memory:8192, vCores:2>. 
Number pending requests 3.
2019-08-07 16:55:03,675 INFO org.apache.flink.yarn.YarnResourceManager - 
Received new container: container_xxx_01_007684 - Remaining pending container 
requests: 3
2019-08-07 16:55:03,675 INFO org.apache.flink.yarn.YarnResourceManager - 
Removing container request Capability[<memory:8192, vCores:2>]Priority[1]. 
Pending container requests 2.
2019-08-07 16:55:03,676 INFO org.apache.flink.yarn.YarnResourceManager - Adding 
keytab hdfs://xxxx.keytab to the AM container local resource bucket
2019-08-07 16:55:03,740 INFO org.apache.flink.yarn.YarnResourceManager - 
Creating container launch context for TaskManagers
2019-08-07 16:55:03,740 INFO org.apache.flink.yarn.YarnResourceManager - 
Starting TaskManagers
2019-08-07 16:55:03,742 INFO 
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy - 
Opening proxy : hadoopxxx
2019-08-07 16:55:03,748 INFO org.apache.flink.yarn.YarnResourceManager - 
Received new container: container_xxx_01_007685 - Remaining pending container 
requests: 2
2019-08-07 16:55:03,748 INFO org.apache.flink.yarn.YarnResourceManager - 
Removing container request Capability[<memory:8192, vCores:2>]Priority[1]. 
Pending container requests 1.
2019-08-07 16:55:03,748 INFO org.apache.flink.yarn.YarnResourceManager - Adding 
keytab hdfs://xxxx.keytab to the AM container local resource bucket
2019-08-07 16:55:03,761 INFO org.apache.flink.yarn.YarnResourceManager - 
Creating container launch context for TaskManagers
2019-08-07 16:55:03,761 INFO org.apache.flink.yarn.YarnResourceManager - 
Starting TaskManagers
2019-08-07 16:55:03,761 INFO 
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy - 
Opening proxy : hadoopxxx
2019-08-07 16:55:03,767 INFO org.apache.flink.yarn.YarnResourceManager - 
Received new container: container_xxx_01_007686 - Remaining pending container 
requests: 1
2019-08-07 16:55:03,767 INFO org.apache.flink.yarn.YarnResourceManager - 
Removing container request Capability[<memory:8192, vCores:2>]Priority[1]. 
Pending container requests 0.
2019-08-07 16:55:03,767 INFO org.apache.flink.yarn.YarnResourceManager - Adding 
keytab hdfs://xxxx.keytab to the AM container local resource bucket
2019-08-07 16:55:03,814 INFO org.apache.flink.yarn.YarnResourceManager - 
Creating container launch context for TaskManagers
2019-08-07 16:55:03,814 INFO org.apache.flink.yarn.YarnResourceManager - 
Starting TaskManagers
2019-08-07 16:55:03,814 INFO 
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy - 
Opening proxy : hadoopxxx
2019-08-07 16:55:09,178 INFO org.apache.flink.yarn.YarnResourceManager - 
Requesting new TaskExecutor container with resources <memory:8192, vCores:2>. 
Number pending requests 1.
2019-08-07 16:55:09,178 INFO org.apache.flink.yarn.YarnResourceManager - 
Requesting new TaskExecutor container with resources <memory:8192, vCores:2>. 
Number pending requests 2.
2019-08-07 16:55:09,178 INFO org.apache.flink.yarn.YarnResourceManager - 
Requesting new TaskExecutor container with resources <memory:8192, vCores:2>. 
Number pending requests 3.
2019-08-07 16:55:14,682 INFO org.apache.flink.yarn.YarnResourceManager - 
Received new container: container_xxx - Remaining pending container requests: 3
2019-08-07 16:55:14,683 INFO org.apache.flink.yarn.YarnResourceManager - 
Removing container request Capability[<memory:8192, vCores:2>]Priority[1]. 
Pending container requests 2.
2019-08-07 16:55:14,683 INFO org.apache.flink.yarn.YarnResourceManager - Adding 
keytab hdfs://xxxx.keytab to the AM container local resource bucket
2019-08-07 16:55:14,778 INFO org.apache.flink.yarn.YarnResourceManager - 
Creating container launch context for TaskManagers
2019-08-07 16:55:14,778 INFO org.apache.flink.yarn.YarnResourceManager - 
Starting TaskManagers
{code}
 


was (Author: [email protected]):
Hi Yang,

You are right, it's right that the files are created when launching a new yarn 
container. And it is better optimized to override taskmanager-conf.yaml. Now, 
it could cause a more serious problem, that it will produce hundreds of 
thousands of small files, because the RM keep retrying to apply for 
taskmanager(container) when the source is enough. And the RM always don't get 
some new correct taskmanagers.

 

Thank you for your comment.

 

> flink-on-yarn sometimes could create many little files that are 
> xxx-taskmanager-conf.yaml
> -----------------------------------------------------------------------------------------
>
>                 Key: FLINK-13590
>                 URL: https://issues.apache.org/jira/browse/FLINK-13590
>             Project: Flink
>          Issue Type: Bug
>          Components: Deployment / YARN
>            Reporter: shuwenjun
>            Priority: Major
>         Attachments: taskmanager-conf-yaml.png
>
>
> Both of 1.7.2 and 1.8.0 are used, but they could create many little files.
>  These files are the configuration file of  taskmanager and when the flink 
> session try to apply a new container, one of the files will be created. And I 
> don't know why sometimes the flink session apply container again and again? 
> Or when one container has lost, it could delete its taskmanager-conf.yaml 
>  



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

Reply via email to