Sihua Zhou commented on FLINK-9120:

Hi [~dhirajpraj] I think maybe we should leave this issue opening ... because 
increasing the no. of retries is just a workaround for this "bug". Is this 
should be a bug or this is what we expected for standalone model? 

> Task Manager Fault Tolerance issue
> ----------------------------------
>                 Key: FLINK-9120
>                 URL: https://issues.apache.org/jira/browse/FLINK-9120
>             Project: Flink
>          Issue Type: Bug
>          Components: Cluster Management, Configuration, Core
>    Affects Versions: 1.4.2
>            Reporter: dhiraj prajapati
>            Priority: Critical
>         Attachments: flink-dhiraj.prajapati-client-ip-10-14-25-115.log, 
> flink-dhiraj.prajapati-client-ip-10-14-25-115.log, 
> flink-dhiraj.prajapati-jobmanager-5-ip-10-14-25-115.log, 
> flink-dhiraj.prajapati-jobmanager-5-ip-10-14-25-115.log, 
> flink-dhiraj.prajapati-taskmanager-5-ip-10-14-25-116.log, 
> flink-dhiraj.prajapati-taskmanager-5-ip-10-14-25-116.log
> HI, 
> I have set up a flink 1.4 cluster with 1 job manager and two task managers. 
> The configs taskmanager.numberOfTaskSlots and parallelism.default were set 
> to 2 on each node. I submitted a job to this cluster and it runs fine. To 
> test fault tolerance, I killed one task manager. I was expecting the job to 
> run fine because one of the 2 task managers was still up and running. 
> However, the job failed. Am I missing something? 

This message was sent by Atlassian JIRA

Reply via email to