Re: Not cleanup Kubernetes Configmaps after execution success

2021-10-31 Thread Hua Wei Chen
HI Yang, Roman, Thanks for the information and sorry for the late reply. Seems like the Kubernetes node restarted during the Flink finalization stage. I think that is the root cause. Regards, Oscar On Wed, Oct 27, 2021 at 4:20 PM Yang Wang wrote: > Hi, > > I think Roman is right. It seems

Re: Not cleanup Kubernetes Configmaps after execution success

2021-10-27 Thread Yang Wang
Hi, I think Roman is right. It seems that the JobManager is relaunched again by K8s after Flink has already deregister the application(aka delete the JobManager K8s deployment). One possible reason might be that kubelet is too late to know the JobManager deployment is deleted. So it relaunch the

Re: Not cleanup Kubernetes Configmaps after execution success

2021-10-26 Thread Roman Khachatryan
Thanks for sharing this, The sequence of events the log seems strange to me: 2021-10-17 03:05:55,801 INFO org.apache.flink.runtime.jobmaster.JobMaster [] - Close ResourceManager connection c1092812cfb2853a5576ffd78e346189: Stopping JobMaster for job 'rt-match_12.4.5_8d48b21a'

Re: Not cleanup Kubernetes Configmaps after execution success

2021-10-25 Thread Yang Wang
Hi Hua Wei, I think you need to share the JobManager logs so that we could check whether Flink had tried to clean up the HA related ConfigMaps. Using the "kubectl logs -f >/tmp/log.jm" could help with dumping the logs. Best, Yang Roman Khachatryan 于2021年10月25日周一 下午5:35写道: > Hi Hua, > > It

Re: Not cleanup Kubernetes Configmaps after execution success

2021-10-25 Thread Roman Khachatryan
Hi Hua, It looks like the ConfigMap misses HA labels for some reason. Could you confirm that you are running in HA mode? Which deployment mode are you using? [1] I'm also pulling in Yan Wang who might know this area better. [1]

Not cleanup Kubernetes Configmaps after execution success

2021-10-24 Thread Hua Wei Chen
Hi all, We have Flink jobs run on batch mode and get the job status via JobHandler. onJobExecuted