hjw created FLINK-27245: --------------------------- Summary: Flink job on Yarn cannot revover when zookeeper in Exception Key: FLINK-27245 URL: https://issues.apache.org/jira/browse/FLINK-27245 Project: Flink Issue Type: Bug Components: Runtime / Coordination Affects Versions: 1.17.0 Environment: Flink :1.17.2 Hdfs:3.1.1 zookeeper:3.5.1
HA defined in Flink-conf,yaml: flink.security.enable: true fs.output.always-create-directory: false fs.overwrite-files: false high-availability.job.delay: 10 s high-availability.storageDir: hdfs:///flink/recovery high-availability.zookeeper.client.acl: creator high-availability.zookeeper.client.connection-timeout: 15000 high-availability.zookeeper.client.max-retry-attempts: 3 high-availability.zookeeper.client.retry-wait: 5000 high-availability.zookeeper.client.session-timeout: 60000 high-availability.zookeeper.path.root: /flink high-availability.zookeeper.quorum: zk01:24002,zk02:24002,zk03:24002 high-availability: zookeeper Reporter: hjw Flink job cannot revover when zookeeper in Exception. I noticed that the data in high-availability.storageDir deleled when Job failed , resulting in failure when pulling up again. {code:java} (SmarterLeaderLatch.java:570) 2022-04-07 19:54:29,002 | INFO | [Suspend state waiting handler] | Connection to Zookeeper is SUSPENDED. Wait it to be back. Already waited 10 seconds. | org.apache.flink.runtime.leaderelection.SmarterLeaderLatch (SmarterLeaderLatch.java:570) 2022-04-07 19:54:29,004 | INFO | [Suspend state waiting handler] | Connection to Zookeeper is SUSPENDED. Wait it to be back. Already waited 10 seconds. | org.apache.flink.runtime.leaderelection.SmarterLeaderLatch (SmarterLeaderLatch.java:570) 2022-04-07 19:54:29,004 | INFO | [Suspend state waiting handler] | Connection to Zookeeper is SUSPENDED. Wait it to be back. Already waited 10 seconds. | org.apache.flink.runtime.leaderelection.SmarterLeaderLatch (SmarterLeaderLatch.java:570) 2022-04-07 19:54:30,002 | INFO | [Suspend state waiting handler] | Connection to Zookeeper is SUSPENDED. Wait it to be back. Already waited 11 seconds. | org.apache.flink.runtime.leaderelection.SmarterLeaderLatch (SmarterLeaderLatch.java:570) 2022-04-07 19:54:30,002 | INFO | [Suspend state waiting handler] | Connection to Zookeeper is SUSPENDED. Wait it to be back. Already waited 11 seconds. | org.apache.flink.runtime.leaderelection.SmarterLeaderLatch (SmarterLeaderLatch.java:570) 2022-04-07 19:54:30,004 | INFO | [Suspend state waiting handler] | Connection to Zookeeper is SUSPENDED. Wait it to be back. Already waited 11 seconds. | org.apache.flink.runtime.leaderelection.SmarterLeaderLatch (SmarterLeaderLatch.java:570) 2022-04-07 19:54:30,004 | INFO | [Suspend state waiting handler] | Connection to Zookeeper is SUSPENDED. Wait it to be back. Already waited 11 seconds. | org.apache.flink.runtime.leaderelection.SmarterLeaderLatch (SmarterLeaderLatch.java:570) 2022-04-07 19:54:30,769 | INFO | [BlobServer shutdown hook] | FileSystemBlobStore cleaning up hdfs:/flink/recovery/application_1625720467511_45233. | org.apache.flink.runtime.blob.FileSystemBlobStor {code} {code:java} 2022-04-07 19:55:29,452 | INFO | [flink-akka.actor.default-dispatcher-4] | Recovered SubmittedJobGraph(1898637f2d11429bd5f5767ea1daaf79, null). | org.apache.flink.runtime.jobmanager.ZooKeeperSubmittedJobGraphStore (ZooKeeperSubmittedJobGraphStore.java:215) 2022-04-07 19:55:29,467 | ERROR | [flink-akka.actor.default-dispatcher-17] | Fatal error occurred in the cluster entrypoint. | org.apache.flink.runtime.entrypoint.ClusterEntrypoint (ClusterEntrypoint.java:408) java.lang.RuntimeException: org.apache.flink.runtime.client.JobExecutionException: Could not set up JobManager at org.apache.flink.util.function.CheckedSupplier.lambda$unchecked$0(CheckedSupplier.java:36) at java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1590) at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:39) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:415) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) Caused by: org.apache.flink.runtime.client.JobExecutionException: Could not set up JobManager at org.apache.flink.runtime.jobmaster.JobManagerRunner.<init>(JobManagerRunner.java:176) at org.apache.flink.runtime.dispatcher.Dispatcher$DefaultJobManagerRunnerFactory.createJobManagerRunner(Dispatcher.java:1058) at org.apache.flink.runtime.dispatcher.Dispatcher.lambda$createJobManagerRunner$5(Dispatcher.java:308) at org.apache.flink.util.function.CheckedSupplier.lambda$unchecked$0(CheckedSupplier.java:34) ... 7 common frames omitted Caused by: java.lang.Exception: Cannot set up the user code libraries: File does not exist: /flink/recovery/application_1625720467511_45233/blob/job_1898637f2d11429bd5f5767ea1daaf79/blob_p-7128d0ae4a06a277e3b1182c99eb616ffd45b590-c90586d4a5d4641fcc0c9e4cab31c131 at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:86) at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:76) at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:153) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1951) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:742) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:439) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:530) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1036) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:928) {code} I get the log of the job failed when zookeeper happend error ,try to restart job manager by yarn and zookeeper . Error happended in 2022/04/07 19:54 thx -- This message was sent by Atlassian Jira (v8.20.1#820001)