Hi, We are trying to use Flink Kubernetes Operator 1.4.0 with Flink 1.16.
However, at the job-manager deployment step we get the following error: ``` Exception in thread "main" java.lang.NullPointerException at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.shutDownAsync(ClusterEntrypoint.java:585) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:242) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runClusterEntrypoint(ClusterEntrypoint.java:729) at org.apache.flink.kubernetes.entrypoint.KubernetesApplicationClusterEntrypoint.main(KubernetesApplicationClusterEntrypoint.java:86) ``` It sems it is related to the following line: ``` this.clusterId = checkNotNull(flinkConfig.getString(KubernetesConfigOptions.CLUSTER_ID), "ClusterId must be specified!"); ``` We specified the CLUSTER_ID but it seems that the flinkConfig object is not handled correctly. We have the following flinkConfiguration defined in deployment.yaml: ``` spec: flinkConfiguration: execution.checkpointing.externalized-checkpoint-retention: RETAIN_ON_CANCELLATION execution.checkpointing.interval: 120s execution.checkpointing.min-pause: 120s execution.checkpointing.mode: AT_LEAST_ONCE execution.checkpointing.snapshot-compression: "false" execution.checkpointing.timeout: 3000s execution.checkpointing.tolerable-failed-checkpoints: "5" execution.checkpointing.unaligned: "false" fs.hdfs.hadoopconf: /opt/hadoop-conf/ high-availability.storageDir: gs://<path/to/environment>/ha high-availability: kubernetes high-availability.cluster-id: <cluster-id> kubernetes.operator.periodic.savepoint.interval: 6h kubernetes.operator.savepoint.history.max.age: 72h kubernetes.operator.savepoint.history.max.count: "15" metrics.reporter.prom.class: org.apache.flink.metrics.prometheus.PrometheusReporter metrics.reporter.prom.port: "2112" metrics.reporters: prom rest.flamegraph.enabled: "false" state.backend: rocksdb state.backend.incremental: "false" state.backend.rocksdb.localdir: /rocksdb state.checkpoint-storage: filesystem state.checkpoints.dir: gs://<path/to/environment>/checkpoints state.savepoints.dir: gs://<path/to/environment>/savepoints taskmanager.memory.managed.fraction: "0" taskmanager.network.memory.buffer-debloat.enabled: "false" taskmanager.network.memory.buffer-debloat.period: "200" taskmanager.network.memory.buffers-per-channel: "2" taskmanager.network.memory.floating-buffers-per-gate: "8" taskmanager.network.memory.max-buffers-per-channel: "10" taskmanager.network.sort-shuffle.min-buffers: "512" taskmanager.numberOfTaskSlots: "1" kubernetes.taskmanager.cpu.limit-factor: "4" kubernetes.taskmanager.cpu: "0.5" kubernetes.cluster-id: <cluster-id> ``` Have someone encountered the issue before? Thanks, PB