After looking a bit more into it, at least for some of the errors, the issue seems to be caused by someone manually deleting the flinkdeployments, leaving the autoscaler configmaps orphaned so to say.
I guess one could delete those configmaps too but is that expected that the autoscaler puts the new jobs into failed state vs trying to handle the situation more gracefully? Regards, Salva On 2025/07/22 10:26:11 Salva Alcántara wrote: > I've enabled the Autoscaler in the Flink Kubernetes Operator for several > jobs and since then I'm observing the following errors: > > ``` > Message: 1 AutoscalerError: Failure executing: PUT at: > https://100.64.0.1:443/api/v1/namespaces/my-namespace/configmaps/autoscaler-my-job . > Message: Operation cannot be fulfilled on configmaps "autoscaler-my-job": > StorageError: invalid object, Code: 4, Key: > /registry/configmaps/my-namespace/autoscaler-my-job, ResourceVersion: 0, > AdditionalErrorMsg: Precondition failed: UID in precondition: > a77dcdcb-6fea-4c3e-bae5-ad876c67c50e, UID in object meta: . Received > status: Status(apiVersion=v1, code=409, details=StatusDetails(causes=[], > group=null, kind=configmaps, name=autoscaler-my-job, > retryAfterSeconds=null, uid=null, additionalProperties={}), kind=Status, > message=Operation cannot be fulfilled on configmaps "autoscaler-my-job": > StorageError: invalid object, Code: 4, Key: > /registry/configmaps/my-namespace/autoscaler-my-job, ResourceVersion: 0, > AdditionalErrorMsg: Precondition failed: UID in precondition: > a77dcdcb-6fea-4c3e-bae5-ad876c67c50e, UID in object meta: , > metadata=ListMeta(_continue=null, remainingItemCount=null, > resourceVersion=null, selfLink=null, additionalProperties={}), > reason=Conflict, status=Failure, additionalProperties={}). > ``` > > The affected jobs get into FAILED state after that. > > Do you have any ideas why this is happening? Is there a way to > automatically resolve these issues? Ideally the operator should resolve the > conflict and go on from there vs failing... > > Regards, > > Salva >