Hello, I am running a small OKD 3.10 cluster on CentOS 7.6 and just noticed that the controller-manager on my master node does not seem to start properly as you can see from the "oc get pods" output below:
NAMESPACE NAME READY STATUS RESTARTS AGE kube-service-catalog controller-manager-cw69k 0/1 CrashLoopBackOff 10905 117d The relevant log entry from /var/log/messages is the following: Dec 12 09:17:49 master origin-node: I1212 09:17:49.977148 3626 kuberuntime_manager.go:513] Container {Name:controller-manager Image:docker.io/openshift/origin-service-catalog:v3.10.0 Command:[/usr/bin/service-catalog] Args:[controller-manager --secure-port 6443 -v 3 --leader-election-namespace kube-service-catalog --leader-elect-resource-lock configmaps --cluster-id-configmap-namespace=kube-service-catalog --broker-relist-interval 5m --feature-gates OriginatingIdentity=true --feature-gates AsyncBindingOperations=true] WorkingDir: Ports:[{Name: HostPort:0 ContainerPort:6443 Protocol:TCP HostIP:}] EnvFrom:[] Env:[{Name:K8S_NAMESPACE Value: ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,}}] Resources:{Limits:map[] Requests:map[]} VolumeMounts:[{Name:service-catalog-ssl ReadOnly:true MountPath:/var/run/kubernetes-service-catalog SubPath: MountPropagation:<nil>} {Name:service-c atalog-controller-token-8mvfj ReadOnly:true MountPath:/var/run/secrets/kubernetes.io/serviceaccount SubPath: MountPropagation:<nil>}] VolumeDevices:[] LivenessProbe:nil ReadinessProbe:nil Lifecycle:nil TerminationMessagePath:/dev/termination-log TerminationMessagePolicy:File ImagePullPolicy:IfNotPresent SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[KILL MKNOD SETGID SETUID],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000100000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,} Stdin:false StdinOnce:false TTY:false} is dead, but RestartPolicy says that we should restart it. Dec 12 09:17:49 master origin-node: I1212 09:17:49.977314 3626 kuberuntime_manager.go:757] checking backoff for container "controller-manager" in pod "controller-manager-cw69k_kube-service-catalog(6634c5e5-a132-11e8-8c17-00163e0302cc)" Dec 12 09:17:49 masterorigin-node: I1212 09:17:49.977490 3626 kuberuntime_manager.go:767] Back-off 5m0s restarting failed container=controller-manager pod=controller-manager-cw69k_kube-service-catalog(6634c5e5-a132-11e8-8c17-00163e0302cc) Dec 12 09:17:49 master origin-node: E1212 09:17:49.977551 3626 pod_workers.go:186] Error syncing pod 6634c5e5-a132-11e8-8c17-00163e0302cc ("controller-manager-cw69k_kube-service-catalog(6634c5e5-a132-11e8-8c17-00163e0302cc)"), skipping: failed to "StartContainer" for "controller-manager" with CrashLoopBackOff: "Back-off 5m0s restarting failed container=controller-manager pod=controller-manager-cw69k_kube-service-catalog(6634c5e5-a132-11e8-8c17-00163e0302cc)" Going into more log details by running "master logs controllers controllers" show the following errors: I1212 08:17:35.011521 1 healthz.go:72] /healthz/log check W1212 08:17:37.585956 1 reflector.go:341] github.com/openshift/origin/vendor/k8s.io/client-go/informers/factory.go:87: watch of *v1beta1.PodDisruptionBudget ended with: The resourceVersion for the provided watch is too old. W1212 08:17:38.587939 1 reflector.go:341] github.com/openshift/origin/vendor/k8s.io/client-go/informers/factory.go:87: watch of *v1beta1.StatefulSet ended with: The resourceVersion for the provided watch is too old. W1212 08:17:38.688482 1 reflector.go:341] github.com/openshift/origin/vendor/k8s.io/client-go/informers/factory.go:87: watch of *v1beta1.CronJob ended with: The resourceVersion for the provided watch is too old. I1212 08:17:45.010840 1 healthz.go:72] /healthz/log check W1212 08:17:50.609257 1 reflector.go:341] github.com/openshift/origin/vendor/github.com/openshift/client-go/security/informers/externalversions/factory.go:58: watch of *v1.RangeAllocation ended with: The resourceVersion for the provided watch is too old. Does anyone have an idea what could be going wrong here? Here is the output of "oc version": oc v3.10.0+0c4577e-1 kubernetes v1.10.0+b81c8f8 features: Basic-Auth GSSAPI Kerberos SPNEGO Regards, Mab _______________________________________________ users mailing list users@lists.openshift.redhat.com http://lists.openshift.redhat.com/openshiftmm/listinfo/users