On Thursday, October 12, 2017 at 6:55:03 PM UTC+2, paolo.m...@sparkfabrik.com wrote: > On Thursday, October 12, 2017 at 6:53:28 PM UTC+2, paolo.m...@sparkfabrik.com > wrote: > > On Thursday, October 12, 2017 at 6:49:01 PM UTC+2, > > paolo.m...@sparkfabrik.com wrote: > > > We are experiencing networking problems over an entire pool of nodes, > > > they seems to have network issues causing also the kube-proxy to fail, > > > the nodes are freshly provisioned as they are configured to autoscale, > > > i've also tried to delete the entire pool but the problem persists, also > > > for new pools. > > > > > > > > > Kubectl output: > > > > > > kube-system > > > kube-proxy-gke-spark-op-services-gitlab-ci-build-fb120c5e-q331 0/1 > > > Init:ImagePullBackOff 0 18m > > > > > > The issue is quite urgent. > > > > > > What happens is that the docker containers provisioned on the nodes (by > > gitlab in this case) suffer of networking issues, i suppose because the > > kube-proxy is missing. > > The kubectl output > > spark-k8s-services git:dev ❯ kubectl logs -f > kube-proxy-gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 -nkube-system > > > Error from server (BadRequest): container "kube-proxy" in pod > "kube-proxy-gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7" is waiting > to start: PodInitializing > > > spark-k8s-services git:dev ❯ kubectl describe pod > kube-proxy-gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 -nkube-system > > Name: kube-proxy-gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 > Namespace: kube-system > Node: gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7/10.132.0.8 > Start Time: Thu, 12 Oct 2017 18:52:22 +0200 > Labels: component=kube-proxy > tier=node > Annotations: kubernetes.io/config.hash=e1aba4d0cdf8ee0f1b89b21ea6b46704 > kubernetes.io/config.mirror=e1aba4d0cdf8ee0f1b89b21ea6b46704 > kubernetes.io/config.seen=2017-10-12T16:52:17.72294518Z > kubernetes.io/config.source=file > scheduler.alpha.kubernetes.io/critical-pod= > Status: Pending > IP: 10.132.0.8 > Init Containers: > touch-lock: > Container ID: > Image: busybox > Image ID: > Port: <none> > Command: > /bin/touch > /run/xtables.lock > State: Waiting > Reason: ImagePullBackOff > Ready: False > Restart Count: 0 > Environment: <none> > Mounts: > /run from run (rw) > Containers: > kube-proxy: > Container ID: > Image: gcr.io/google_containers/kube-proxy:v1.7.6 > Image ID: > Port: <none> > Command: > /bin/sh > -c > echo -998 > /proc/$$$/oom_score_adj && kube-proxy > --master=https://104.199.47.251 --kubeconfig=/var/lib/kube-proxy/kubeconfig > --cluster-cidr=10.0.0.0/14 --resource-container="" --v=2 > --feature-gates=ExperimentalCriticalPodAnnotation=true > --iptables-sync-period=1m --iptables-min-sync-period=10s > 1>>/var/log/kube-proxy.log 2>&1 > State: Waiting > Reason: PodInitializing > Ready: False > Restart Count: 0 > Requests: > cpu: 100m > Environment: <none> > Mounts: > /etc/ssl/certs from etc-ssl-certs (ro) > /run/xtables.lock from iptableslock (rw) > /usr/share/ca-certificates from usr-ca-certs (ro) > /var/lib/kube-proxy/kubeconfig from kubeconfig (rw) > /var/log from varlog (rw) > Conditions: > Type Status > Initialized False > Ready False > PodScheduled True > Volumes: > usr-ca-certs: > Type: HostPath (bare host directory volume) > Path: /usr/share/ca-certificates > etc-ssl-certs: > Type: HostPath (bare host directory volume) > Path: /etc/ssl/certs > kubeconfig: > Type: HostPath (bare host directory volume) > Path: /var/lib/kube-proxy/kubeconfig > varlog: > Type: HostPath (bare host directory volume) > Path: /var/log > run: > Type: HostPath (bare host directory volume) > Path: /run > iptableslock: > Type: HostPath (bare host directory volume) > Path: /run/xtables.lock > QoS Class: Burstable > Node-Selectors: <none> > Tolerations: :NoExecute > Events: <none>
A dump of k8s events: ``` kube-system 2017-10-12 18:58:11 +0200 CEST 2017-10-12 18:52:24 +0200 CEST 6 kube-proxy-gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 Pod spec.initContainers{touch-lock} Normal Pulling kubelet, gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 pulling image "busybox" kube-system 2017-10-12 18:58:11 +0200 CEST 2017-10-12 18:52:31 +0200 CEST 6 kube-proxy-gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 Pod spec.initContainers{touch-lock} Warning Failed kubelet, gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 Failed to pull image "busybox": rpc error: code = 2 desc = error pulling image configuration: unknown blob kube-system 2017-10-12 18:58:11 +0200 CEST 2017-10-12 18:52:31 +0200 CEST 36 kube-proxy-gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 Pod Warning FailedSync kubelet, gke-spark-op-services-gitlab-ci-build-fb120c5e-tct7 Error syncing pod ``` -- You received this message because you are subscribed to the Google Groups "Kubernetes user discussion and Q&A" group. To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-users+unsubscr...@googlegroups.com. To post to this group, send email to kubernetes-users@googlegroups.com. Visit this group at https://groups.google.com/group/kubernetes-users. For more options, visit https://groups.google.com/d/optout.