We recently had an instance check failure on one of our k8s 1.6.2 nodes, resulting in it being replaced by the AWS Autoscaler. The new node booted and joined the cluster, but pods scheduled to it are unable to access network resources outside of the cluster unless host network is enabled for the pod.
SSHing into the host OS, networking is properly configured and I can perform any arbitrary network operation (apt-get update, curl ..., telnet, ssh, etc etc) to any destination. Entering a pod that has hostNetworking: true enabled, this continues to be the case: all network connections to any arbitrary destination on the internet will be successful. Entering a pod that does not, however, I can only connect to other IPs inside the cluster. $ kubectl get svc NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE [ .. redacted .. ] feature-flip 100.71.179.196 <nodes> 9292:30800/TCP 27d $ curl http://feature-flip:9292 {"status": "ok"} $ wget https://www.google.com/ Connecting to www.google.com (172.217.7.228:80) wget: can't connect to remote host (172.217.7.228): Connection timed out On the same host, a pod with hostNetworking will succeed. I'm using flannel networking, and the hosts that don't work are part of an EC2 ASG with the hosts that do work. The base AMI and launchconfiguration are the same. Comparing a variety of possible avenues of difference, I have found no obvious difference between the working and non-working nodes. -- You received this message because you are subscribed to the Google Groups "Kubernetes user discussion and Q&A" group. To unsubscribe from this group and stop receiving emails from it, send an email to kubernetes-users+unsubscr...@googlegroups.com. To post to this group, send email to kubernetes-users@googlegroups.com. Visit this group at https://groups.google.com/group/kubernetes-users. For more options, visit https://groups.google.com/d/optout.