> TASK [openshift_control_plane : Wait for control plane pods to appear] ********* > Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 ************ > FAILED - RETRYING: Wait for control plane pods to appear (60 retries left). > FAILED - RETRYING: Wait for control plane pods to appear (59 retries left). > It eventually counts all the way down to zero and fails.
This looks a lot like the issues I saw when the host firewall (iptables) was blocking another OKD all-in-one-host install script [1]. Disabling iptables allowed the installation to continue for my proof of concept "cluster". [1]https://github.com/gshipley/installcentos The other error I had with [1] was it was trying to install a couple of packages (zile and python2-pip) from EPEL with the repo disabled. On Tue, May 28, 2019 at 10:41 AM Jayme <[email protected]> wrote: > Shirly, > > Oh and I should mention that I did verify that NetworkManager was > installed on the master0 VM and enabled/started the second go around. So > that service is there and running. > > # systemctl list-unit-files | grep Network > dbus-org.freedesktop.NetworkManager.service > enabled > NetworkManager-dispatcher.service > enabled > NetworkManager-wait-online.service > enabled > NetworkManager.service > enabled > > On Tue, May 28, 2019 at 11:13 AM Jayme <[email protected]> wrote: > >> Shirly, >> >> I appreciate the help with this. Unfortunately I am still running in to >> the same problem. So far I've tried to install/enable/start NetworkManager >> on the existing "master0" server and re-ran the plans from the installer >> VM. I ran in to the same problem waiting for control plane pods and same >> errors in syslog. >> >> So I wiped everything out, killed the template along with the installer >> and master VMs. On oVirt engine (I am running 4.3.3.7-1 stable) I did have >> ovirt-engine-metrics-1.3.0x rpm installed, no yum updates available on an >> update check. So I installed >> http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm then >> proceeded to install the latest version of ovirt-engine-metrics which gave >> me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine. >> >> After that package was installed I proceeded to follow steps from >> beginning outlined at: >> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >> -- >> I ran in to the docker check issue again (same as my initial email) so I >> disabled that and again got as far as starting control plane pods before >> failure. >> >> Not sure where to go from here at this point. The only thing I can think >> of that I did differently vs the instructions outlined above is that I have >> not crated the wildcard DNS record, however I did set configs to create >> /etc/hosts entries and they /etc/hosts on the machines have the proper IPs >> assigned for all hostnames (automatically added by the ansible plays). >> >> Any ideas how I can get past the plane pods issue? >> >> Thanks! >> >> On Tue, May 28, 2019 at 4:23 AM Shirly Radco <[email protected]> wrote: >> >>> Hi, >>> >>> The latest release of 4.3.z should already include a fix for this issue, >>> ovirt-engine-metrics-1.3.1 rpm. >>> >>> The issue is that it requires the NetworkManagar to be installed, >>> running and enabled for it to work. >>> >>> You can install it manually on the master0 vm , start and enable it or >>> you can also install the updated rpm from the nightly builds if your >>> environment is oVirt 4.2.z: >>> >>> https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm >>> >>> Relevant patches are: >>> https://gerrit.ovirt.org/#/c/99959/ >>> https://gerrit.ovirt.org/#/c/99718/ >>> >>> Best regards, >>> >>> -- >>> >>> Shirly Radco >>> >>> BI Senior Software Engineer >>> >>> Red Hat <https://www.redhat.com/> >>> >>> <https://www.redhat.com/> >>> >>> >>> On Mon, May 27, 2019 at 4:41 PM Jayme <[email protected]> wrote: >>> >>>> I managed to get past that but am running in to another problem later >>>> in the process on the control plane pods to appear task. I thought >>>> perhaps a glitch in the process from the failed docker step previously so >>>> after a few more runs I tried killing everything and restarting the metrics >>>> process again from the very beginning and end up hitting the same issue >>>> with control plane pods even though all other steps/tasks seem to be >>>> working. >>>> >>>> I'm just getting this: >>>> >>>> TASK [openshift_control_plane : Wait for control plane pods to appear] >>>> ********* >>>> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >>>> ************ >>>> FAILED - RETRYING: Wait for control plane pods to appear (60 retries >>>> left). >>>> FAILED - RETRYING: Wait for control plane pods to appear (59 retries >>>> left). >>>> FAILED - RETRYING: Wait for control plane pods to appear (58 retries >>>> left). >>>> FAILED - RETRYING: Wait for control plane pods to appear (57 retries >>>> left). >>>> FAILED - RETRYING: Wait for control plane pods to appear (56 retries >>>> left). >>>> >>>> It eventually counts all the way down to zero and fails. >>>> >>>> In syslog of the master0 server I'm seeing some errors related to cni >>>> config >>>> >>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None >>>> kind=pod force=False all_namespaces=None field_selector=None >>>> namespace=kube-system delete_after=False >>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list >>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150 >>>> cni.go:172] Unable to update cni config: No networks found in >>>> /etc/cni/net.d >>>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150 >>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>>> cni config uninitialized >>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None >>>> kind=pod force=False all_namespaces=None field_selector=None >>>> namespace=kube-system delete_after=False >>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list >>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150 >>>> cni.go:172] Unable to update cni config: No networks found in >>>> /etc/cni/net.d >>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150 >>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>>> cni config uninitialized >>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150 >>>> summary.go:102] Failed to get system container stats for >>>> "/system.slice/origin-node.service": failed to get cgroup stats for >>>> "/system.slice/origin-node.service": failed to get container info for >>>> "/system.slice/origin-node.service": unknown container >>>> "/system.slice/origin-node.service" >>>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150 >>>> cni.go:172] Unable to update cni config: No networks found in >>>> /etc/cni/net.d >>>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150 >>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>>> cni config uninitialized >>>> >>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <[email protected]> wrote: >>>> >>>>> Hi Jayme, >>>>> >>>>> Thank you for reaching out. >>>>> Please try rerunning the ansible playbook. >>>>> If this doesn't work, try adding to the integ.ini in the metrics vm >>>>> openshift_disable_check=docker_storage >>>>> and rerun the ansible playbook again. >>>>> >>>>> Please update how it goes. >>>>> >>>>> Best regards, >>>>> >>>>> -- >>>>> >>>>> Shirly Radco >>>>> >>>>> BI Senior Software Engineer >>>>> >>>>> Red Hat <https://www.redhat.com/> >>>>> >>>>> <https://www.redhat.com/> >>>>> >>>>> >>>>> On Sun, May 26, 2019 at 9:34 PM Jayme <[email protected]> wrote: >>>>> >>>>>> I'm running in to this ansible error during oVirt metrics >>>>>> installation (following procedures at: >>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>>> ) >>>>>> >>>>>> This is happening late in the process, after successfully deploying >>>>>> the installation VM and then running second step from the metrics VM. >>>>>> >>>>>> CHECK [memory_availability : master0.xxxxxx.com] >>>>>> ************************************************************************************************* >>>>>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, "checks": >>>>>> {"disk_availability": {}, "docker_image_availability": {"changed": true}, >>>>>> "docker_storage": {"failed": true, "failures": >>>>>> [["OpenShiftCheckException", >>>>>> "Could not find imported module support code for docker_info. Looked for >>>>>> either AnsibleDockerClient.py or docker_common.py\nTraceback (most recent >>>>>> call last):\n File >>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>> line 225, in run_check\n result = check.run()\n File >>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>>>>> {})\n File >>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>> line 211, in execute_module\n result = >>>>>> self._execute_module(module_name, >>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>> line 809, in _execute_module\n (module_style, shebang, module_data, >>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>> line 203, in _configure_module\n environment=final_environment)\n >>>>>> File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>> '.join(msg))\nAnsibleError: Could not find imported module support code >>>>>> for >>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>> docker_common.py\n"]], "msg": "Could not find imported module support >>>>>> code >>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>> line 225, in run_check\n result = check.run()\n File >>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>>>>> {})\n File >>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>> line 211, in execute_module\n result = >>>>>> self._execute_module(module_name, >>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>> line 809, in _execute_module\n (module_style, shebang, module_data, >>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>> line 203, in _configure_module\n environment=final_environment)\n >>>>>> File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>> '.join(msg))\nAnsibleError: Could not find imported module support code >>>>>> for >>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>> docker_common.py\n"}, "memory_availability": {}, "package_availability": >>>>>> {"changed": false, "invocation": {"module_args": {"packages": ["PyYAML", >>>>>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", >>>>>> "firewalld", >>>>>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables", >>>>>> "iptables-services", "iscsi-initiator-utils", "libselinux-python", >>>>>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients", >>>>>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2", >>>>>> "yum-utils"]}}}, "package_version": {"changed": false, "invocation": >>>>>> {"module_args": {"package_list": [{"check_multi": false, "name": >>>>>> "origin", >>>>>> "version": ""}, {"check_multi": false, "name": "origin-master", >>>>>> "version": >>>>>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}], >>>>>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed", >>>>>> "playbook_context": "install"} >>>>>> >>>>>> NO MORE HOSTS LEFT >>>>>> ******************************************************************************************************************************************* >>>>>> >>>>>> PLAY RECAP >>>>>> *************************************************************************************************************************************************** >>>>>> localhost : ok=35 changed=1 unreachable=0 >>>>>> failed=0 skipped=16 rescued=0 ignored=0 >>>>>> master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 failed=1 >>>>>> skipped=165 rescued=0 ignored=0 >>>>>> >>>>>> >>>>>> INSTALLER STATUS >>>>>> ********************************************************************************************************************************************* >>>>>> Initialization : Complete (0:00:16) >>>>>> Health Check : In Progress (0:00:36) >>>>>> This phase can be restarted by running: >>>>>> playbooks/openshift-checks/pre-install.yml >>>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) 0:01:56.339 >>>>>> ************ >>>>>> >>>>>> =============================================================================== >>>>>> Run health checks (install) - EL >>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>> 36.15s >>>>>> os_firewall : Ensure iptables services are not enabled >>>>>> ------------------------------------------------------------------------------------------------ >>>>>> 2.74s >>>>>> openshift_repos : Ensure libselinux-python is installed >>>>>> ----------------------------------------------------------------------------------------------- >>>>>> 1.77s >>>>>> openshift_repos : refresh cache >>>>>> ----------------------------------------------------------------------------------------------------------------------- >>>>>> 1.60s >>>>>> Gather Cluster facts >>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>> 1.51s >>>>>> container_runtime : Fixup SELinux permissions for docker >>>>>> ---------------------------------------------------------------------------------------------- >>>>>> 1.33s >>>>>> container_runtime : Place additional/blocked/insecure registries in >>>>>> /etc/containers/registries.conf >>>>>> --------------------------------------------------- 1.30s >>>>>> Ensure openshift-ansible installer package deps are installed >>>>>> ----------------------------------------------------------------------------------------- >>>>>> 1.29s >>>>>> container_runtime : Install Docker >>>>>> -------------------------------------------------------------------------------------------------------------------- >>>>>> 1.17s >>>>>> Initialize openshift.node.sdn_mtu >>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>> 1.13s >>>>>> os_firewall : Install firewalld packages >>>>>> -------------------------------------------------------------------------------------------------------------- >>>>>> 1.13s >>>>>> container_runtime : Set various Docker options >>>>>> -------------------------------------------------------------------------------------------------------- >>>>>> 1.11s >>>>>> install NetworkManager >>>>>> -------------------------------------------------------------------------------------------------------------------------------- >>>>>> 1.10s >>>>>> openshift_repos : Configure correct origin release repository >>>>>> ----------------------------------------------------------------------------------------- >>>>>> 1.05s >>>>>> container_runtime : Get current installed Docker version >>>>>> ---------------------------------------------------------------------------------------------- >>>>>> 1.04s >>>>>> openshift_repos : Configure origin gpg keys >>>>>> ----------------------------------------------------------------------------------------------------------- >>>>>> 1.04s >>>>>> openshift_repos : Remove openshift_additional.repo file >>>>>> ----------------------------------------------------------------------------------------------- >>>>>> 0.99s >>>>>> container_runtime : Setup the docker-storage for overlay >>>>>> ---------------------------------------------------------------------------------------------- >>>>>> 0.96s >>>>>> Detecting Operating System from ostree_booted >>>>>> --------------------------------------------------------------------------------------------------------- >>>>>> 0.95s >>>>>> Gather Cluster facts >>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>> 0.92s >>>>>> >>>>>> >>>>>> Failure summary: >>>>>> >>>>>> >>>>>> 1. Hosts: master0.xxxxxxx.com >>>>>> Play: OpenShift Health Checks >>>>>> Task: Run health checks (install) - EL >>>>>> Message: One or more checks failed >>>>>> Details: check "docker_storage": >>>>>> Could not find imported module support code for >>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>> docker_common.py >>>>>> Traceback (most recent call last): >>>>>> File >>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py", >>>>>> line 225, in run_check >>>>>> result = check.run() >>>>>> File >>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py", >>>>>> line 53, in run >>>>>> docker_info = self.execute_module("docker_info", >>>>>> {}) >>>>>> File >>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py", >>>>>> line 211, in execute_module >>>>>> result = self._execute_module(module_name, >>>>>> module_args, self.tmp, self.task_vars) >>>>>> File >>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>> line >>>>>> 809, in _execute_module >>>>>> (module_style, shebang, module_data, module_path) >>>>>> = self._configure_module(module_name=module_name, >>>>>> module_args=module_args, >>>>>> task_vars=task_vars) >>>>>> File >>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>> line >>>>>> 203, in _configure_module >>>>>> environment=final_environment) >>>>>> File >>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>> line >>>>>> 1023, in modify_module >>>>>> environment=environment) >>>>>> File >>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>> line >>>>>> 859, in _find_module_utils >>>>>> recursive_finder(module_name, b_module_data, >>>>>> py_module_names, py_module_cache, zf) >>>>>> File >>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>> line >>>>>> 621, in recursive_finder >>>>>> raise AnsibleError(' '.join(msg)) >>>>>> AnsibleError: Could not find imported module support >>>>>> code for docker_info. Looked for either AnsibleDockerClient.py or >>>>>> docker_common.py >>>>>> >>>>>> >>>>>> The execution of "install_okd.yaml" includes checks designed to fail >>>>>> early if the requirements of the playbook are not met. One or more of >>>>>> these >>>>>> checks failed. To disregard these results,explicitly disable checks by >>>>>> setting an Ansible variable: >>>>>> openshift_disable_check=docker_storage >>>>>> Failing check names are shown in the failure details above. Some >>>>>> checks may be configurable by variables if your requirements are >>>>>> different >>>>>> from the defaults; consult check documentation. >>>>>> Variables can be set in the inventory or passed on the command line >>>>>> using the -e flag to ansible-playbook. >>>>>> _______________________________________________ >>>>>> Users mailing list -- [email protected] >>>>>> To unsubscribe send an email to [email protected] >>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>> oVirt Code of Conduct: >>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>> List Archives: >>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/ >>>>>> >>>>> _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/CN6G76Z7MVNK6AEZQ2I7SHULGXSXHCNQ/ >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/Q6ML5P4FPK6Q7M6ZV6DGSHZKZ3BLQAOR/

