Hi, The latest release of 4.3.z should already include a fix for this issue, ovirt-engine-metrics-1.3.1 rpm.
The issue is that it requires the NetworkManagar to be installed, running and enabled for it to work. You can install it manually on the master0 vm , start and enable it or you can also install the updated rpm from the nightly builds if your environment is oVirt 4.2.z: https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm Relevant patches are: https://gerrit.ovirt.org/#/c/99959/ https://gerrit.ovirt.org/#/c/99718/ Best regards, -- Shirly Radco BI Senior Software Engineer Red Hat <https://www.redhat.com/> <https://www.redhat.com/> On Mon, May 27, 2019 at 4:41 PM Jayme <jay...@gmail.com> wrote: > I managed to get past that but am running in to another problem later in > the process on the control plane pods to appear task. I thought perhaps a > glitch in the process from the failed docker step previously so after a few > more runs I tried killing everything and restarting the metrics process > again from the very beginning and end up hitting the same issue with > control plane pods even though all other steps/tasks seem to be working. > > I'm just getting this: > > TASK [openshift_control_plane : Wait for control plane pods to appear] > ********* > Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 > ************ > FAILED - RETRYING: Wait for control plane pods to appear (60 retries left). > FAILED - RETRYING: Wait for control plane pods to appear (59 retries left). > FAILED - RETRYING: Wait for control plane pods to appear (58 retries left). > FAILED - RETRYING: Wait for control plane pods to appear (57 retries left). > FAILED - RETRYING: Wait for control plane pods to appear (56 retries left). > > It eventually counts all the way down to zero and fails. > > In syslog of the master0 server I'm seeing some errors related to cni > config > > May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None kind=pod > force=False all_namespaces=None field_selector=None namespace=kube-system > delete_after=False kubeconfig=/etc/origin/master/admin.kubeconfig > content=None state=list debug=False selector=None name= > master-api-master0.xxxxxx.com > May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150 > cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d > May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150 > kubelet.go:2101] Container runtime network not ready: NetworkReady=false > reason:NetworkPluginNotReady message:docker: network plugin is not ready: > cni config uninitialized > May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None kind=pod > force=False all_namespaces=None field_selector=None namespace=kube-system > delete_after=False kubeconfig=/etc/origin/master/admin.kubeconfig > content=None state=list debug=False selector=None name= > master-api-master0.xxxxxx.com > May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150 > cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d > May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150 > kubelet.go:2101] Container runtime network not ready: NetworkReady=false > reason:NetworkPluginNotReady message:docker: network plugin is not ready: > cni config uninitialized > May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150 > summary.go:102] Failed to get system container stats for > "/system.slice/origin-node.service": failed to get cgroup stats for > "/system.slice/origin-node.service": failed to get container info for > "/system.slice/origin-node.service": unknown container > "/system.slice/origin-node.service" > May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150 > cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d > May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150 > kubelet.go:2101] Container runtime network not ready: NetworkReady=false > reason:NetworkPluginNotReady message:docker: network plugin is not ready: > cni config uninitialized > > On Mon, May 27, 2019 at 9:35 AM Shirly Radco <sra...@redhat.com> wrote: > >> Hi Jayme, >> >> Thank you for reaching out. >> Please try rerunning the ansible playbook. >> If this doesn't work, try adding to the integ.ini in the metrics vm >> openshift_disable_check=docker_storage >> and rerun the ansible playbook again. >> >> Please update how it goes. >> >> Best regards, >> >> -- >> >> Shirly Radco >> >> BI Senior Software Engineer >> >> Red Hat <https://www.redhat.com/> >> >> <https://www.redhat.com/> >> >> >> On Sun, May 26, 2019 at 9:34 PM Jayme <jay...@gmail.com> wrote: >> >>> I'm running in to this ansible error during oVirt metrics installation >>> (following procedures at: >>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>> ) >>> >>> This is happening late in the process, after successfully deploying the >>> installation VM and then running second step from the metrics VM. >>> >>> CHECK [memory_availability : master0.xxxxxx.com] >>> ************************************************************************************************* >>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, "checks": >>> {"disk_availability": {}, "docker_image_availability": {"changed": true}, >>> "docker_storage": {"failed": true, "failures": [["OpenShiftCheckException", >>> "Could not find imported module support code for docker_info. Looked for >>> either AnsibleDockerClient.py or docker_common.py\nTraceback (most recent >>> call last):\n File >>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>> line 225, in run_check\n result = check.run()\n File >>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>> {})\n File >>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>> line 211, in execute_module\n result = self._execute_module(module_name, >>> module_args, self.tmp, self.task_vars)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>> line 809, in _execute_module\n (module_style, shebang, module_data, >>> module_path) = self._configure_module(module_name=module_name, >>> module_args=module_args, task_vars=task_vars)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>> line 203, in _configure_module\n environment=final_environment)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>> line 1023, in modify_module\n environment=environment)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>> line 859, in _find_module_utils\n recursive_finder(module_name, >>> b_module_data, py_module_names, py_module_cache, zf)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>> line 621, in recursive_finder\n raise AnsibleError(' >>> '.join(msg))\nAnsibleError: Could not find imported module support code for >>> docker_info. Looked for either AnsibleDockerClient.py or >>> docker_common.py\n"]], "msg": "Could not find imported module support code >>> for docker_info. Looked for either AnsibleDockerClient.py or >>> docker_common.py\nTraceback (most recent call last):\n File >>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>> line 225, in run_check\n result = check.run()\n File >>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>> {})\n File >>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>> line 211, in execute_module\n result = self._execute_module(module_name, >>> module_args, self.tmp, self.task_vars)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>> line 809, in _execute_module\n (module_style, shebang, module_data, >>> module_path) = self._configure_module(module_name=module_name, >>> module_args=module_args, task_vars=task_vars)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>> line 203, in _configure_module\n environment=final_environment)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>> line 1023, in modify_module\n environment=environment)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>> line 859, in _find_module_utils\n recursive_finder(module_name, >>> b_module_data, py_module_names, py_module_cache, zf)\n File >>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>> line 621, in recursive_finder\n raise AnsibleError(' >>> '.join(msg))\nAnsibleError: Could not find imported module support code for >>> docker_info. Looked for either AnsibleDockerClient.py or >>> docker_common.py\n"}, "memory_availability": {}, "package_availability": >>> {"changed": false, "invocation": {"module_args": {"packages": ["PyYAML", >>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", "firewalld", >>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables", >>> "iptables-services", "iscsi-initiator-utils", "libselinux-python", >>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients", >>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2", >>> "yum-utils"]}}}, "package_version": {"changed": false, "invocation": >>> {"module_args": {"package_list": [{"check_multi": false, "name": "origin", >>> "version": ""}, {"check_multi": false, "name": "origin-master", "version": >>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}], >>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed", >>> "playbook_context": "install"} >>> >>> NO MORE HOSTS LEFT >>> ******************************************************************************************************************************************* >>> >>> PLAY RECAP >>> *************************************************************************************************************************************************** >>> localhost : ok=35 changed=1 unreachable=0 >>> failed=0 skipped=16 rescued=0 ignored=0 >>> master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 failed=1 >>> skipped=165 rescued=0 ignored=0 >>> >>> >>> INSTALLER STATUS >>> ********************************************************************************************************************************************* >>> Initialization : Complete (0:00:16) >>> Health Check : In Progress (0:00:36) >>> This phase can be restarted by running: >>> playbooks/openshift-checks/pre-install.yml >>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) 0:01:56.339 >>> ************ >>> >>> =============================================================================== >>> Run health checks (install) - EL >>> --------------------------------------------------------------------------------------------------------------------- >>> 36.15s >>> os_firewall : Ensure iptables services are not enabled >>> ------------------------------------------------------------------------------------------------ >>> 2.74s >>> openshift_repos : Ensure libselinux-python is installed >>> ----------------------------------------------------------------------------------------------- >>> 1.77s >>> openshift_repos : refresh cache >>> ----------------------------------------------------------------------------------------------------------------------- >>> 1.60s >>> Gather Cluster facts >>> ---------------------------------------------------------------------------------------------------------------------------------- >>> 1.51s >>> container_runtime : Fixup SELinux permissions for docker >>> ---------------------------------------------------------------------------------------------- >>> 1.33s >>> container_runtime : Place additional/blocked/insecure registries in >>> /etc/containers/registries.conf >>> --------------------------------------------------- 1.30s >>> Ensure openshift-ansible installer package deps are installed >>> ----------------------------------------------------------------------------------------- >>> 1.29s >>> container_runtime : Install Docker >>> -------------------------------------------------------------------------------------------------------------------- >>> 1.17s >>> Initialize openshift.node.sdn_mtu >>> --------------------------------------------------------------------------------------------------------------------- >>> 1.13s >>> os_firewall : Install firewalld packages >>> -------------------------------------------------------------------------------------------------------------- >>> 1.13s >>> container_runtime : Set various Docker options >>> -------------------------------------------------------------------------------------------------------- >>> 1.11s >>> install NetworkManager >>> -------------------------------------------------------------------------------------------------------------------------------- >>> 1.10s >>> openshift_repos : Configure correct origin release repository >>> ----------------------------------------------------------------------------------------- >>> 1.05s >>> container_runtime : Get current installed Docker version >>> ---------------------------------------------------------------------------------------------- >>> 1.04s >>> openshift_repos : Configure origin gpg keys >>> ----------------------------------------------------------------------------------------------------------- >>> 1.04s >>> openshift_repos : Remove openshift_additional.repo file >>> ----------------------------------------------------------------------------------------------- >>> 0.99s >>> container_runtime : Setup the docker-storage for overlay >>> ---------------------------------------------------------------------------------------------- >>> 0.96s >>> Detecting Operating System from ostree_booted >>> --------------------------------------------------------------------------------------------------------- >>> 0.95s >>> Gather Cluster facts >>> ---------------------------------------------------------------------------------------------------------------------------------- >>> 0.92s >>> >>> >>> Failure summary: >>> >>> >>> 1. Hosts: master0.xxxxxxx.com >>> Play: OpenShift Health Checks >>> Task: Run health checks (install) - EL >>> Message: One or more checks failed >>> Details: check "docker_storage": >>> Could not find imported module support code for >>> docker_info. Looked for either AnsibleDockerClient.py or docker_common.py >>> Traceback (most recent call last): >>> File >>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py", >>> line 225, in run_check >>> result = check.run() >>> File >>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py", >>> line 53, in run >>> docker_info = self.execute_module("docker_info", {}) >>> File >>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py", >>> line 211, in execute_module >>> result = self._execute_module(module_name, >>> module_args, self.tmp, self.task_vars) >>> File >>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line >>> 809, in _execute_module >>> (module_style, shebang, module_data, module_path) = >>> self._configure_module(module_name=module_name, module_args=module_args, >>> task_vars=task_vars) >>> File >>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line >>> 203, in _configure_module >>> environment=final_environment) >>> File >>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>> 1023, in modify_module >>> environment=environment) >>> File >>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>> 859, in _find_module_utils >>> recursive_finder(module_name, b_module_data, >>> py_module_names, py_module_cache, zf) >>> File >>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>> 621, in recursive_finder >>> raise AnsibleError(' '.join(msg)) >>> AnsibleError: Could not find imported module support code >>> for docker_info. Looked for either AnsibleDockerClient.py or >>> docker_common.py >>> >>> >>> The execution of "install_okd.yaml" includes checks designed to fail >>> early if the requirements of the playbook are not met. One or more of these >>> checks failed. To disregard these results,explicitly disable checks by >>> setting an Ansible variable: >>> openshift_disable_check=docker_storage >>> Failing check names are shown in the failure details above. Some checks >>> may be configurable by variables if your requirements are different from >>> the defaults; consult check documentation. >>> Variables can be set in the inventory or passed on the command line >>> using the -e flag to ansible-playbook. >>> _______________________________________________ >>> Users mailing list -- users@ovirt.org >>> To unsubscribe send an email to users-le...@ovirt.org >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>> oVirt Code of Conduct: >>> https://www.ovirt.org/community/about/community-guidelines/ >>> List Archives: >>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/ >>> >>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/5MPGNLV4BRZSPI2LP4LVF2KEGJPRWCJZ/