Shirly, I appreciate the help with this. Unfortunately I am still running in to the same problem. So far I've tried to install/enable/start NetworkManager on the existing "master0" server and re-ran the plans from the installer VM. I ran in to the same problem waiting for control plane pods and same errors in syslog.
So I wiped everything out, killed the template along with the installer and master VMs. On oVirt engine (I am running 4.3.3.7-1 stable) I did have ovirt-engine-metrics-1.3.0x rpm installed, no yum updates available on an update check. So I installed http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm then proceeded to install the latest version of ovirt-engine-metrics which gave me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine. After that package was installed I proceeded to follow steps from beginning outlined at: https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html -- I ran in to the docker check issue again (same as my initial email) so I disabled that and again got as far as starting control plane pods before failure. Not sure where to go from here at this point. The only thing I can think of that I did differently vs the instructions outlined above is that I have not crated the wildcard DNS record, however I did set configs to create /etc/hosts entries and they /etc/hosts on the machines have the proper IPs assigned for all hostnames (automatically added by the ansible plays). Any ideas how I can get past the plane pods issue? Thanks! On Tue, May 28, 2019 at 4:23 AM Shirly Radco <[email protected]> wrote: > Hi, > > The latest release of 4.3.z should already include a fix for this issue, > ovirt-engine-metrics-1.3.1 rpm. > > The issue is that it requires the NetworkManagar to be installed, running > and enabled for it to work. > > You can install it manually on the master0 vm , start and enable it or you > can also install the updated rpm from the nightly builds if your > environment is oVirt 4.2.z: > > https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm > > Relevant patches are: > https://gerrit.ovirt.org/#/c/99959/ > https://gerrit.ovirt.org/#/c/99718/ > > Best regards, > > -- > > Shirly Radco > > BI Senior Software Engineer > > Red Hat <https://www.redhat.com/> > > <https://www.redhat.com/> > > > On Mon, May 27, 2019 at 4:41 PM Jayme <[email protected]> wrote: > >> I managed to get past that but am running in to another problem later in >> the process on the control plane pods to appear task. I thought perhaps a >> glitch in the process from the failed docker step previously so after a few >> more runs I tried killing everything and restarting the metrics process >> again from the very beginning and end up hitting the same issue with >> control plane pods even though all other steps/tasks seem to be working. >> >> I'm just getting this: >> >> TASK [openshift_control_plane : Wait for control plane pods to appear] >> ********* >> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >> ************ >> FAILED - RETRYING: Wait for control plane pods to appear (60 retries >> left). >> FAILED - RETRYING: Wait for control plane pods to appear (59 retries >> left). >> FAILED - RETRYING: Wait for control plane pods to appear (58 retries >> left). >> FAILED - RETRYING: Wait for control plane pods to appear (57 retries >> left). >> FAILED - RETRYING: Wait for control plane pods to appear (56 retries >> left). >> >> It eventually counts all the way down to zero and fails. >> >> In syslog of the master0 server I'm seeing some errors related to cni >> config >> >> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None kind=pod >> force=False all_namespaces=None field_selector=None namespace=kube-system >> delete_after=False kubeconfig=/etc/origin/master/admin.kubeconfig >> content=None state=list debug=False selector=None name= >> master-api-master0.xxxxxx.com >> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150 >> cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d >> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150 >> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >> cni config uninitialized >> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None kind=pod >> force=False all_namespaces=None field_selector=None namespace=kube-system >> delete_after=False kubeconfig=/etc/origin/master/admin.kubeconfig >> content=None state=list debug=False selector=None name= >> master-api-master0.xxxxxx.com >> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150 >> cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d >> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150 >> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >> cni config uninitialized >> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150 >> summary.go:102] Failed to get system container stats for >> "/system.slice/origin-node.service": failed to get cgroup stats for >> "/system.slice/origin-node.service": failed to get container info for >> "/system.slice/origin-node.service": unknown container >> "/system.slice/origin-node.service" >> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150 >> cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d >> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150 >> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >> cni config uninitialized >> >> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <[email protected]> wrote: >> >>> Hi Jayme, >>> >>> Thank you for reaching out. >>> Please try rerunning the ansible playbook. >>> If this doesn't work, try adding to the integ.ini in the metrics vm >>> openshift_disable_check=docker_storage >>> and rerun the ansible playbook again. >>> >>> Please update how it goes. >>> >>> Best regards, >>> >>> -- >>> >>> Shirly Radco >>> >>> BI Senior Software Engineer >>> >>> Red Hat <https://www.redhat.com/> >>> >>> <https://www.redhat.com/> >>> >>> >>> On Sun, May 26, 2019 at 9:34 PM Jayme <[email protected]> wrote: >>> >>>> I'm running in to this ansible error during oVirt metrics installation >>>> (following procedures at: >>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>> ) >>>> >>>> This is happening late in the process, after successfully deploying the >>>> installation VM and then running second step from the metrics VM. >>>> >>>> CHECK [memory_availability : master0.xxxxxx.com] >>>> ************************************************************************************************* >>>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, "checks": >>>> {"disk_availability": {}, "docker_image_availability": {"changed": true}, >>>> "docker_storage": {"failed": true, "failures": [["OpenShiftCheckException", >>>> "Could not find imported module support code for docker_info. Looked for >>>> either AnsibleDockerClient.py or docker_common.py\nTraceback (most recent >>>> call last):\n File >>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>> line 225, in run_check\n result = check.run()\n File >>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>>> {})\n File >>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>> line 211, in execute_module\n result = self._execute_module(module_name, >>>> module_args, self.tmp, self.task_vars)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>> line 809, in _execute_module\n (module_style, shebang, module_data, >>>> module_path) = self._configure_module(module_name=module_name, >>>> module_args=module_args, task_vars=task_vars)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>> line 203, in _configure_module\n environment=final_environment)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>> line 1023, in modify_module\n environment=environment)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>> line 621, in recursive_finder\n raise AnsibleError(' >>>> '.join(msg))\nAnsibleError: Could not find imported module support code for >>>> docker_info. Looked for either AnsibleDockerClient.py or >>>> docker_common.py\n"]], "msg": "Could not find imported module support code >>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>> docker_common.py\nTraceback (most recent call last):\n File >>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>> line 225, in run_check\n result = check.run()\n File >>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>>> {})\n File >>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>> line 211, in execute_module\n result = self._execute_module(module_name, >>>> module_args, self.tmp, self.task_vars)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>> line 809, in _execute_module\n (module_style, shebang, module_data, >>>> module_path) = self._configure_module(module_name=module_name, >>>> module_args=module_args, task_vars=task_vars)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>> line 203, in _configure_module\n environment=final_environment)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>> line 1023, in modify_module\n environment=environment)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>> line 621, in recursive_finder\n raise AnsibleError(' >>>> '.join(msg))\nAnsibleError: Could not find imported module support code for >>>> docker_info. Looked for either AnsibleDockerClient.py or >>>> docker_common.py\n"}, "memory_availability": {}, "package_availability": >>>> {"changed": false, "invocation": {"module_args": {"packages": ["PyYAML", >>>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", "firewalld", >>>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables", >>>> "iptables-services", "iscsi-initiator-utils", "libselinux-python", >>>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients", >>>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2", >>>> "yum-utils"]}}}, "package_version": {"changed": false, "invocation": >>>> {"module_args": {"package_list": [{"check_multi": false, "name": "origin", >>>> "version": ""}, {"check_multi": false, "name": "origin-master", "version": >>>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}], >>>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed", >>>> "playbook_context": "install"} >>>> >>>> NO MORE HOSTS LEFT >>>> ******************************************************************************************************************************************* >>>> >>>> PLAY RECAP >>>> *************************************************************************************************************************************************** >>>> localhost : ok=35 changed=1 unreachable=0 >>>> failed=0 skipped=16 rescued=0 ignored=0 >>>> master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 failed=1 >>>> skipped=165 rescued=0 ignored=0 >>>> >>>> >>>> INSTALLER STATUS >>>> ********************************************************************************************************************************************* >>>> Initialization : Complete (0:00:16) >>>> Health Check : In Progress (0:00:36) >>>> This phase can be restarted by running: >>>> playbooks/openshift-checks/pre-install.yml >>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) 0:01:56.339 >>>> ************ >>>> >>>> =============================================================================== >>>> Run health checks (install) - EL >>>> --------------------------------------------------------------------------------------------------------------------- >>>> 36.15s >>>> os_firewall : Ensure iptables services are not enabled >>>> ------------------------------------------------------------------------------------------------ >>>> 2.74s >>>> openshift_repos : Ensure libselinux-python is installed >>>> ----------------------------------------------------------------------------------------------- >>>> 1.77s >>>> openshift_repos : refresh cache >>>> ----------------------------------------------------------------------------------------------------------------------- >>>> 1.60s >>>> Gather Cluster facts >>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>> 1.51s >>>> container_runtime : Fixup SELinux permissions for docker >>>> ---------------------------------------------------------------------------------------------- >>>> 1.33s >>>> container_runtime : Place additional/blocked/insecure registries in >>>> /etc/containers/registries.conf >>>> --------------------------------------------------- 1.30s >>>> Ensure openshift-ansible installer package deps are installed >>>> ----------------------------------------------------------------------------------------- >>>> 1.29s >>>> container_runtime : Install Docker >>>> -------------------------------------------------------------------------------------------------------------------- >>>> 1.17s >>>> Initialize openshift.node.sdn_mtu >>>> --------------------------------------------------------------------------------------------------------------------- >>>> 1.13s >>>> os_firewall : Install firewalld packages >>>> -------------------------------------------------------------------------------------------------------------- >>>> 1.13s >>>> container_runtime : Set various Docker options >>>> -------------------------------------------------------------------------------------------------------- >>>> 1.11s >>>> install NetworkManager >>>> -------------------------------------------------------------------------------------------------------------------------------- >>>> 1.10s >>>> openshift_repos : Configure correct origin release repository >>>> ----------------------------------------------------------------------------------------- >>>> 1.05s >>>> container_runtime : Get current installed Docker version >>>> ---------------------------------------------------------------------------------------------- >>>> 1.04s >>>> openshift_repos : Configure origin gpg keys >>>> ----------------------------------------------------------------------------------------------------------- >>>> 1.04s >>>> openshift_repos : Remove openshift_additional.repo file >>>> ----------------------------------------------------------------------------------------------- >>>> 0.99s >>>> container_runtime : Setup the docker-storage for overlay >>>> ---------------------------------------------------------------------------------------------- >>>> 0.96s >>>> Detecting Operating System from ostree_booted >>>> --------------------------------------------------------------------------------------------------------- >>>> 0.95s >>>> Gather Cluster facts >>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>> 0.92s >>>> >>>> >>>> Failure summary: >>>> >>>> >>>> 1. Hosts: master0.xxxxxxx.com >>>> Play: OpenShift Health Checks >>>> Task: Run health checks (install) - EL >>>> Message: One or more checks failed >>>> Details: check "docker_storage": >>>> Could not find imported module support code for >>>> docker_info. Looked for either AnsibleDockerClient.py or docker_common.py >>>> Traceback (most recent call last): >>>> File >>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py", >>>> line 225, in run_check >>>> result = check.run() >>>> File >>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py", >>>> line 53, in run >>>> docker_info = self.execute_module("docker_info", {}) >>>> File >>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py", >>>> line 211, in execute_module >>>> result = self._execute_module(module_name, >>>> module_args, self.tmp, self.task_vars) >>>> File >>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line >>>> 809, in _execute_module >>>> (module_style, shebang, module_data, module_path) = >>>> self._configure_module(module_name=module_name, module_args=module_args, >>>> task_vars=task_vars) >>>> File >>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", line >>>> 203, in _configure_module >>>> environment=final_environment) >>>> File >>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>>> 1023, in modify_module >>>> environment=environment) >>>> File >>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>>> 859, in _find_module_utils >>>> recursive_finder(module_name, b_module_data, >>>> py_module_names, py_module_cache, zf) >>>> File >>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>>> 621, in recursive_finder >>>> raise AnsibleError(' '.join(msg)) >>>> AnsibleError: Could not find imported module support >>>> code for docker_info. Looked for either AnsibleDockerClient.py or >>>> docker_common.py >>>> >>>> >>>> The execution of "install_okd.yaml" includes checks designed to fail >>>> early if the requirements of the playbook are not met. One or more of these >>>> checks failed. To disregard these results,explicitly disable checks by >>>> setting an Ansible variable: >>>> openshift_disable_check=docker_storage >>>> Failing check names are shown in the failure details above. Some checks >>>> may be configurable by variables if your requirements are different from >>>> the defaults; consult check documentation. >>>> Variables can be set in the inventory or passed on the command line >>>> using the -e flag to ansible-playbook. >>>> _______________________________________________ >>>> Users mailing list -- [email protected] >>>> To unsubscribe send an email to [email protected] >>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>> oVirt Code of Conduct: >>>> https://www.ovirt.org/community/about/community-guidelines/ >>>> List Archives: >>>> https://lists.ovirt.org/archives/list/[email protected]/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/ >>>> >>>
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/25KVJORBMANHAP36BDLFSVCLT6EAEDWM/

