Shirly, Oh and I should mention that I did verify that NetworkManager was installed on the master0 VM and enabled/started the second go around. So that service is there and running.
# systemctl list-unit-files | grep Network dbus-org.freedesktop.NetworkManager.service enabled NetworkManager-dispatcher.service enabled NetworkManager-wait-online.service enabled NetworkManager.service enabled On Tue, May 28, 2019 at 11:13 AM Jayme <jay...@gmail.com> wrote: > Shirly, > > I appreciate the help with this. Unfortunately I am still running in to > the same problem. So far I've tried to install/enable/start NetworkManager > on the existing "master0" server and re-ran the plans from the installer > VM. I ran in to the same problem waiting for control plane pods and same > errors in syslog. > > So I wiped everything out, killed the template along with the installer > and master VMs. On oVirt engine (I am running 4.3.3.7-1 stable) I did have > ovirt-engine-metrics-1.3.0x rpm installed, no yum updates available on an > update check. So I installed > http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm then > proceeded to install the latest version of ovirt-engine-metrics which gave > me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine. > > After that package was installed I proceeded to follow steps from > beginning outlined at: > https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html > -- > I ran in to the docker check issue again (same as my initial email) so I > disabled that and again got as far as starting control plane pods before > failure. > > Not sure where to go from here at this point. The only thing I can think > of that I did differently vs the instructions outlined above is that I have > not crated the wildcard DNS record, however I did set configs to create > /etc/hosts entries and they /etc/hosts on the machines have the proper IPs > assigned for all hostnames (automatically added by the ansible plays). > > Any ideas how I can get past the plane pods issue? > > Thanks! > > On Tue, May 28, 2019 at 4:23 AM Shirly Radco <sra...@redhat.com> wrote: > >> Hi, >> >> The latest release of 4.3.z should already include a fix for this issue, >> ovirt-engine-metrics-1.3.1 rpm. >> >> The issue is that it requires the NetworkManagar to be installed, running >> and enabled for it to work. >> >> You can install it manually on the master0 vm , start and enable it or >> you can also install the updated rpm from the nightly builds if your >> environment is oVirt 4.2.z: >> >> https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm >> >> Relevant patches are: >> https://gerrit.ovirt.org/#/c/99959/ >> https://gerrit.ovirt.org/#/c/99718/ >> >> Best regards, >> >> -- >> >> Shirly Radco >> >> BI Senior Software Engineer >> >> Red Hat <https://www.redhat.com/> >> >> <https://www.redhat.com/> >> >> >> On Mon, May 27, 2019 at 4:41 PM Jayme <jay...@gmail.com> wrote: >> >>> I managed to get past that but am running in to another problem later in >>> the process on the control plane pods to appear task. I thought perhaps a >>> glitch in the process from the failed docker step previously so after a few >>> more runs I tried killing everything and restarting the metrics process >>> again from the very beginning and end up hitting the same issue with >>> control plane pods even though all other steps/tasks seem to be working. >>> >>> I'm just getting this: >>> >>> TASK [openshift_control_plane : Wait for control plane pods to appear] >>> ********* >>> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >>> ************ >>> FAILED - RETRYING: Wait for control plane pods to appear (60 retries >>> left). >>> FAILED - RETRYING: Wait for control plane pods to appear (59 retries >>> left). >>> FAILED - RETRYING: Wait for control plane pods to appear (58 retries >>> left). >>> FAILED - RETRYING: Wait for control plane pods to appear (57 retries >>> left). >>> FAILED - RETRYING: Wait for control plane pods to appear (56 retries >>> left). >>> >>> It eventually counts all the way down to zero and fails. >>> >>> In syslog of the master0 server I'm seeing some errors related to cni >>> config >>> >>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None kind=pod >>> force=False all_namespaces=None field_selector=None namespace=kube-system >>> delete_after=False kubeconfig=/etc/origin/master/admin.kubeconfig >>> content=None state=list debug=False selector=None name= >>> master-api-master0.xxxxxx.com >>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150 >>> cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d >>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150 >>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>> cni config uninitialized >>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None kind=pod >>> force=False all_namespaces=None field_selector=None namespace=kube-system >>> delete_after=False kubeconfig=/etc/origin/master/admin.kubeconfig >>> content=None state=list debug=False selector=None name= >>> master-api-master0.xxxxxx.com >>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150 >>> cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d >>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150 >>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>> cni config uninitialized >>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150 >>> summary.go:102] Failed to get system container stats for >>> "/system.slice/origin-node.service": failed to get cgroup stats for >>> "/system.slice/origin-node.service": failed to get container info for >>> "/system.slice/origin-node.service": unknown container >>> "/system.slice/origin-node.service" >>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150 >>> cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d >>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150 >>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>> cni config uninitialized >>> >>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <sra...@redhat.com> wrote: >>> >>>> Hi Jayme, >>>> >>>> Thank you for reaching out. >>>> Please try rerunning the ansible playbook. >>>> If this doesn't work, try adding to the integ.ini in the metrics vm >>>> openshift_disable_check=docker_storage >>>> and rerun the ansible playbook again. >>>> >>>> Please update how it goes. >>>> >>>> Best regards, >>>> >>>> -- >>>> >>>> Shirly Radco >>>> >>>> BI Senior Software Engineer >>>> >>>> Red Hat <https://www.redhat.com/> >>>> >>>> <https://www.redhat.com/> >>>> >>>> >>>> On Sun, May 26, 2019 at 9:34 PM Jayme <jay...@gmail.com> wrote: >>>> >>>>> I'm running in to this ansible error during oVirt metrics installation >>>>> (following procedures at: >>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>> ) >>>>> >>>>> This is happening late in the process, after successfully deploying >>>>> the installation VM and then running second step from the metrics VM. >>>>> >>>>> CHECK [memory_availability : master0.xxxxxx.com] >>>>> ************************************************************************************************* >>>>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, "checks": >>>>> {"disk_availability": {}, "docker_image_availability": {"changed": true}, >>>>> "docker_storage": {"failed": true, "failures": >>>>> [["OpenShiftCheckException", >>>>> "Could not find imported module support code for docker_info. Looked for >>>>> either AnsibleDockerClient.py or docker_common.py\nTraceback (most recent >>>>> call last):\n File >>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>> line 225, in run_check\n result = check.run()\n File >>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>>>> {})\n File >>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>> line 211, in execute_module\n result = >>>>> self._execute_module(module_name, >>>>> module_args, self.tmp, self.task_vars)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>> line 809, in _execute_module\n (module_style, shebang, module_data, >>>>> module_path) = self._configure_module(module_name=module_name, >>>>> module_args=module_args, task_vars=task_vars)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>> line 203, in _configure_module\n environment=final_environment)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>> line 1023, in modify_module\n environment=environment)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>> '.join(msg))\nAnsibleError: Could not find imported module support code >>>>> for >>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>> docker_common.py\n"]], "msg": "Could not find imported module support code >>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>> line 225, in run_check\n result = check.run()\n File >>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>>>> {})\n File >>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>> line 211, in execute_module\n result = >>>>> self._execute_module(module_name, >>>>> module_args, self.tmp, self.task_vars)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>> line 809, in _execute_module\n (module_style, shebang, module_data, >>>>> module_path) = self._configure_module(module_name=module_name, >>>>> module_args=module_args, task_vars=task_vars)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>> line 203, in _configure_module\n environment=final_environment)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>> line 1023, in modify_module\n environment=environment)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>> '.join(msg))\nAnsibleError: Could not find imported module support code >>>>> for >>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>> docker_common.py\n"}, "memory_availability": {}, "package_availability": >>>>> {"changed": false, "invocation": {"module_args": {"packages": ["PyYAML", >>>>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", >>>>> "firewalld", >>>>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables", >>>>> "iptables-services", "iscsi-initiator-utils", "libselinux-python", >>>>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients", >>>>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2", >>>>> "yum-utils"]}}}, "package_version": {"changed": false, "invocation": >>>>> {"module_args": {"package_list": [{"check_multi": false, "name": "origin", >>>>> "version": ""}, {"check_multi": false, "name": "origin-master", "version": >>>>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}], >>>>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed", >>>>> "playbook_context": "install"} >>>>> >>>>> NO MORE HOSTS LEFT >>>>> ******************************************************************************************************************************************* >>>>> >>>>> PLAY RECAP >>>>> *************************************************************************************************************************************************** >>>>> localhost : ok=35 changed=1 unreachable=0 >>>>> failed=0 skipped=16 rescued=0 ignored=0 >>>>> master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 failed=1 >>>>> skipped=165 rescued=0 ignored=0 >>>>> >>>>> >>>>> INSTALLER STATUS >>>>> ********************************************************************************************************************************************* >>>>> Initialization : Complete (0:00:16) >>>>> Health Check : In Progress (0:00:36) >>>>> This phase can be restarted by running: >>>>> playbooks/openshift-checks/pre-install.yml >>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) 0:01:56.339 >>>>> ************ >>>>> >>>>> =============================================================================== >>>>> Run health checks (install) - EL >>>>> --------------------------------------------------------------------------------------------------------------------- >>>>> 36.15s >>>>> os_firewall : Ensure iptables services are not enabled >>>>> ------------------------------------------------------------------------------------------------ >>>>> 2.74s >>>>> openshift_repos : Ensure libselinux-python is installed >>>>> ----------------------------------------------------------------------------------------------- >>>>> 1.77s >>>>> openshift_repos : refresh cache >>>>> ----------------------------------------------------------------------------------------------------------------------- >>>>> 1.60s >>>>> Gather Cluster facts >>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>> 1.51s >>>>> container_runtime : Fixup SELinux permissions for docker >>>>> ---------------------------------------------------------------------------------------------- >>>>> 1.33s >>>>> container_runtime : Place additional/blocked/insecure registries in >>>>> /etc/containers/registries.conf >>>>> --------------------------------------------------- 1.30s >>>>> Ensure openshift-ansible installer package deps are installed >>>>> ----------------------------------------------------------------------------------------- >>>>> 1.29s >>>>> container_runtime : Install Docker >>>>> -------------------------------------------------------------------------------------------------------------------- >>>>> 1.17s >>>>> Initialize openshift.node.sdn_mtu >>>>> --------------------------------------------------------------------------------------------------------------------- >>>>> 1.13s >>>>> os_firewall : Install firewalld packages >>>>> -------------------------------------------------------------------------------------------------------------- >>>>> 1.13s >>>>> container_runtime : Set various Docker options >>>>> -------------------------------------------------------------------------------------------------------- >>>>> 1.11s >>>>> install NetworkManager >>>>> -------------------------------------------------------------------------------------------------------------------------------- >>>>> 1.10s >>>>> openshift_repos : Configure correct origin release repository >>>>> ----------------------------------------------------------------------------------------- >>>>> 1.05s >>>>> container_runtime : Get current installed Docker version >>>>> ---------------------------------------------------------------------------------------------- >>>>> 1.04s >>>>> openshift_repos : Configure origin gpg keys >>>>> ----------------------------------------------------------------------------------------------------------- >>>>> 1.04s >>>>> openshift_repos : Remove openshift_additional.repo file >>>>> ----------------------------------------------------------------------------------------------- >>>>> 0.99s >>>>> container_runtime : Setup the docker-storage for overlay >>>>> ---------------------------------------------------------------------------------------------- >>>>> 0.96s >>>>> Detecting Operating System from ostree_booted >>>>> --------------------------------------------------------------------------------------------------------- >>>>> 0.95s >>>>> Gather Cluster facts >>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>> 0.92s >>>>> >>>>> >>>>> Failure summary: >>>>> >>>>> >>>>> 1. Hosts: master0.xxxxxxx.com >>>>> Play: OpenShift Health Checks >>>>> Task: Run health checks (install) - EL >>>>> Message: One or more checks failed >>>>> Details: check "docker_storage": >>>>> Could not find imported module support code for >>>>> docker_info. Looked for either AnsibleDockerClient.py or docker_common.py >>>>> Traceback (most recent call last): >>>>> File >>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py", >>>>> line 225, in run_check >>>>> result = check.run() >>>>> File >>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py", >>>>> line 53, in run >>>>> docker_info = self.execute_module("docker_info", {}) >>>>> File >>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py", >>>>> line 211, in execute_module >>>>> result = self._execute_module(module_name, >>>>> module_args, self.tmp, self.task_vars) >>>>> File >>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>> line >>>>> 809, in _execute_module >>>>> (module_style, shebang, module_data, module_path) = >>>>> self._configure_module(module_name=module_name, module_args=module_args, >>>>> task_vars=task_vars) >>>>> File >>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>> line >>>>> 203, in _configure_module >>>>> environment=final_environment) >>>>> File >>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>>>> 1023, in modify_module >>>>> environment=environment) >>>>> File >>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>>>> 859, in _find_module_utils >>>>> recursive_finder(module_name, b_module_data, >>>>> py_module_names, py_module_cache, zf) >>>>> File >>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", line >>>>> 621, in recursive_finder >>>>> raise AnsibleError(' '.join(msg)) >>>>> AnsibleError: Could not find imported module support >>>>> code for docker_info. Looked for either AnsibleDockerClient.py or >>>>> docker_common.py >>>>> >>>>> >>>>> The execution of "install_okd.yaml" includes checks designed to fail >>>>> early if the requirements of the playbook are not met. One or more of >>>>> these >>>>> checks failed. To disregard these results,explicitly disable checks by >>>>> setting an Ansible variable: >>>>> openshift_disable_check=docker_storage >>>>> Failing check names are shown in the failure details above. Some >>>>> checks may be configurable by variables if your requirements are different >>>>> from the defaults; consult check documentation. >>>>> Variables can be set in the inventory or passed on the command line >>>>> using the -e flag to ansible-playbook. >>>>> _______________________________________________ >>>>> Users mailing list -- users@ovirt.org >>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>> oVirt Code of Conduct: >>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>> List Archives: >>>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/ >>>>> >>>>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/CN6G76Z7MVNK6AEZQ2I7SHULGXSXHCNQ/