Shirly, No problem, I understand. I will provide all of the requested info in a bug report. Thanks again for your help!
On Wed, May 29, 2019 at 11:44 AM Shirly Radco <[email protected]> wrote: > Hi Jayme, > > It getting hard to debug your issue over the mailing list. > Can you please open a bug in bugzilla and attach all the information you > know? > The versions you are using, config files from engine machine > and ansible log, vars.yaml , integ.ini from the master0 vm in /root > directory, . > > Sorry for the inconvenience. > > Best, > > > -- > > Shirly Radco > > BI Senior Software Engineer > > Red Hat <https://www.redhat.com/> > > <https://www.redhat.com/> > > > On Tue, May 28, 2019 at 10:47 PM Jayme <[email protected]> wrote: > >> I actually see the pods running on master0 if I do this: >> >> @master0 master]# oc project kube-system >> Now using project "kube-system" on server " >> https://openshift-master.cloud.xxxxxxx.com:8443". >> [root@master0 master]# oc get pods >> NAME READY >> STATUS RESTARTS AGE >> master-api-master0.cloud.xxxxxxxx.com 1/1 Running 0 >> 22m >> master-controllers-master0.cloud.xxxxxxxx.com 1/1 Running 0 >> 22m >> master-etcd-master0.cloud.xxxxxxxxxx 1/1 Running 0 >> 22m >> >> So I wonder why the ansible "Wait for control plane pods to appear" task >> is looping >> >> - name: Wait for control plane pods to appear >> oc_obj: >> state: list >> kind: pod >> name: "master-{{ item }}-{{ l_kubelet_node_name | lower }}" >> namespace: kube-system >> register: control_plane_pods >> until: >> - "'results' in control_plane_pods" >> - "'results' in control_plane_pods.results" >> - control_plane_pods.results.results | length > 0 >> retries: 60 >> delay: 5 >> with_items: >> - "{{ 'etcd' if inventory_hostname in groups['oo_etcd_to_config'] else >> omit }}" >> - api >> - controllers >> ignore_errors: true >> >> On Tue, May 28, 2019 at 4:23 PM Jayme <[email protected]> wrote: >> >>> I just tried again from scratch this time making sure a proper wildcard >>> DNS entry existed and without using the set /etc/hosts option and am still >>> running in to the pods issue. Can anyone confirm if this requires a public >>> external IP to work? I am working on an internal DNS zone here and natted >>> ips. >>> >>> On Tue, May 28, 2019 at 3:28 PM Edward Berger <[email protected]> >>> wrote: >>> >>>> In my case it was a single bare metal host, so that would be equivalent >>>> to disabling iptables on the master0 VM you're installing to, in your ovirt >>>> scenario. >>>> >>>> On Tue, May 28, 2019 at 1:25 PM Jayme <[email protected]> wrote: >>>> >>>>> Do you mean the iptables firewall on the server being installed to >>>>> i.e. master0 or the actual oVirt host that the master0 VM is running on? >>>>> I >>>>> did try flushing iptables rules on master0 VM then ran plays again from >>>>> installer VM but fail at the same point. >>>>> >>>>> Does this log message have anything to do with the issue, /etc/cni >>>>> directory does not even exist on master0 VM. >>>>> >>>>> May 28 17:23:35 master0 origin-node: W0528 17:23:35.012902 10434 >>>>> cni.go:172] Unable to update cni config: No networks found in >>>>> /etc/cni/net.d >>>>> May 28 17:23:35 master0 origin-node: E0528 17:23:35.013398 10434 >>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>>>> cni config uninitialized >>>>> >>>>> >>>>> >>>>> On Tue, May 28, 2019 at 1:19 PM Edward Berger <[email protected]> >>>>> wrote: >>>>> >>>>>> > TASK [openshift_control_plane : Wait for control plane pods to >>>>>> appear] ********* >>>>>> > Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >>>>>> ************ >>>>>> > FAILED - RETRYING: Wait for control plane pods to appear (60 >>>>>> retries left). >>>>>> > FAILED - RETRYING: Wait for control plane pods to appear (59 >>>>>> retries left). >>>>>> > It eventually counts all the way down to zero and fails. >>>>>> >>>>>> This looks a lot like the issues I saw when the host firewall >>>>>> (iptables) was blocking another OKD all-in-one-host install script [1]. >>>>>> Disabling iptables allowed the installation to continue for my proof >>>>>> of concept "cluster". >>>>>> >>>>>> [1]https://github.com/gshipley/installcentos >>>>>> >>>>>> The other error I had with [1] was it was trying to install a couple >>>>>> of packages (zile and python2-pip) from EPEL with the repo disabled. >>>>>> >>>>>> >>>>>> >>>>>> On Tue, May 28, 2019 at 10:41 AM Jayme <[email protected]> wrote: >>>>>> >>>>>>> Shirly, >>>>>>> >>>>>>> Oh and I should mention that I did verify that NetworkManager was >>>>>>> installed on the master0 VM and enabled/started the second go around. >>>>>>> So >>>>>>> that service is there and running. >>>>>>> >>>>>>> # systemctl list-unit-files | grep Network >>>>>>> dbus-org.freedesktop.NetworkManager.service >>>>>>> enabled >>>>>>> NetworkManager-dispatcher.service >>>>>>> enabled >>>>>>> NetworkManager-wait-online.service >>>>>>> enabled >>>>>>> NetworkManager.service >>>>>>> enabled >>>>>>> >>>>>>> On Tue, May 28, 2019 at 11:13 AM Jayme <[email protected]> wrote: >>>>>>> >>>>>>>> Shirly, >>>>>>>> >>>>>>>> I appreciate the help with this. Unfortunately I am still running >>>>>>>> in to the same problem. So far I've tried to install/enable/start >>>>>>>> NetworkManager on the existing "master0" server and re-ran the plans >>>>>>>> from >>>>>>>> the installer VM. I ran in to the same problem waiting for control >>>>>>>> plane >>>>>>>> pods and same errors in syslog. >>>>>>>> >>>>>>>> So I wiped everything out, killed the template along with the >>>>>>>> installer and master VMs. On oVirt engine (I am running 4.3.3.7-1 >>>>>>>> stable) >>>>>>>> I did have ovirt-engine-metrics-1.3.0x rpm installed, no yum updates >>>>>>>> available on an update check. So I installed >>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm >>>>>>>> then proceeded to install the latest version of ovirt-engine-metrics >>>>>>>> which >>>>>>>> gave me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine. >>>>>>>> >>>>>>>> After that package was installed I proceeded to follow steps from >>>>>>>> beginning outlined at: >>>>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>>>>> -- >>>>>>>> I ran in to the docker check issue again (same as my initial email) so >>>>>>>> I >>>>>>>> disabled that and again got as far as starting control plane pods >>>>>>>> before >>>>>>>> failure. >>>>>>>> >>>>>>>> Not sure where to go from here at this point. The only thing I can >>>>>>>> think of that I did differently vs the instructions outlined above is >>>>>>>> that >>>>>>>> I have not crated the wildcard DNS record, however I did set configs to >>>>>>>> create /etc/hosts entries and they /etc/hosts on the machines have the >>>>>>>> proper IPs assigned for all hostnames (automatically added by the >>>>>>>> ansible >>>>>>>> plays). >>>>>>>> >>>>>>>> Any ideas how I can get past the plane pods issue? >>>>>>>> >>>>>>>> Thanks! >>>>>>>> >>>>>>>> On Tue, May 28, 2019 at 4:23 AM Shirly Radco <[email protected]> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> The latest release of 4.3.z should already include a fix for this >>>>>>>>> issue, ovirt-engine-metrics-1.3.1 rpm. >>>>>>>>> >>>>>>>>> The issue is that it requires the NetworkManagar to be installed, >>>>>>>>> running and enabled for it to work. >>>>>>>>> >>>>>>>>> You can install it manually on the master0 vm , start and enable >>>>>>>>> it or you can also install the updated rpm from the nightly builds if >>>>>>>>> your >>>>>>>>> environment is oVirt 4.2.z: >>>>>>>>> >>>>>>>>> https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm >>>>>>>>> >>>>>>>>> Relevant patches are: >>>>>>>>> https://gerrit.ovirt.org/#/c/99959/ >>>>>>>>> https://gerrit.ovirt.org/#/c/99718/ >>>>>>>>> >>>>>>>>> Best regards, >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> Shirly Radco >>>>>>>>> >>>>>>>>> BI Senior Software Engineer >>>>>>>>> >>>>>>>>> Red Hat <https://www.redhat.com/> >>>>>>>>> >>>>>>>>> <https://www.redhat.com/> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Mon, May 27, 2019 at 4:41 PM Jayme <[email protected]> wrote: >>>>>>>>> >>>>>>>>>> I managed to get past that but am running in to another problem >>>>>>>>>> later in the process on the control plane pods to appear task. I >>>>>>>>>> thought >>>>>>>>>> perhaps a glitch in the process from the failed docker step >>>>>>>>>> previously so >>>>>>>>>> after a few more runs I tried killing everything and restarting the >>>>>>>>>> metrics >>>>>>>>>> process again from the very beginning and end up hitting the same >>>>>>>>>> issue >>>>>>>>>> with control plane pods even though all other steps/tasks seem to be >>>>>>>>>> working. >>>>>>>>>> >>>>>>>>>> I'm just getting this: >>>>>>>>>> >>>>>>>>>> TASK [openshift_control_plane : Wait for control plane pods to >>>>>>>>>> appear] ********* >>>>>>>>>> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) >>>>>>>>>> 0:14:33.857 ************ >>>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (60 >>>>>>>>>> retries left). >>>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (59 >>>>>>>>>> retries left). >>>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (58 >>>>>>>>>> retries left). >>>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (57 >>>>>>>>>> retries left). >>>>>>>>>> FAILED - RETRYING: Wait for control plane pods to appear (56 >>>>>>>>>> retries left). >>>>>>>>>> >>>>>>>>>> It eventually counts all the way down to zero and fails. >>>>>>>>>> >>>>>>>>>> In syslog of the master0 server I'm seeing some errors related to >>>>>>>>>> cni config >>>>>>>>>> >>>>>>>>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None >>>>>>>>>> kind=pod force=False all_namespaces=None field_selector=None >>>>>>>>>> namespace=kube-system delete_after=False >>>>>>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None >>>>>>>>>> state=list >>>>>>>>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>>>>>>>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 >>>>>>>>>> 20150 cni.go:172] Unable to update cni config: No networks found in >>>>>>>>>> /etc/cni/net.d >>>>>>>>>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 >>>>>>>>>> 20150 kubelet.go:2101] Container runtime network not ready: >>>>>>>>>> NetworkReady=false reason:NetworkPluginNotReady message:docker: >>>>>>>>>> network >>>>>>>>>> plugin is not ready: cni config uninitialized >>>>>>>>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None >>>>>>>>>> kind=pod force=False all_namespaces=None field_selector=None >>>>>>>>>> namespace=kube-system delete_after=False >>>>>>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None >>>>>>>>>> state=list >>>>>>>>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>>>>>>>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 >>>>>>>>>> 20150 cni.go:172] Unable to update cni config: No networks found in >>>>>>>>>> /etc/cni/net.d >>>>>>>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 >>>>>>>>>> 20150 kubelet.go:2101] Container runtime network not ready: >>>>>>>>>> NetworkReady=false reason:NetworkPluginNotReady message:docker: >>>>>>>>>> network >>>>>>>>>> plugin is not ready: cni config uninitialized >>>>>>>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 >>>>>>>>>> 20150 summary.go:102] Failed to get system container stats for >>>>>>>>>> "/system.slice/origin-node.service": failed to get cgroup stats for >>>>>>>>>> "/system.slice/origin-node.service": failed to get container info for >>>>>>>>>> "/system.slice/origin-node.service": unknown container >>>>>>>>>> "/system.slice/origin-node.service" >>>>>>>>>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 >>>>>>>>>> 20150 cni.go:172] Unable to update cni config: No networks found in >>>>>>>>>> /etc/cni/net.d >>>>>>>>>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 >>>>>>>>>> 20150 kubelet.go:2101] Container runtime network not ready: >>>>>>>>>> NetworkReady=false reason:NetworkPluginNotReady message:docker: >>>>>>>>>> network >>>>>>>>>> plugin is not ready: cni config uninitialized >>>>>>>>>> >>>>>>>>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <[email protected]> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Jayme, >>>>>>>>>>> >>>>>>>>>>> Thank you for reaching out. >>>>>>>>>>> Please try rerunning the ansible playbook. >>>>>>>>>>> If this doesn't work, try adding to the integ.ini in the metrics >>>>>>>>>>> vm >>>>>>>>>>> openshift_disable_check=docker_storage >>>>>>>>>>> and rerun the ansible playbook again. >>>>>>>>>>> >>>>>>>>>>> Please update how it goes. >>>>>>>>>>> >>>>>>>>>>> Best regards, >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> >>>>>>>>>>> Shirly Radco >>>>>>>>>>> >>>>>>>>>>> BI Senior Software Engineer >>>>>>>>>>> >>>>>>>>>>> Red Hat <https://www.redhat.com/> >>>>>>>>>>> >>>>>>>>>>> <https://www.redhat.com/> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Sun, May 26, 2019 at 9:34 PM Jayme <[email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> I'm running in to this ansible error during oVirt metrics >>>>>>>>>>>> installation (following procedures at: >>>>>>>>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>>>>>>>>> ) >>>>>>>>>>>> >>>>>>>>>>>> This is happening late in the process, after successfully >>>>>>>>>>>> deploying the installation VM and then running second step from >>>>>>>>>>>> the metrics >>>>>>>>>>>> VM. >>>>>>>>>>>> >>>>>>>>>>>> CHECK [memory_availability : master0.xxxxxx.com] >>>>>>>>>>>> ************************************************************************************************* >>>>>>>>>>>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, >>>>>>>>>>>> "checks": {"disk_availability": {}, "docker_image_availability": >>>>>>>>>>>> {"changed": true}, "docker_storage": {"failed": true, "failures": >>>>>>>>>>>> [["OpenShiftCheckException", "Could not find imported module >>>>>>>>>>>> support code >>>>>>>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>>>>>>>> line 225, in run_check\n result = check.run()\n File >>>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>>>>>>>> line 53, in run\n docker_info = >>>>>>>>>>>> self.execute_module(\"docker_info\", >>>>>>>>>>>> {})\n File >>>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>>>>>>>> line 211, in execute_module\n result = >>>>>>>>>>>> self._execute_module(module_name, >>>>>>>>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>>>> line 809, in _execute_module\n (module_style, shebang, >>>>>>>>>>>> module_data, >>>>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>>>> line 203, in _configure_module\n >>>>>>>>>>>> environment=final_environment)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>>>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>>>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support >>>>>>>>>>>> code for >>>>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>>> docker_common.py\n"]], "msg": "Could not find imported module >>>>>>>>>>>> support code >>>>>>>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>>>>>>>> line 225, in run_check\n result = check.run()\n File >>>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>>>>>>>> line 53, in run\n docker_info = >>>>>>>>>>>> self.execute_module(\"docker_info\", >>>>>>>>>>>> {})\n File >>>>>>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>>>>>>>> line 211, in execute_module\n result = >>>>>>>>>>>> self._execute_module(module_name, >>>>>>>>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>>>> line 809, in _execute_module\n (module_style, shebang, >>>>>>>>>>>> module_data, >>>>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>>>>>> line 203, in _configure_module\n >>>>>>>>>>>> environment=final_environment)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>>>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>>>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support >>>>>>>>>>>> code for >>>>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>>> docker_common.py\n"}, "memory_availability": {}, >>>>>>>>>>>> "package_availability": >>>>>>>>>>>> {"changed": false, "invocation": {"module_args": {"packages": >>>>>>>>>>>> ["PyYAML", >>>>>>>>>>>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", >>>>>>>>>>>> "firewalld", >>>>>>>>>>>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables", >>>>>>>>>>>> "iptables-services", "iscsi-initiator-utils", "libselinux-python", >>>>>>>>>>>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients", >>>>>>>>>>>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2", >>>>>>>>>>>> "yum-utils"]}}}, "package_version": {"changed": false, >>>>>>>>>>>> "invocation": >>>>>>>>>>>> {"module_args": {"package_list": [{"check_multi": false, "name": >>>>>>>>>>>> "origin", >>>>>>>>>>>> "version": ""}, {"check_multi": false, "name": "origin-master", >>>>>>>>>>>> "version": >>>>>>>>>>>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}], >>>>>>>>>>>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed", >>>>>>>>>>>> "playbook_context": "install"} >>>>>>>>>>>> >>>>>>>>>>>> NO MORE HOSTS LEFT >>>>>>>>>>>> ******************************************************************************************************************************************* >>>>>>>>>>>> >>>>>>>>>>>> PLAY RECAP >>>>>>>>>>>> *************************************************************************************************************************************************** >>>>>>>>>>>> localhost : ok=35 changed=1 unreachable=0 >>>>>>>>>>>> failed=0 skipped=16 rescued=0 ignored=0 >>>>>>>>>>>> master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 >>>>>>>>>>>> failed=1 skipped=165 rescued=0 ignored=0 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> INSTALLER STATUS >>>>>>>>>>>> ********************************************************************************************************************************************* >>>>>>>>>>>> Initialization : Complete (0:00:16) >>>>>>>>>>>> Health Check : In Progress (0:00:36) >>>>>>>>>>>> This phase can be restarted by running: >>>>>>>>>>>> playbooks/openshift-checks/pre-install.yml >>>>>>>>>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) >>>>>>>>>>>> 0:01:56.339 ************ >>>>>>>>>>>> >>>>>>>>>>>> =============================================================================== >>>>>>>>>>>> Run health checks (install) - EL >>>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 36.15s >>>>>>>>>>>> os_firewall : Ensure iptables services are not enabled >>>>>>>>>>>> ------------------------------------------------------------------------------------------------ >>>>>>>>>>>> 2.74s >>>>>>>>>>>> openshift_repos : Ensure libselinux-python is installed >>>>>>>>>>>> ----------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.77s >>>>>>>>>>>> openshift_repos : refresh cache >>>>>>>>>>>> ----------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.60s >>>>>>>>>>>> Gather Cluster facts >>>>>>>>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.51s >>>>>>>>>>>> container_runtime : Fixup SELinux permissions for docker >>>>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.33s >>>>>>>>>>>> container_runtime : Place additional/blocked/insecure >>>>>>>>>>>> registries in /etc/containers/registries.conf >>>>>>>>>>>> --------------------------------------------------- 1.30s >>>>>>>>>>>> Ensure openshift-ansible installer package deps are installed >>>>>>>>>>>> ----------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.29s >>>>>>>>>>>> container_runtime : Install Docker >>>>>>>>>>>> -------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.17s >>>>>>>>>>>> Initialize openshift.node.sdn_mtu >>>>>>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.13s >>>>>>>>>>>> os_firewall : Install firewalld packages >>>>>>>>>>>> -------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.13s >>>>>>>>>>>> container_runtime : Set various Docker options >>>>>>>>>>>> -------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.11s >>>>>>>>>>>> install NetworkManager >>>>>>>>>>>> -------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.10s >>>>>>>>>>>> openshift_repos : Configure correct origin release repository >>>>>>>>>>>> ----------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.05s >>>>>>>>>>>> container_runtime : Get current installed Docker version >>>>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.04s >>>>>>>>>>>> openshift_repos : Configure origin gpg keys >>>>>>>>>>>> ----------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 1.04s >>>>>>>>>>>> openshift_repos : Remove openshift_additional.repo file >>>>>>>>>>>> ----------------------------------------------------------------------------------------------- >>>>>>>>>>>> 0.99s >>>>>>>>>>>> container_runtime : Setup the docker-storage for overlay >>>>>>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>>>>>> 0.96s >>>>>>>>>>>> Detecting Operating System from ostree_booted >>>>>>>>>>>> --------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 0.95s >>>>>>>>>>>> Gather Cluster facts >>>>>>>>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>>>>>>>> 0.92s >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Failure summary: >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> 1. Hosts: master0.xxxxxxx.com >>>>>>>>>>>> Play: OpenShift Health Checks >>>>>>>>>>>> Task: Run health checks (install) - EL >>>>>>>>>>>> Message: One or more checks failed >>>>>>>>>>>> Details: check "docker_storage": >>>>>>>>>>>> Could not find imported module support code for >>>>>>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>>>>>> docker_common.py >>>>>>>>>>>> Traceback (most recent call last): >>>>>>>>>>>> File >>>>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py", >>>>>>>>>>>> line 225, in run_check >>>>>>>>>>>> result = check.run() >>>>>>>>>>>> File >>>>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py", >>>>>>>>>>>> line 53, in run >>>>>>>>>>>> docker_info = >>>>>>>>>>>> self.execute_module("docker_info", {}) >>>>>>>>>>>> File >>>>>>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py", >>>>>>>>>>>> line 211, in execute_module >>>>>>>>>>>> result = self._execute_module(module_name, >>>>>>>>>>>> module_args, self.tmp, self.task_vars) >>>>>>>>>>>> File >>>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>>>>>>>> line >>>>>>>>>>>> 809, in _execute_module >>>>>>>>>>>> (module_style, shebang, module_data, >>>>>>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>>>>>> module_args=module_args, task_vars=task_vars) >>>>>>>>>>>> File >>>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>>>>>>>> line >>>>>>>>>>>> 203, in _configure_module >>>>>>>>>>>> environment=final_environment) >>>>>>>>>>>> File >>>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>>>> line >>>>>>>>>>>> 1023, in modify_module >>>>>>>>>>>> environment=environment) >>>>>>>>>>>> File >>>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>>>> line >>>>>>>>>>>> 859, in _find_module_utils >>>>>>>>>>>> recursive_finder(module_name, b_module_data, >>>>>>>>>>>> py_module_names, py_module_cache, zf) >>>>>>>>>>>> File >>>>>>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>>>>>> line >>>>>>>>>>>> 621, in recursive_finder >>>>>>>>>>>> raise AnsibleError(' '.join(msg)) >>>>>>>>>>>> AnsibleError: Could not find imported module >>>>>>>>>>>> support code for docker_info. Looked for either >>>>>>>>>>>> AnsibleDockerClient.py or >>>>>>>>>>>> docker_common.py >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> The execution of "install_okd.yaml" includes checks designed to >>>>>>>>>>>> fail early if the requirements of the playbook are not met. One or >>>>>>>>>>>> more of >>>>>>>>>>>> these checks failed. To disregard these results,explicitly disable >>>>>>>>>>>> checks >>>>>>>>>>>> by setting an Ansible variable: >>>>>>>>>>>> openshift_disable_check=docker_storage >>>>>>>>>>>> Failing check names are shown in the failure details above. >>>>>>>>>>>> Some checks may be configurable by variables if your requirements >>>>>>>>>>>> are >>>>>>>>>>>> different from the defaults; consult check documentation. >>>>>>>>>>>> Variables can be set in the inventory or passed on the command >>>>>>>>>>>> line using the -e flag to ansible-playbook. >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing list -- [email protected] >>>>>>>>>>>> To unsubscribe send an email to [email protected] >>>>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>>>> oVirt Code of Conduct: >>>>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>>>>>> List Archives: >>>>>>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/ >>>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>> Users mailing list -- [email protected] >>>>>>> To unsubscribe send an email to [email protected] >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>> oVirt Code of Conduct: >>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>> List Archives: >>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/CN6G76Z7MVNK6AEZQ2I7SHULGXSXHCNQ/ >>>>>>> >>>>>>
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/POSOYKSJADXNYBDMV3VHVGFSUNT4OCEP/

