> TASK [openshift_control_plane : Wait for control plane pods to appear]
*********
> Monday 27 May 2019  13:31:54 +0000 (0:00:00.180)       0:14:33.857
************
> FAILED - RETRYING: Wait for control plane pods to appear (60 retries
left).
> FAILED - RETRYING: Wait for control plane pods to appear (59 retries
left).
>    It eventually counts all the way down to zero and fails.

This looks a lot like the issues I saw when the host firewall (iptables)
was blocking another OKD all-in-one-host install script [1].
Disabling iptables allowed the installation to continue for my proof of
concept "cluster".

[1]https://github.com/gshipley/installcentos

The other error I had with [1] was it was trying to install a couple of
packages (zile and python2-pip) from EPEL with the repo disabled.



On Tue, May 28, 2019 at 10:41 AM Jayme <[email protected]> wrote:

> Shirly,
>
> Oh and I should mention that I did verify that NetworkManager was
> installed on the master0 VM and enabled/started the second go around.  So
> that service is there and running.
>
> # systemctl list-unit-files | grep Network
> dbus-org.freedesktop.NetworkManager.service
> enabled
> NetworkManager-dispatcher.service
> enabled
> NetworkManager-wait-online.service
>  enabled
> NetworkManager.service
>  enabled
>
> On Tue, May 28, 2019 at 11:13 AM Jayme <[email protected]> wrote:
>
>> Shirly,
>>
>> I appreciate the help with this.  Unfortunately I am still running in to
>> the same problem.  So far I've tried to install/enable/start NetworkManager
>> on the existing "master0" server and re-ran the plans from the installer
>> VM.  I ran in to the same problem waiting for control plane pods and same
>> errors in syslog.
>>
>> So I wiped everything out, killed the template along with the installer
>> and master VMs.  On oVirt engine (I am running 4.3.3.7-1 stable) I did have
>> ovirt-engine-metrics-1.3.0x rpm installed, no yum updates available on an
>> update check.  So I installed
>> http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm then
>> proceeded to install the latest version of ovirt-engine-metrics which gave
>> me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine.
>>
>> After that package was installed I proceeded to follow steps from
>> beginning outlined at:
>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html
>>  --
>> I ran in to the docker check issue again (same as my initial email) so I
>> disabled that and again got as far as starting control plane pods before
>> failure.
>>
>> Not sure where to go from here at this point.  The only thing I can think
>> of that I did differently vs the instructions outlined above is that I have
>> not crated the wildcard DNS record, however I did set configs to create
>> /etc/hosts entries and they /etc/hosts on the machines have the proper IPs
>> assigned for all hostnames (automatically added by the ansible plays).
>>
>> Any ideas how I can get past the plane pods issue?
>>
>> Thanks!
>>
>> On Tue, May 28, 2019 at 4:23 AM Shirly Radco <[email protected]> wrote:
>>
>>> Hi,
>>>
>>> The latest release of 4.3.z should already include a fix for this issue,
>>> ovirt-engine-metrics-1.3.1 rpm.
>>>
>>> The issue is that it requires the NetworkManagar to be installed,
>>> running and enabled for it to work.
>>>
>>> You can install it manually on the master0 vm , start and enable it or
>>> you can also install the updated rpm from the nightly builds if your
>>> environment is oVirt 4.2.z:
>>>
>>> https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm
>>>
>>> Relevant patches are:
>>> https://gerrit.ovirt.org/#/c/99959/
>>> https://gerrit.ovirt.org/#/c/99718/
>>>
>>> Best regards,
>>>
>>> --
>>>
>>> Shirly Radco
>>>
>>> BI Senior Software Engineer
>>>
>>> Red Hat <https://www.redhat.com/>
>>>
>>> <https://www.redhat.com/>
>>>
>>>
>>> On Mon, May 27, 2019 at 4:41 PM Jayme <[email protected]> wrote:
>>>
>>>> I managed to get past that but am running in to another problem later
>>>> in the process on the control plane pods to appear task.   I thought
>>>> perhaps a glitch in the process from the failed docker step previously so
>>>> after a few more runs I tried killing everything and restarting the metrics
>>>> process again from the very beginning and end up hitting the same issue
>>>> with control plane pods even though all other steps/tasks seem to be
>>>> working.
>>>>
>>>> I'm just getting this:
>>>>
>>>> TASK [openshift_control_plane : Wait for control plane pods to appear]
>>>> *********
>>>> Monday 27 May 2019  13:31:54 +0000 (0:00:00.180)       0:14:33.857
>>>> ************
>>>> FAILED - RETRYING: Wait for control plane pods to appear (60 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (59 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (58 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (57 retries
>>>> left).
>>>> FAILED - RETRYING: Wait for control plane pods to appear (56 retries
>>>> left).
>>>>
>>>> It eventually counts all the way down to zero and fails.
>>>>
>>>> In syslog of the master0 server I'm seeing some errors related to cni
>>>> config
>>>>
>>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None
>>>> kind=pod force=False all_namespaces=None field_selector=None
>>>> namespace=kube-system delete_after=False
>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list
>>>> debug=False selector=None name=master-api-master0.xxxxxx.com
>>>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230   20150
>>>> cni.go:172] Unable to update cni config: No networks found in 
>>>> /etc/cni/net.d
>>>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670   20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready:
>>>> cni config uninitialized
>>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None
>>>> kind=pod force=False all_namespaces=None field_selector=None
>>>> namespace=kube-system delete_after=False
>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list
>>>> debug=False selector=None name=master-api-master0.xxxxxx.com
>>>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911   20150
>>>> cni.go:172] Unable to update cni config: No networks found in 
>>>> /etc/cni/net.d
>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321   20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready:
>>>> cni config uninitialized
>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705   20150
>>>> summary.go:102] Failed to get system container stats for
>>>> "/system.slice/origin-node.service": failed to get cgroup stats for
>>>> "/system.slice/origin-node.service": failed to get container info for
>>>> "/system.slice/origin-node.service": unknown container
>>>> "/system.slice/origin-node.service"
>>>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450   20150
>>>> cni.go:172] Unable to update cni config: No networks found in 
>>>> /etc/cni/net.d
>>>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850   20150
>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false
>>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready:
>>>> cni config uninitialized
>>>>
>>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <[email protected]> wrote:
>>>>
>>>>> Hi Jayme,
>>>>>
>>>>> Thank you for reaching out.
>>>>> Please try rerunning the ansible playbook.
>>>>> If this doesn't work, try adding to the integ.ini in the metrics vm
>>>>> openshift_disable_check=docker_storage
>>>>> and rerun the ansible playbook again.
>>>>>
>>>>> Please update how it goes.
>>>>>
>>>>> Best regards,
>>>>>
>>>>> --
>>>>>
>>>>> Shirly Radco
>>>>>
>>>>> BI Senior Software Engineer
>>>>>
>>>>> Red Hat <https://www.redhat.com/>
>>>>>
>>>>> <https://www.redhat.com/>
>>>>>
>>>>>
>>>>> On Sun, May 26, 2019 at 9:34 PM Jayme <[email protected]> wrote:
>>>>>
>>>>>> I'm running in to this ansible error during oVirt metrics
>>>>>> installation (following procedures at:
>>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html
>>>>>>  )
>>>>>>
>>>>>> This is happening late in the process, after successfully deploying
>>>>>> the installation VM and then running second step from the metrics VM.
>>>>>>
>>>>>> CHECK [memory_availability : master0.xxxxxx.com]
>>>>>> *************************************************************************************************
>>>>>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, "checks":
>>>>>> {"disk_availability": {}, "docker_image_availability": {"changed": true},
>>>>>> "docker_storage": {"failed": true, "failures": 
>>>>>> [["OpenShiftCheckException",
>>>>>> "Could not find imported module support code for docker_info.  Looked for
>>>>>> either AnsibleDockerClient.py or docker_common.py\nTraceback (most recent
>>>>>> call last):\n  File
>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
>>>>>> line 225, in run_check\n    result = check.run()\n  File
>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
>>>>>> line 53, in run\n    docker_info = self.execute_module(\"docker_info\",
>>>>>> {})\n  File
>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
>>>>>> line 211, in execute_module\n    result = 
>>>>>> self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)\n  File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 809, in _execute_module\n    (module_style, shebang, module_data,
>>>>>> module_path) = self._configure_module(module_name=module_name,
>>>>>> module_args=module_args, task_vars=task_vars)\n  File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 203, in _configure_module\n    environment=final_environment)\n  
>>>>>> File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 1023, in modify_module\n    environment=environment)\n  File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 859, in _find_module_utils\n    recursive_finder(module_name,
>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n  File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 621, in recursive_finder\n    raise AnsibleError('
>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support code 
>>>>>> for
>>>>>> docker_info.  Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\n"]], "msg": "Could not find imported module support 
>>>>>> code
>>>>>> for docker_info.  Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\nTraceback (most recent call last):\n  File
>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\",
>>>>>> line 225, in run_check\n    result = check.run()\n  File
>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\",
>>>>>> line 53, in run\n    docker_info = self.execute_module(\"docker_info\",
>>>>>> {})\n  File
>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\",
>>>>>> line 211, in execute_module\n    result = 
>>>>>> self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)\n  File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 809, in _execute_module\n    (module_style, shebang, module_data,
>>>>>> module_path) = self._configure_module(module_name=module_name,
>>>>>> module_args=module_args, task_vars=task_vars)\n  File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\",
>>>>>> line 203, in _configure_module\n    environment=final_environment)\n  
>>>>>> File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 1023, in modify_module\n    environment=environment)\n  File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 859, in _find_module_utils\n    recursive_finder(module_name,
>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n  File
>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\",
>>>>>> line 621, in recursive_finder\n    raise AnsibleError('
>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support code 
>>>>>> for
>>>>>> docker_info.  Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py\n"}, "memory_availability": {}, "package_availability":
>>>>>> {"changed": false, "invocation": {"module_args": {"packages": ["PyYAML",
>>>>>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", 
>>>>>> "firewalld",
>>>>>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables",
>>>>>> "iptables-services", "iscsi-initiator-utils", "libselinux-python",
>>>>>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients",
>>>>>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2",
>>>>>> "yum-utils"]}}}, "package_version": {"changed": false, "invocation":
>>>>>> {"module_args": {"package_list": [{"check_multi": false, "name": 
>>>>>> "origin",
>>>>>> "version": ""}, {"check_multi": false, "name": "origin-master", 
>>>>>> "version":
>>>>>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}],
>>>>>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed",
>>>>>> "playbook_context": "install"}
>>>>>>
>>>>>> NO MORE HOSTS LEFT
>>>>>> *******************************************************************************************************************************************
>>>>>>
>>>>>> PLAY RECAP
>>>>>> ***************************************************************************************************************************************************
>>>>>> localhost                  : ok=35   changed=1    unreachable=0
>>>>>>  failed=0    skipped=16   rescued=0    ignored=0
>>>>>> master0.xxxxxxx.com : ok=96   changed=6    unreachable=0    failed=1
>>>>>>    skipped=165  rescued=0    ignored=0
>>>>>>
>>>>>>
>>>>>> INSTALLER STATUS
>>>>>> *********************************************************************************************************************************************
>>>>>> Initialization  : Complete (0:00:16)
>>>>>> Health Check    : In Progress (0:00:36)
>>>>>> This phase can be restarted by running:
>>>>>> playbooks/openshift-checks/pre-install.yml
>>>>>> Sunday 26 May 2019  16:36:25 +0000 (0:00:36.151)       0:01:56.339
>>>>>> ************
>>>>>>
>>>>>> ===============================================================================
>>>>>> Run health checks (install) - EL
>>>>>> ---------------------------------------------------------------------------------------------------------------------
>>>>>> 36.15s
>>>>>> os_firewall : Ensure iptables services are not enabled
>>>>>> ------------------------------------------------------------------------------------------------
>>>>>> 2.74s
>>>>>> openshift_repos : Ensure libselinux-python is installed
>>>>>> -----------------------------------------------------------------------------------------------
>>>>>> 1.77s
>>>>>> openshift_repos : refresh cache
>>>>>> -----------------------------------------------------------------------------------------------------------------------
>>>>>> 1.60s
>>>>>> Gather Cluster facts
>>>>>> ----------------------------------------------------------------------------------------------------------------------------------
>>>>>> 1.51s
>>>>>> container_runtime : Fixup SELinux permissions for docker
>>>>>> ----------------------------------------------------------------------------------------------
>>>>>> 1.33s
>>>>>> container_runtime : Place additional/blocked/insecure registries in
>>>>>> /etc/containers/registries.conf
>>>>>> --------------------------------------------------- 1.30s
>>>>>> Ensure openshift-ansible installer package deps are installed
>>>>>> -----------------------------------------------------------------------------------------
>>>>>> 1.29s
>>>>>> container_runtime : Install Docker
>>>>>> --------------------------------------------------------------------------------------------------------------------
>>>>>> 1.17s
>>>>>> Initialize openshift.node.sdn_mtu
>>>>>> ---------------------------------------------------------------------------------------------------------------------
>>>>>> 1.13s
>>>>>> os_firewall : Install firewalld packages
>>>>>> --------------------------------------------------------------------------------------------------------------
>>>>>> 1.13s
>>>>>> container_runtime : Set various Docker options
>>>>>> --------------------------------------------------------------------------------------------------------
>>>>>> 1.11s
>>>>>> install NetworkManager
>>>>>> --------------------------------------------------------------------------------------------------------------------------------
>>>>>> 1.10s
>>>>>> openshift_repos : Configure correct origin release repository
>>>>>> -----------------------------------------------------------------------------------------
>>>>>> 1.05s
>>>>>> container_runtime : Get current installed Docker version
>>>>>> ----------------------------------------------------------------------------------------------
>>>>>> 1.04s
>>>>>> openshift_repos : Configure origin gpg keys
>>>>>> -----------------------------------------------------------------------------------------------------------
>>>>>> 1.04s
>>>>>> openshift_repos : Remove openshift_additional.repo file
>>>>>> -----------------------------------------------------------------------------------------------
>>>>>> 0.99s
>>>>>> container_runtime : Setup the docker-storage for overlay
>>>>>> ----------------------------------------------------------------------------------------------
>>>>>> 0.96s
>>>>>> Detecting Operating System from ostree_booted
>>>>>> ---------------------------------------------------------------------------------------------------------
>>>>>> 0.95s
>>>>>> Gather Cluster facts
>>>>>> ----------------------------------------------------------------------------------------------------------------------------------
>>>>>> 0.92s
>>>>>>
>>>>>>
>>>>>> Failure summary:
>>>>>>
>>>>>>
>>>>>>   1. Hosts:    master0.xxxxxxx.com
>>>>>>      Play:     OpenShift Health Checks
>>>>>>      Task:     Run health checks (install) - EL
>>>>>>      Message:  One or more checks failed
>>>>>>      Details:  check "docker_storage":
>>>>>>                Could not find imported module support code for
>>>>>> docker_info.  Looked for either AnsibleDockerClient.py or 
>>>>>> docker_common.py
>>>>>>                Traceback (most recent call last):
>>>>>>                  File
>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py",
>>>>>> line 225, in run_check
>>>>>>                    result = check.run()
>>>>>>                  File
>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py",
>>>>>> line 53, in run
>>>>>>                    docker_info = self.execute_module("docker_info",
>>>>>> {})
>>>>>>                  File
>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py",
>>>>>> line 211, in execute_module
>>>>>>                    result = self._execute_module(module_name,
>>>>>> module_args, self.tmp, self.task_vars)
>>>>>>                  File
>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", 
>>>>>> line
>>>>>> 809, in _execute_module
>>>>>>                    (module_style, shebang, module_data, module_path)
>>>>>> = self._configure_module(module_name=module_name, 
>>>>>> module_args=module_args,
>>>>>> task_vars=task_vars)
>>>>>>                  File
>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", 
>>>>>> line
>>>>>> 203, in _configure_module
>>>>>>                    environment=final_environment)
>>>>>>                  File
>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", 
>>>>>> line
>>>>>> 1023, in modify_module
>>>>>>                    environment=environment)
>>>>>>                  File
>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", 
>>>>>> line
>>>>>> 859, in _find_module_utils
>>>>>>                    recursive_finder(module_name, b_module_data,
>>>>>> py_module_names, py_module_cache, zf)
>>>>>>                  File
>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", 
>>>>>> line
>>>>>> 621, in recursive_finder
>>>>>>                    raise AnsibleError(' '.join(msg))
>>>>>>                AnsibleError: Could not find imported module support
>>>>>> code for docker_info.  Looked for either AnsibleDockerClient.py or
>>>>>> docker_common.py
>>>>>>
>>>>>>
>>>>>> The execution of "install_okd.yaml" includes checks designed to fail
>>>>>> early if the requirements of the playbook are not met. One or more of 
>>>>>> these
>>>>>> checks failed. To disregard these results,explicitly disable checks by
>>>>>> setting an Ansible variable:
>>>>>>    openshift_disable_check=docker_storage
>>>>>> Failing check names are shown in the failure details above. Some
>>>>>> checks may be configurable by variables if your requirements are 
>>>>>> different
>>>>>> from the defaults; consult check documentation.
>>>>>> Variables can be set in the inventory or passed on the command line
>>>>>> using the -e flag to ansible-playbook.
>>>>>> _______________________________________________
>>>>>> Users mailing list -- [email protected]
>>>>>> To unsubscribe send an email to [email protected]
>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>>>>> oVirt Code of Conduct:
>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>>>>>> List Archives:
>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/
>>>>>>
>>>>> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/CN6G76Z7MVNK6AEZQ2I7SHULGXSXHCNQ/
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/Q6ML5P4FPK6Q7M6ZV6DGSHZKZ3BLQAOR/

Reply via email to