I should also mention one more thing, I am attempting to install on an internal domain, not externally accessible.
On Tue, May 28, 2019 at 2:25 PM Jayme <[email protected]> wrote: > Do you mean the iptables firewall on the server being installed to i.e. > master0 or the actual oVirt host that the master0 VM is running on? I did > try flushing iptables rules on master0 VM then ran plays again from > installer VM but fail at the same point. > > Does this log message have anything to do with the issue, /etc/cni > directory does not even exist on master0 VM. > > May 28 17:23:35 master0 origin-node: W0528 17:23:35.012902 10434 > cni.go:172] Unable to update cni config: No networks found in /etc/cni/net.d > May 28 17:23:35 master0 origin-node: E0528 17:23:35.013398 10434 > kubelet.go:2101] Container runtime network not ready: NetworkReady=false > reason:NetworkPluginNotReady message:docker: network plugin is not ready: > cni config uninitialized > > > > On Tue, May 28, 2019 at 1:19 PM Edward Berger <[email protected]> wrote: > >> > TASK [openshift_control_plane : Wait for control plane pods to appear] >> ********* >> > Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >> ************ >> > FAILED - RETRYING: Wait for control plane pods to appear (60 retries >> left). >> > FAILED - RETRYING: Wait for control plane pods to appear (59 retries >> left). >> > It eventually counts all the way down to zero and fails. >> >> This looks a lot like the issues I saw when the host firewall (iptables) >> was blocking another OKD all-in-one-host install script [1]. >> Disabling iptables allowed the installation to continue for my proof of >> concept "cluster". >> >> [1]https://github.com/gshipley/installcentos >> >> The other error I had with [1] was it was trying to install a couple of >> packages (zile and python2-pip) from EPEL with the repo disabled. >> >> >> >> On Tue, May 28, 2019 at 10:41 AM Jayme <[email protected]> wrote: >> >>> Shirly, >>> >>> Oh and I should mention that I did verify that NetworkManager was >>> installed on the master0 VM and enabled/started the second go around. So >>> that service is there and running. >>> >>> # systemctl list-unit-files | grep Network >>> dbus-org.freedesktop.NetworkManager.service >>> enabled >>> NetworkManager-dispatcher.service >>> enabled >>> NetworkManager-wait-online.service >>> enabled >>> NetworkManager.service >>> enabled >>> >>> On Tue, May 28, 2019 at 11:13 AM Jayme <[email protected]> wrote: >>> >>>> Shirly, >>>> >>>> I appreciate the help with this. Unfortunately I am still running in >>>> to the same problem. So far I've tried to install/enable/start >>>> NetworkManager on the existing "master0" server and re-ran the plans from >>>> the installer VM. I ran in to the same problem waiting for control plane >>>> pods and same errors in syslog. >>>> >>>> So I wiped everything out, killed the template along with the installer >>>> and master VMs. On oVirt engine (I am running 4.3.3.7-1 stable) I did have >>>> ovirt-engine-metrics-1.3.0x rpm installed, no yum updates available on an >>>> update check. So I installed >>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release43-pre.rpm then >>>> proceeded to install the latest version of ovirt-engine-metrics which gave >>>> me: ovirt-engine-metrics-1.3.1-1.el7.noarch on hosted engine. >>>> >>>> After that package was installed I proceeded to follow steps from >>>> beginning outlined at: >>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>> -- >>>> I ran in to the docker check issue again (same as my initial email) so I >>>> disabled that and again got as far as starting control plane pods before >>>> failure. >>>> >>>> Not sure where to go from here at this point. The only thing I can >>>> think of that I did differently vs the instructions outlined above is that >>>> I have not crated the wildcard DNS record, however I did set configs to >>>> create /etc/hosts entries and they /etc/hosts on the machines have the >>>> proper IPs assigned for all hostnames (automatically added by the ansible >>>> plays). >>>> >>>> Any ideas how I can get past the plane pods issue? >>>> >>>> Thanks! >>>> >>>> On Tue, May 28, 2019 at 4:23 AM Shirly Radco <[email protected]> wrote: >>>> >>>>> Hi, >>>>> >>>>> The latest release of 4.3.z should already include a fix for this >>>>> issue, ovirt-engine-metrics-1.3.1 rpm. >>>>> >>>>> The issue is that it requires the NetworkManagar to be installed, >>>>> running and enabled for it to work. >>>>> >>>>> You can install it manually on the master0 vm , start and enable it or >>>>> you can also install the updated rpm from the nightly builds if your >>>>> environment is oVirt 4.2.z: >>>>> >>>>> https://resources.ovirt.org/pub/ovirt-4.2-snapshot/rpm/el7/noarch/ovirt-engine-metrics-1.2.3-0.0.master.20190523112218.gitbc6e4fa.el7.noarch.rpm >>>>> >>>>> Relevant patches are: >>>>> https://gerrit.ovirt.org/#/c/99959/ >>>>> https://gerrit.ovirt.org/#/c/99718/ >>>>> >>>>> Best regards, >>>>> >>>>> -- >>>>> >>>>> Shirly Radco >>>>> >>>>> BI Senior Software Engineer >>>>> >>>>> Red Hat <https://www.redhat.com/> >>>>> >>>>> <https://www.redhat.com/> >>>>> >>>>> >>>>> On Mon, May 27, 2019 at 4:41 PM Jayme <[email protected]> wrote: >>>>> >>>>>> I managed to get past that but am running in to another problem later >>>>>> in the process on the control plane pods to appear task. I thought >>>>>> perhaps a glitch in the process from the failed docker step previously so >>>>>> after a few more runs I tried killing everything and restarting the >>>>>> metrics >>>>>> process again from the very beginning and end up hitting the same issue >>>>>> with control plane pods even though all other steps/tasks seem to be >>>>>> working. >>>>>> >>>>>> I'm just getting this: >>>>>> >>>>>> TASK [openshift_control_plane : Wait for control plane pods to >>>>>> appear] ********* >>>>>> Monday 27 May 2019 13:31:54 +0000 (0:00:00.180) 0:14:33.857 >>>>>> ************ >>>>>> FAILED - RETRYING: Wait for control plane pods to appear (60 retries >>>>>> left). >>>>>> FAILED - RETRYING: Wait for control plane pods to appear (59 retries >>>>>> left). >>>>>> FAILED - RETRYING: Wait for control plane pods to appear (58 retries >>>>>> left). >>>>>> FAILED - RETRYING: Wait for control plane pods to appear (57 retries >>>>>> left). >>>>>> FAILED - RETRYING: Wait for control plane pods to appear (56 retries >>>>>> left). >>>>>> >>>>>> It eventually counts all the way down to zero and fails. >>>>>> >>>>>> In syslog of the master0 server I'm seeing some errors related to cni >>>>>> config >>>>>> >>>>>> May 27 13:39:07 master0 ansible-oc_obj: Invoked with files=None >>>>>> kind=pod force=False all_namespaces=None field_selector=None >>>>>> namespace=kube-system delete_after=False >>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list >>>>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>>>> May 27 13:39:09 master0 origin-node: W0527 13:39:09.064230 20150 >>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>> /etc/cni/net.d >>>>>> May 27 13:39:09 master0 origin-node: E0527 13:39:09.064670 20150 >>>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>>>>> cni config uninitialized >>>>>> May 27 13:39:13 master0 ansible-oc_obj: Invoked with files=None >>>>>> kind=pod force=False all_namespaces=None field_selector=None >>>>>> namespace=kube-system delete_after=False >>>>>> kubeconfig=/etc/origin/master/admin.kubeconfig content=None state=list >>>>>> debug=False selector=None name=master-api-master0.xxxxxx.com >>>>>> May 27 13:39:14 master0 origin-node: W0527 13:39:14.066911 20150 >>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>> /etc/cni/net.d >>>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.067321 20150 >>>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>>>>> cni config uninitialized >>>>>> May 27 13:39:14 master0 origin-node: E0527 13:39:14.814705 20150 >>>>>> summary.go:102] Failed to get system container stats for >>>>>> "/system.slice/origin-node.service": failed to get cgroup stats for >>>>>> "/system.slice/origin-node.service": failed to get container info for >>>>>> "/system.slice/origin-node.service": unknown container >>>>>> "/system.slice/origin-node.service" >>>>>> May 27 13:39:19 master0 origin-node: W0527 13:39:19.069450 20150 >>>>>> cni.go:172] Unable to update cni config: No networks found in >>>>>> /etc/cni/net.d >>>>>> May 27 13:39:19 master0 origin-node: E0527 13:39:19.069850 20150 >>>>>> kubelet.go:2101] Container runtime network not ready: NetworkReady=false >>>>>> reason:NetworkPluginNotReady message:docker: network plugin is not ready: >>>>>> cni config uninitialized >>>>>> >>>>>> On Mon, May 27, 2019 at 9:35 AM Shirly Radco <[email protected]> >>>>>> wrote: >>>>>> >>>>>>> Hi Jayme, >>>>>>> >>>>>>> Thank you for reaching out. >>>>>>> Please try rerunning the ansible playbook. >>>>>>> If this doesn't work, try adding to the integ.ini in the metrics vm >>>>>>> openshift_disable_check=docker_storage >>>>>>> and rerun the ansible playbook again. >>>>>>> >>>>>>> Please update how it goes. >>>>>>> >>>>>>> Best regards, >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> Shirly Radco >>>>>>> >>>>>>> BI Senior Software Engineer >>>>>>> >>>>>>> Red Hat <https://www.redhat.com/> >>>>>>> >>>>>>> <https://www.redhat.com/> >>>>>>> >>>>>>> >>>>>>> On Sun, May 26, 2019 at 9:34 PM Jayme <[email protected]> wrote: >>>>>>> >>>>>>>> I'm running in to this ansible error during oVirt metrics >>>>>>>> installation (following procedures at: >>>>>>>> https://ovirt.org/documentation/metrics-install-guide/Installing_Metrics_Store.html >>>>>>>> ) >>>>>>>> >>>>>>>> This is happening late in the process, after successfully deploying >>>>>>>> the installation VM and then running second step from the metrics VM. >>>>>>>> >>>>>>>> CHECK [memory_availability : master0.xxxxxx.com] >>>>>>>> ************************************************************************************************* >>>>>>>> fatal: [master0.xxxxxxx.com]: FAILED! => {"changed": true, >>>>>>>> "checks": {"disk_availability": {}, "docker_image_availability": >>>>>>>> {"changed": true}, "docker_storage": {"failed": true, "failures": >>>>>>>> [["OpenShiftCheckException", "Could not find imported module support >>>>>>>> code >>>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>>>> line 225, in run_check\n result = check.run()\n File >>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>>>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>>>>>>> {})\n File >>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>>>> line 211, in execute_module\n result = >>>>>>>> self._execute_module(module_name, >>>>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>> line 809, in _execute_module\n (module_style, shebang, module_data, >>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>> line 203, in _configure_module\n environment=final_environment)\n >>>>>>>> File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support >>>>>>>> code for >>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>> docker_common.py\n"]], "msg": "Could not find imported module support >>>>>>>> code >>>>>>>> for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>> docker_common.py\nTraceback (most recent call last):\n File >>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py\", >>>>>>>> line 225, in run_check\n result = check.run()\n File >>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py\", >>>>>>>> line 53, in run\n docker_info = self.execute_module(\"docker_info\", >>>>>>>> {})\n File >>>>>>>> \"/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py\", >>>>>>>> line 211, in execute_module\n result = >>>>>>>> self._execute_module(module_name, >>>>>>>> module_args, self.tmp, self.task_vars)\n File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>> line 809, in _execute_module\n (module_style, shebang, module_data, >>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>> module_args=module_args, task_vars=task_vars)\n File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py\", >>>>>>>> line 203, in _configure_module\n environment=final_environment)\n >>>>>>>> File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>> line 1023, in modify_module\n environment=environment)\n File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>> line 859, in _find_module_utils\n recursive_finder(module_name, >>>>>>>> b_module_data, py_module_names, py_module_cache, zf)\n File >>>>>>>> \"/usr/lib/python2.7/site-packages/ansible/executor/module_common.py\", >>>>>>>> line 621, in recursive_finder\n raise AnsibleError(' >>>>>>>> '.join(msg))\nAnsibleError: Could not find imported module support >>>>>>>> code for >>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>> docker_common.py\n"}, "memory_availability": {}, >>>>>>>> "package_availability": >>>>>>>> {"changed": false, "invocation": {"module_args": {"packages": >>>>>>>> ["PyYAML", >>>>>>>> "bash-completion", "bind", "ceph-common", "dnsmasq", "docker", >>>>>>>> "firewalld", >>>>>>>> "flannel", "glusterfs-fuse", "httpd-tools", "iptables", >>>>>>>> "iptables-services", "iscsi-initiator-utils", "libselinux-python", >>>>>>>> "nfs-utils", "ntp", "openssl", "origin", "origin-clients", >>>>>>>> "origin-hyperkube", "origin-node", "pyparted", "python-httplib2", >>>>>>>> "yum-utils"]}}}, "package_version": {"changed": false, "invocation": >>>>>>>> {"module_args": {"package_list": [{"check_multi": false, "name": >>>>>>>> "origin", >>>>>>>> "version": ""}, {"check_multi": false, "name": "origin-master", >>>>>>>> "version": >>>>>>>> ""}, {"check_multi": false, "name": "origin-node", "version": ""}], >>>>>>>> "package_mgr": "yum"}}}}, "msg": "One or more checks failed", >>>>>>>> "playbook_context": "install"} >>>>>>>> >>>>>>>> NO MORE HOSTS LEFT >>>>>>>> ******************************************************************************************************************************************* >>>>>>>> >>>>>>>> PLAY RECAP >>>>>>>> *************************************************************************************************************************************************** >>>>>>>> localhost : ok=35 changed=1 unreachable=0 >>>>>>>> failed=0 skipped=16 rescued=0 ignored=0 >>>>>>>> master0.xxxxxxx.com : ok=96 changed=6 unreachable=0 >>>>>>>> failed=1 skipped=165 rescued=0 ignored=0 >>>>>>>> >>>>>>>> >>>>>>>> INSTALLER STATUS >>>>>>>> ********************************************************************************************************************************************* >>>>>>>> Initialization : Complete (0:00:16) >>>>>>>> Health Check : In Progress (0:00:36) >>>>>>>> This phase can be restarted by running: >>>>>>>> playbooks/openshift-checks/pre-install.yml >>>>>>>> Sunday 26 May 2019 16:36:25 +0000 (0:00:36.151) 0:01:56.339 >>>>>>>> ************ >>>>>>>> >>>>>>>> =============================================================================== >>>>>>>> Run health checks (install) - EL >>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>> 36.15s >>>>>>>> os_firewall : Ensure iptables services are not enabled >>>>>>>> ------------------------------------------------------------------------------------------------ >>>>>>>> 2.74s >>>>>>>> openshift_repos : Ensure libselinux-python is installed >>>>>>>> ----------------------------------------------------------------------------------------------- >>>>>>>> 1.77s >>>>>>>> openshift_repos : refresh cache >>>>>>>> ----------------------------------------------------------------------------------------------------------------------- >>>>>>>> 1.60s >>>>>>>> Gather Cluster facts >>>>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>>>> 1.51s >>>>>>>> container_runtime : Fixup SELinux permissions for docker >>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>> 1.33s >>>>>>>> container_runtime : Place additional/blocked/insecure registries in >>>>>>>> /etc/containers/registries.conf >>>>>>>> --------------------------------------------------- 1.30s >>>>>>>> Ensure openshift-ansible installer package deps are installed >>>>>>>> ----------------------------------------------------------------------------------------- >>>>>>>> 1.29s >>>>>>>> container_runtime : Install Docker >>>>>>>> -------------------------------------------------------------------------------------------------------------------- >>>>>>>> 1.17s >>>>>>>> Initialize openshift.node.sdn_mtu >>>>>>>> --------------------------------------------------------------------------------------------------------------------- >>>>>>>> 1.13s >>>>>>>> os_firewall : Install firewalld packages >>>>>>>> -------------------------------------------------------------------------------------------------------------- >>>>>>>> 1.13s >>>>>>>> container_runtime : Set various Docker options >>>>>>>> -------------------------------------------------------------------------------------------------------- >>>>>>>> 1.11s >>>>>>>> install NetworkManager >>>>>>>> -------------------------------------------------------------------------------------------------------------------------------- >>>>>>>> 1.10s >>>>>>>> openshift_repos : Configure correct origin release repository >>>>>>>> ----------------------------------------------------------------------------------------- >>>>>>>> 1.05s >>>>>>>> container_runtime : Get current installed Docker version >>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>> 1.04s >>>>>>>> openshift_repos : Configure origin gpg keys >>>>>>>> ----------------------------------------------------------------------------------------------------------- >>>>>>>> 1.04s >>>>>>>> openshift_repos : Remove openshift_additional.repo file >>>>>>>> ----------------------------------------------------------------------------------------------- >>>>>>>> 0.99s >>>>>>>> container_runtime : Setup the docker-storage for overlay >>>>>>>> ---------------------------------------------------------------------------------------------- >>>>>>>> 0.96s >>>>>>>> Detecting Operating System from ostree_booted >>>>>>>> --------------------------------------------------------------------------------------------------------- >>>>>>>> 0.95s >>>>>>>> Gather Cluster facts >>>>>>>> ---------------------------------------------------------------------------------------------------------------------------------- >>>>>>>> 0.92s >>>>>>>> >>>>>>>> >>>>>>>> Failure summary: >>>>>>>> >>>>>>>> >>>>>>>> 1. Hosts: master0.xxxxxxx.com >>>>>>>> Play: OpenShift Health Checks >>>>>>>> Task: Run health checks (install) - EL >>>>>>>> Message: One or more checks failed >>>>>>>> Details: check "docker_storage": >>>>>>>> Could not find imported module support code for >>>>>>>> docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>> docker_common.py >>>>>>>> Traceback (most recent call last): >>>>>>>> File >>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/action_plugins/openshift_health_check.py", >>>>>>>> line 225, in run_check >>>>>>>> result = check.run() >>>>>>>> File >>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/docker_storage.py", >>>>>>>> line 53, in run >>>>>>>> docker_info = self.execute_module("docker_info", >>>>>>>> {}) >>>>>>>> File >>>>>>>> "/usr/share/ansible/openshift-ansible/roles/openshift_health_checker/openshift_checks/__init__.py", >>>>>>>> line 211, in execute_module >>>>>>>> result = self._execute_module(module_name, >>>>>>>> module_args, self.tmp, self.task_vars) >>>>>>>> File >>>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>>>> line >>>>>>>> 809, in _execute_module >>>>>>>> (module_style, shebang, module_data, >>>>>>>> module_path) = self._configure_module(module_name=module_name, >>>>>>>> module_args=module_args, task_vars=task_vars) >>>>>>>> File >>>>>>>> "/usr/lib/python2.7/site-packages/ansible/plugins/action/__init__.py", >>>>>>>> line >>>>>>>> 203, in _configure_module >>>>>>>> environment=final_environment) >>>>>>>> File >>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>> line >>>>>>>> 1023, in modify_module >>>>>>>> environment=environment) >>>>>>>> File >>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>> line >>>>>>>> 859, in _find_module_utils >>>>>>>> recursive_finder(module_name, b_module_data, >>>>>>>> py_module_names, py_module_cache, zf) >>>>>>>> File >>>>>>>> "/usr/lib/python2.7/site-packages/ansible/executor/module_common.py", >>>>>>>> line >>>>>>>> 621, in recursive_finder >>>>>>>> raise AnsibleError(' '.join(msg)) >>>>>>>> AnsibleError: Could not find imported module support >>>>>>>> code for docker_info. Looked for either AnsibleDockerClient.py or >>>>>>>> docker_common.py >>>>>>>> >>>>>>>> >>>>>>>> The execution of "install_okd.yaml" includes checks designed to >>>>>>>> fail early if the requirements of the playbook are not met. One or >>>>>>>> more of >>>>>>>> these checks failed. To disregard these results,explicitly disable >>>>>>>> checks >>>>>>>> by setting an Ansible variable: >>>>>>>> openshift_disable_check=docker_storage >>>>>>>> Failing check names are shown in the failure details above. Some >>>>>>>> checks may be configurable by variables if your requirements are >>>>>>>> different >>>>>>>> from the defaults; consult check documentation. >>>>>>>> Variables can be set in the inventory or passed on the command line >>>>>>>> using the -e flag to ansible-playbook. >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list -- [email protected] >>>>>>>> To unsubscribe send an email to [email protected] >>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>> oVirt Code of Conduct: >>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>>> List Archives: >>>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/SEFPOF36T7G4GIIGHERUBKTNOPEMVFSM/ >>>>>>>> >>>>>>> _______________________________________________ >>> Users mailing list -- [email protected] >>> To unsubscribe send an email to [email protected] >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>> oVirt Code of Conduct: >>> https://www.ovirt.org/community/about/community-guidelines/ >>> List Archives: >>> https://lists.ovirt.org/archives/list/[email protected]/message/CN6G76Z7MVNK6AEZQ2I7SHULGXSXHCNQ/ >>> >>
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/BCADLACSZRPEXS5ENYRUX3GVWGUWMFSD/

