I guess  your only option is to edit 
/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/main.yml  and  
replace  'package' with 'dnf' (keep the beginning 2  "spaces" deeper  than  '- 
name' -> just where  "package" starts).

Best Regards,
Strahil Nikolov

На 15 юли 2020 г. 22:39:09 GMT+03:00, clam2...@gmail.com написа:
>Thank you very much Strahil for your continued assistance.  I have
>tried cleaning up and redeploying four additional times and am still
>experiencing the same error.
>
>To Summarize
>
>(1)
>Attempt 1: change gluster_infra_thick_lvs --> size: 100G to size:
>'100%PVS' and change gluster_infra_thinpools --> lvsize: 500G to
>lvsize: '100%PVS'
>Result 1: deployment failed -->
>TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools
>for RHEL systems.] ***
>task path:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/main.yml:33
>fatal: [fmov1n3.sn.dtcorp.com]: FAILED! => {"changed": false, "msg":
>"The Python 2 yum module is needed for this module. If you require
>Python 3 support use the `dnf` Ansible module instead."}
>fatal: [fmov1n1.sn.dtcorp.com]: FAILED! => {"changed": false, "msg":
>"The Python 2 yum module is needed for this module. If you require
>Python 3 support use the `dnf` Ansible module instead."}
>fatal: [fmov1n2.sn.dtcorp.com]: FAILED! => {"changed": false, "msg":
>"The Python 2 yum module is needed for this module. If you require
>Python 3 support use the `dnf` Ansible module instead."}
>
>(2)
>Attempt 2: same as Attempt 1, but substituted 99G for '100%PVS'
>Result 2: same as Result 1
>
>(3)
>Attempt 3: same as Attempt 1, but added
>vars:
>  ansible_python_interpreter: /usr/bin/python3
>Result 3: same as Result 1
>
>(4)
>Attempt 4: reboot all three nodes, same as Attempt 1 but omitted
>previously edited size arguments as I read in documentation at
>https://github.com/gluster/gluster-ansible-infra that the size/lvsize
>arguements for variables gluster_infra_thick_lvs and
>gluster_infra_lv_logicalvols are optional and default to 100% size of
>LV.
>
>At the end of this post are the latest version of the playbook and log
>output.  As best I can tell the nodes are fully updated, default
>installs using verified images of v4.4.1.1.
>
>From /var/log/cockpit/ovirt-dashboard/gluster-deployment.log I see that
>line 33 in task path:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/main.yml is
>what is causing the deployment to fail at this point
>
>- name: Change to Install lvm tools for RHEL systems.
>  package:
>    name: device-mapper-persistent-data
>    state: present
>  when: ansible_os_family == 'RedHat'
>
>But package device-mapper-persistent-data is installed:
>
>[root@fmov1n1 ~]# dnf install device-mapper-persistent-data
>Last metadata expiration check: 0:32:10 ago on Wed 15 Jul 2020 06:44:19
>PM UTC.
>Package device-mapper-persistent-data-0.8.5-3.el8.x86_64 is already
>installed.
>Dependencies resolved.
>Nothing to do.
>Complete!
>
>[root@fmov1n1 ~]# dnf info device-mapper-persistent-data
>Last metadata expiration check: 0:31:44 ago on Wed 15 Jul 2020 06:44:19
>PM UTC.
>Installed Packages
>Name         : device-mapper-persistent-data
>Version      : 0.8.5
>Release      : 3.el8
>Architecture : x86_64
>Size         : 1.4 M
>Source       : device-mapper-persistent-data-0.8.5-3.el8.src.rpm
>Repository   : @System
>Summary      : Device-mapper Persistent Data Tools
>URL          : https://github.com/jthornber/thin-provisioning-tools
>License      : GPLv3+
>Description  : thin-provisioning-tools contains
>check,dump,restore,repair,rmap
>    : and metadata_size tools to manage device-mapper thin provisioning
>      : target metadata devices; cache check,dump,metadata_size,restore
>      : and repair tools to manage device-mapper cache metadata devices
>   : are included and era check, dump, restore and invalidate to manage
>             : snapshot eras
>
>I can't figure out why Ansible v2.9.10 is not calling DNF.  Ansible DNF
>package is installed:
>
>[root@fmov1n1 modules]# ansible-doc -t module dnf
>> DNF   
>(/usr/lib/python3.6/site-packages/ansible/modules/packaging/os/dnf.py)
>
>Installs, upgrade, removes, and lists packages and groups with the
>`dnf' package
>        manager.
>
>  * This module is maintained by The Ansible Core Team
>...
>
>
>I am unsure how to further troubleshoot from here!
>
>Thank you again!!!
>Charles
>
>---
>Latest Gluster Playbook (edited from Wizard output)
>
>hc_nodes:
>  hosts:
>    fmov1n1.sn.dtcorp.com:
>      gluster_infra_volume_groups:
>        - vgname: gluster_vg_nvme0n1
>          pvname: /dev/mapper/vdo_nvme0n1
>        - vgname: gluster_vg_nvme2n1
>          pvname: /dev/mapper/vdo_nvme2n1
>        - vgname: gluster_vg_nvme1n1
>          pvname: /dev/mapper/vdo_nvme1n1
>      gluster_infra_mount_devices:
>        - path: /gluster_bricks/engine
>          lvname: gluster_lv_engine
>          vgname: gluster_vg_nvme0n1
>        - path: /gluster_bricks/data
>          lvname: gluster_lv_data
>          vgname: gluster_vg_nvme2n1
>        - path: /gluster_bricks/vmstore
>          lvname: gluster_lv_vmstore
>          vgname: gluster_vg_nvme1n1
>      gluster_infra_vdo:
>        - name: vdo_nvme0n1
>          device: /dev/nvme0n1
>          slabsize: 2G
>          logicalsize: 100G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>        - name: vdo_nvme2n1
>          device: /dev/nvme2n1
>          slabsize: 2G
>          logicalsize: 500G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>        - name: vdo_nvme1n1
>          device: /dev/nvme1n1
>          slabsize: 2G
>          logicalsize: 500G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>      blacklist_mpath_devices:
>        - nvme0n1
>        - nvme2n1
>        - nvme1n1
>      gluster_infra_thick_lvs:
>        - vgname: gluster_vg_nvme0n1
>          lvname: gluster_lv_engine
>      gluster_infra_thinpools:
>        - vgname: gluster_vg_nvme2n1
>          thinpoolname: gluster_thinpool_gluster_vg_nvme2n1
>          poolmetadatasize: 1G
>        - vgname: gluster_vg_nvme1n1
>          thinpoolname: gluster_thinpool_gluster_vg_nvme1n1
>          poolmetadatasize: 1G
>      gluster_infra_lv_logicalvols:
>        - vgname: gluster_vg_nvme2n1
>          thinpool: gluster_thinpool_gluster_vg_nvme2n1
>          lvname: gluster_lv_data
>        - vgname: gluster_vg_nvme1n1
>          thinpool: gluster_thinpool_gluster_vg_nvme1n1
>          lvname: gluster_lv_vmstore
>    fmov1n2.sn.dtcorp.com:
>      gluster_infra_volume_groups:
>        - vgname: gluster_vg_nvme0n1
>          pvname: /dev/mapper/vdo_nvme0n1
>        - vgname: gluster_vg_nvme2n1
>          pvname: /dev/mapper/vdo_nvme2n1
>        - vgname: gluster_vg_nvme1n1
>          pvname: /dev/mapper/vdo_nvme1n1
>      gluster_infra_mount_devices:
>        - path: /gluster_bricks/engine
>          lvname: gluster_lv_engine
>          vgname: gluster_vg_nvme0n1
>        - path: /gluster_bricks/data
>          lvname: gluster_lv_data
>          vgname: gluster_vg_nvme2n1
>        - path: /gluster_bricks/vmstore
>          lvname: gluster_lv_vmstore
>          vgname: gluster_vg_nvme1n1
>      gluster_infra_vdo:
>        - name: vdo_nvme0n1
>          device: /dev/nvme0n1
>          slabsize: 2G
>          logicalsize: 100G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>        - name: vdo_nvme2n1
>          device: /dev/nvme2n1
>          slabsize: 2G
>          logicalsize: 500G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>        - name: vdo_nvme1n1
>          device: /dev/nvme1n1
>          slabsize: 2G
>          logicalsize: 500G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>      blacklist_mpath_devices:
>        - nvme0n1
>        - nvme2n1
>        - nvme1n1
>      gluster_infra_thick_lvs:
>        - vgname: gluster_vg_nvme0n1
>          lvname: gluster_lv_engine
>      gluster_infra_thinpools:
>        - vgname: gluster_vg_nvme2n1
>          thinpoolname: gluster_thinpool_gluster_vg_nvme2n1
>          poolmetadatasize: 1G
>        - vgname: gluster_vg_nvme1n1
>          thinpoolname: gluster_thinpool_gluster_vg_nvme1n1
>          poolmetadatasize: 1G
>      gluster_infra_lv_logicalvols:
>        - vgname: gluster_vg_nvme2n1
>          thinpool: gluster_thinpool_gluster_vg_nvme2n1
>          lvname: gluster_lv_data
>        - vgname: gluster_vg_nvme1n1
>          thinpool: gluster_thinpool_gluster_vg_nvme1n1
>          lvname: gluster_lv_vmstore
>    fmov1n3.sn.dtcorp.com:
>      gluster_infra_volume_groups:
>        - vgname: gluster_vg_nvme0n1
>          pvname: /dev/mapper/vdo_nvme0n1
>        - vgname: gluster_vg_nvme2n1
>          pvname: /dev/mapper/vdo_nvme2n1
>        - vgname: gluster_vg_nvme1n1
>          pvname: /dev/mapper/vdo_nvme1n1
>      gluster_infra_mount_devices:
>        - path: /gluster_bricks/engine
>          lvname: gluster_lv_engine
>          vgname: gluster_vg_nvme0n1
>        - path: /gluster_bricks/data
>          lvname: gluster_lv_data
>          vgname: gluster_vg_nvme2n1
>        - path: /gluster_bricks/vmstore
>          lvname: gluster_lv_vmstore
>          vgname: gluster_vg_nvme1n1
>      gluster_infra_vdo:
>        - name: vdo_nvme0n1
>          device: /dev/nvme0n1
>          slabsize: 2G
>          logicalsize: 100G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>        - name: vdo_nvme2n1
>          device: /dev/nvme2n1
>          slabsize: 2G
>          logicalsize: 500G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>        - name: vdo_nvme1n1
>          device: /dev/nvme1n1
>          slabsize: 2G
>          logicalsize: 500G
>          blockmapcachesize: 128M
>          emulate512: 'off'
>          writepolicy: auto
>          maxDiscardSize: 16M
>      blacklist_mpath_devices:
>        - nvme0n1
>        - nvme2n1
>        - nvme1n1
>      gluster_infra_thick_lvs:
>        - vgname: gluster_vg_nvme0n1
>          lvname: gluster_lv_engine
>      gluster_infra_thinpools:
>        - vgname: gluster_vg_nvme2n1
>          thinpoolname: gluster_thinpool_gluster_vg_nvme2n1
>          poolmetadatasize: 1G
>        - vgname: gluster_vg_nvme1n1
>          thinpoolname: gluster_thinpool_gluster_vg_nvme1n1
>          poolmetadatasize: 1G
>      gluster_infra_lv_logicalvols:
>        - vgname: gluster_vg_nvme2n1
>          thinpool: gluster_thinpool_gluster_vg_nvme2n1
>          lvname: gluster_lv_data
>        - vgname: gluster_vg_nvme1n1
>          thinpool: gluster_thinpool_gluster_vg_nvme1n1
>          lvname: gluster_lv_vmstore
>  vars:
>    ansible_python_interpreter: /usr/bin/python3
>    gluster_infra_disktype: JBOD
>    gluster_set_selinux_labels: true
>    gluster_infra_fw_ports:
>      - 2049/tcp
>      - 54321/tcp
>      - 5900/tcp
>      - 5900-6923/tcp
>      - 5666/tcp
>      - 16514/tcp
>    gluster_infra_fw_permanent: true
>    gluster_infra_fw_state: enabled
>    gluster_infra_fw_zone: public
>    gluster_infra_fw_services:
>      - glusterfs
>    gluster_features_force_varlogsizecheck: false
>    cluster_nodes:
>      - fmov1n1.sn.dtcorp.com
>      - fmov1n2.sn.dtcorp.com
>      - fmov1n3.sn.dtcorp.com
>    gluster_features_hci_cluster: '{{ cluster_nodes }}'
>    gluster_features_hci_volumes:
>      - volname: engine
>        brick: /gluster_bricks/engine/engine
>        arbiter: 0
>      - volname: data
>        brick: /gluster_bricks/data/data
>        arbiter: 0
>      - volname: vmstore
>        brick: /gluster_bricks/vmstore/vmstore
>        arbiter: 0
>
>---
>Latest /var/log/cockpit/ovirt-dashboard/gluster-deployment.log
>
>
>[root@fmov1n1 modules]# cat
>/var/log/cockpit/ovirt-dashboard/gluster-deployment.log
>ansible-playbook 2.9.10
>  config file = /etc/ansible/ansible.cfg
>configured module search path = ['/root/.ansible/plugins/modules',
>'/usr/share/ansible/plugins/modules']
>ansible python module location =
>/usr/lib/python3.6/site-packages/ansible
>  executable location = /root/../usr/bin/ansible-playbook
>python version = 3.6.8 (default, Apr 16 2020, 01:36:27) [GCC 8.3.1
>20191121 (Red Hat 8.3.1-5)]
>Using /etc/ansible/ansible.cfg as config file
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/blacklist_mpath_devices.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/lvm_config.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vdo_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/main-lvm.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vg_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thick_lv_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thin_pool_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thin_volume_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/cache_setup.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vdo_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/main-lvm.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vg_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thick_lv_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thin_pool_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thin_volume_create.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/cache_setup.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/fscreate.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/mount.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/lvm_kernelparams.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/fstrim_service.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/luks_device_encrypt.yml
>statically imported:
>/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/bind_tang_server.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/nfs_ganesha/tasks/prerequisites.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/nfs_ganesha/tasks/distribute_keys.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/nfs_ganesha/tasks/master_tasks.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/nfs_ganesha/tasks/enable_ganesha.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/nfs_ganesha/tasks/add_new_nodes.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/gluster_hci/tasks/prerequisites.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/gluster_hci/tasks/glusterd_ipv6.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/gluster_hci/tasks/hci_volumes.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/gluster_hci/tasks/ssl-setup.yml
>statically imported:
>/etc/ansible/roles/gluster.features/roles/ctdb/tasks/setup_ctdb.yml
>
>PLAYBOOK: hc_wizard.yml
>********************************************************
>1 plays in
>/root/../usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.yml
>
>PLAY [Setup backend]
>***********************************************************
>
>TASK [Gathering Facts]
>*********************************************************
>task path: /usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.yml:4
>ok: [fmov1n2.sn.dtcorp.com]
>ok: [fmov1n1.sn.dtcorp.com]
>ok: [fmov1n3.sn.dtcorp.com]
>
>TASK [Check if valid hostnames are provided]
>***********************************
>task path: /usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.yml:16
>changed: [fmov1n1.sn.dtcorp.com] => (item=fmov1n1.sn.dtcorp.com) =>
>{"ansible_loop_var": "item", "changed": true, "cmd": ["getent",
>"ahosts", "fmov1n1.sn.dtcorp.com"], "delta": "0:00:00.006835", "end":
>"2020-07-15 18:03:58.366109", "item": "fmov1n1.sn.dtcorp.com", "rc": 0,
>"start": "2020-07-15 18:03:58.359274", "stderr": "", "stderr_lines":
>[], "stdout": "172.16.16.21    STREAM
>fmov1n1.sn.dtcorp.com\n172.16.16.21    DGRAM  \n172.16.16.21    RAW   
>", "stdout_lines": ["172.16.16.21    STREAM fmov1n1.sn.dtcorp.com",
>"172.16.16.21    DGRAM  ", "172.16.16.21    RAW    "]}
>changed: [fmov1n1.sn.dtcorp.com] => (item=fmov1n2.sn.dtcorp.com) =>
>{"ansible_loop_var": "item", "changed": true, "cmd": ["getent",
>"ahosts", "fmov1n2.sn.dtcorp.com"], "delta": "0:00:00.004972", "end":
>"2020-07-15 18:03:58.569094", "item": "fmov1n2.sn.dtcorp.com", "rc": 0,
>"start": "2020-07-15 18:03:58.564122", "stderr": "", "stderr_lines":
>[], "stdout": "172.16.16.22    STREAM
>fmov1n2.sn.dtcorp.com\n172.16.16.22    DGRAM  \n172.16.16.22    RAW   
>", "stdout_lines": ["172.16.16.22    STREAM fmov1n2.sn.dtcorp.com",
>"172.16.16.22    DGRAM  ", "172.16.16.22    RAW    "]}
>changed: [fmov1n1.sn.dtcorp.com] => (item=fmov1n3.sn.dtcorp.com) =>
>{"ansible_loop_var": "item", "changed": true, "cmd": ["getent",
>"ahosts", "fmov1n3.sn.dtcorp.com"], "delta": "0:00:00.004759", "end":
>"2020-07-15 18:03:58.769052", "item": "fmov1n3.sn.dtcorp.com", "rc": 0,
>"start": "2020-07-15 18:03:58.764293", "stderr": "", "stderr_lines":
>[], "stdout": "172.16.16.23    STREAM
>fmov1n3.sn.dtcorp.com\n172.16.16.23    DGRAM  \n172.16.16.23    RAW   
>", "stdout_lines": ["172.16.16.23    STREAM fmov1n3.sn.dtcorp.com",
>"172.16.16.23    DGRAM  ", "172.16.16.23    RAW    "]}
>
>TASK [Check if provided hostnames are valid]
>***********************************
>task path: /usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.yml:29
>ok: [fmov1n1.sn.dtcorp.com] => {
>    "changed": false,
>    "msg": "All assertions passed"
>}
>ok: [fmov1n2.sn.dtcorp.com] => {
>    "changed": false,
>    "msg": "All assertions passed"
>}
>ok: [fmov1n3.sn.dtcorp.com] => {
>    "changed": false,
>    "msg": "All assertions passed"
>}
>
>TASK [Check if /var/log has enough disk space]
>*********************************
>task path: /usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.yml:38
>skipping: [fmov1n1.sn.dtcorp.com] => {"changed": false, "skip_reason":
>"Conditional result was False"}
>skipping: [fmov1n2.sn.dtcorp.com] => {"changed": false, "skip_reason":
>"Conditional result was False"}
>skipping: [fmov1n3.sn.dtcorp.com] => {"changed": false, "skip_reason":
>"Conditional result was False"}
>
>TASK [Check if the /var is greater than 15G]
>***********************************
>task path: /usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.yml:43
>skipping: [fmov1n1.sn.dtcorp.com] => {"changed": false, "skip_reason":
>"Conditional result was False"}
>skipping: [fmov1n2.sn.dtcorp.com] => {"changed": false, "skip_reason":
>"Conditional result was False"}
>skipping: [fmov1n3.sn.dtcorp.com] => {"changed": false, "skip_reason":
>"Conditional result was False"}
>
>TASK [Check if disks have logical block size of 512B]
>**************************
>task path: /usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.yml:53
>skipping: [fmov1n1.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme0n1", "vgname": "gluster_vg_nvme0n1"},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n1.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme2n1", "vgname": "gluster_vg_nvme2n1"},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n1.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme1n1", "vgname": "gluster_vg_nvme1n1"},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n2.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme0n1", "vgname": "gluster_vg_nvme0n1"},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n2.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme2n1", "vgname": "gluster_vg_nvme2n1"},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n2.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme1n1", "vgname": "gluster_vg_nvme1n1"},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n3.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme0n1", "vgname": "gluster_vg_nvme0n1"},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n3.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme2n1", "vgname": "gluster_vg_nvme2n1"},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n3.sn.dtcorp.com] => (item={'vgname':
>'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'})  =>
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme1n1", "vgname": "gluster_vg_nvme1n1"},
>"skip_reason": "Conditional result was False"}
>
>TASK [Check if logical block size is 512 bytes]
>********************************
>task path: /usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.yml:61
>skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)  =>
>{"ansible_loop_var": "item", "changed": false, "item":
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme0n1", "vgname": "gluster_vg_nvme0n1"},
>"skip_reason": "Conditional result was False", "skipped": true},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)  =>
>{"ansible_loop_var": "item", "changed": false, "item":
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme2n1", "vgname": "gluster_vg_nvme2n1"},
>"skip_reason": "Conditional result was False", "skipped": true},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)  =>
>{"ansible_loop_var": "item", "changed": false, "item":
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme1n1", "vgname": "gluster_vg_nvme1n1"},
>"skip_reason": "Conditional result was False", "skipped": true},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n2.sn.dtcorp.com] => (item=Logical Block Size)  =>
>{"ansible_loop_var": "item", "changed": false, "item":
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme0n1", "vgname": "gluster_vg_nvme0n1"},
>"skip_reason": "Conditional result was False", "skipped": true},
>"skip_reason": "Conditional result was False"}
>skipping: [fmov1n2.sn.dtcorp.com] => (item=Logical Block Size)  =>
>{"ansible_loop_var": "item", "changed": false, "item":
>{"ansible_loop_var": "item", "changed": false, "item": {"pvname":
>"/dev/mapper/vdo_nvme2n1", "vgname": "gluster_vg_nvme2n1"},
>"skip_reason": "Conditional resul
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MG6SXD62XE5ORJFQNXPJKBEAFJMQMMNN/

Reply via email to