Thank you Strahil. I think I edited the oVirt Node Cockpit Hyperconverged Wizard Gluster Deployment Ansible playbook as detailed in your post and received the following new failure:
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.] *** fatal: [fmov1n3.sn.dtcorp.com]: FAILED! => {"changed": false, "msg": "The Python 2 yum module is needed for this module. If you require Python 3 support use the `dnf` Ansible module instead."} fatal: [fmov1n1.sn.dtcorp.com]: FAILED! => {"changed": false, "msg": "The Python 2 yum module is needed for this module. If you require Python 3 support use the `dnf` Ansible module instead."} fatal: [fmov1n2.sn.dtcorp.com]: FAILED! => {"changed": false, "msg": "The Python 2 yum module is needed for this module. If you require Python 3 support use the `dnf` Ansible module instead."} Any further assistance is most appreciated!!! Respectfully, Charles --- Gluster Deployment Ansible Playbook hc_nodes: hosts: fmov1n1.sn.dtcorp.com: gluster_infra_volume_groups: - vgname: gluster_vg_nvme0n1 pvname: /dev/mapper/vdo_nvme0n1 - vgname: gluster_vg_nvme2n1 pvname: /dev/mapper/vdo_nvme2n1 - vgname: gluster_vg_nvme1n1 pvname: /dev/mapper/vdo_nvme1n1 gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_nvme0n1 - path: /gluster_bricks/data lvname: gluster_lv_data vgname: gluster_vg_nvme2n1 - path: /gluster_bricks/vmstore lvname: gluster_lv_vmstore vgname: gluster_vg_nvme1n1 gluster_infra_vdo: - name: vdo_nvme0n1 device: /dev/nvme0n1 slabsize: 2G logicalsize: 1000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M - name: vdo_nvme2n1 device: /dev/nvme2n1 slabsize: 32G logicalsize: 5000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M - name: vdo_nvme1n1 device: /dev/nvme1n1 slabsize: 32G logicalsize: 5000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M blacklist_mpath_devices: - nvme0n1 - nvme2n1 - nvme1n1 gluster_infra_thick_lvs: - vgname: gluster_vg_nvme0n1 lvname: gluster_lv_engine size: '100%PVS' gluster_infra_thinpools: - vgname: gluster_vg_nvme2n1 thinpoolname: gluster_thinpool_gluster_vg_nvme2n1 poolmetadatasize: 3G - vgname: gluster_vg_nvme1n1 thinpoolname: gluster_thinpool_gluster_vg_nvme1n1 poolmetadatasize: 3G gluster_infra_lv_logicalvols: - vgname: gluster_vg_nvme2n1 thinpool: gluster_thinpool_gluster_vg_nvme2n1 lvname: gluster_lv_data lvsize: '100%PVS' - vgname: gluster_vg_nvme1n1 thinpool: gluster_thinpool_gluster_vg_nvme1n1 lvname: gluster_lv_vmstore lvsize: '100%PVS' fmov1n2.sn.dtcorp.com: gluster_infra_volume_groups: - vgname: gluster_vg_nvme0n1 pvname: /dev/mapper/vdo_nvme0n1 - vgname: gluster_vg_nvme2n1 pvname: /dev/mapper/vdo_nvme2n1 - vgname: gluster_vg_nvme1n1 pvname: /dev/mapper/vdo_nvme1n1 gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_nvme0n1 - path: /gluster_bricks/data lvname: gluster_lv_data vgname: gluster_vg_nvme2n1 - path: /gluster_bricks/vmstore lvname: gluster_lv_vmstore vgname: gluster_vg_nvme1n1 gluster_infra_vdo: - name: vdo_nvme0n1 device: /dev/nvme0n1 slabsize: 2G logicalsize: 1000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M - name: vdo_nvme2n1 device: /dev/nvme2n1 slabsize: 32G logicalsize: 5000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M - name: vdo_nvme1n1 device: /dev/nvme1n1 slabsize: 32G logicalsize: 5000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M blacklist_mpath_devices: - nvme0n1 - nvme2n1 - nvme1n1 gluster_infra_thick_lvs: - vgname: gluster_vg_nvme0n1 lvname: gluster_lv_engine size: '100%PVS' gluster_infra_thinpools: - vgname: gluster_vg_nvme2n1 thinpoolname: gluster_thinpool_gluster_vg_nvme2n1 poolmetadatasize: 3G - vgname: gluster_vg_nvme1n1 thinpoolname: gluster_thinpool_gluster_vg_nvme1n1 poolmetadatasize: 3G gluster_infra_lv_logicalvols: - vgname: gluster_vg_nvme2n1 thinpool: gluster_thinpool_gluster_vg_nvme2n1 lvname: gluster_lv_data lvsize: '100%PVS' - vgname: gluster_vg_nvme1n1 thinpool: gluster_thinpool_gluster_vg_nvme1n1 lvname: gluster_lv_vmstore lvsize: '100%PVS' fmov1n3.sn.dtcorp.com: gluster_infra_volume_groups: - vgname: gluster_vg_nvme0n1 pvname: /dev/mapper/vdo_nvme0n1 - vgname: gluster_vg_nvme2n1 pvname: /dev/mapper/vdo_nvme2n1 - vgname: gluster_vg_nvme1n1 pvname: /dev/mapper/vdo_nvme1n1 gluster_infra_mount_devices: - path: /gluster_bricks/engine lvname: gluster_lv_engine vgname: gluster_vg_nvme0n1 - path: /gluster_bricks/data lvname: gluster_lv_data vgname: gluster_vg_nvme2n1 - path: /gluster_bricks/vmstore lvname: gluster_lv_vmstore vgname: gluster_vg_nvme1n1 gluster_infra_vdo: - name: vdo_nvme0n1 device: /dev/nvme0n1 slabsize: 2G logicalsize: 1000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M - name: vdo_nvme2n1 device: /dev/nvme2n1 slabsize: 32G logicalsize: 5000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M - name: vdo_nvme1n1 device: /dev/nvme1n1 slabsize: 32G logicalsize: 5000G blockmapcachesize: 128M emulate512: 'off' writepolicy: auto maxDiscardSize: 16M blacklist_mpath_devices: - nvme0n1 - nvme2n1 - nvme1n1 gluster_infra_thick_lvs: - vgname: gluster_vg_nvme0n1 lvname: gluster_lv_engine size: '100%PVS' gluster_infra_thinpools: - vgname: gluster_vg_nvme2n1 thinpoolname: gluster_thinpool_gluster_vg_nvme2n1 poolmetadatasize: 3G - vgname: gluster_vg_nvme1n1 thinpoolname: gluster_thinpool_gluster_vg_nvme1n1 poolmetadatasize: 3G gluster_infra_lv_logicalvols: - vgname: gluster_vg_nvme2n1 thinpool: gluster_thinpool_gluster_vg_nvme2n1 lvname: gluster_lv_data lvsize: '100%PVS' - vgname: gluster_vg_nvme1n1 thinpool: gluster_thinpool_gluster_vg_nvme1n1 lvname: gluster_lv_vmstore lvsize: '100%PVS' vars: gluster_infra_disktype: JBOD gluster_set_selinux_labels: true gluster_infra_fw_ports: - 2049/tcp - 54321/tcp - 5900/tcp - 5900-6923/tcp - 5666/tcp - 16514/tcp gluster_infra_fw_permanent: true gluster_infra_fw_state: enabled gluster_infra_fw_zone: public gluster_infra_fw_services: - glusterfs gluster_features_force_varlogsizecheck: false cluster_nodes: - fmov1n1.sn.dtcorp.com - fmov1n2.sn.dtcorp.com - fmov1n3.sn.dtcorp.com gluster_features_hci_cluster: '{{ cluster_nodes }}' gluster_features_hci_volumes: - volname: engine brick: /gluster_bricks/engine/engine arbiter: 0 - volname: data brick: /gluster_bricks/data/data arbiter: 0 - volname: vmstore brick: /gluster_bricks/vmstore/vmstore arbiter: 0 --- /var/log/cockpit/ovirt-dashboard/gluster-deployment.log PLAY [Setup backend] *********************************************************** TASK [Gathering Facts] ********************************************************* ok: [fmov1n1.sn.dtcorp.com] ok: [fmov1n2.sn.dtcorp.com] ok: [fmov1n3.sn.dtcorp.com] TASK [Check if valid hostnames are provided] *********************************** changed: [fmov1n1.sn.dtcorp.com] => (item=fmov1n1.sn.dtcorp.com) changed: [fmov1n1.sn.dtcorp.com] => (item=fmov1n2.sn.dtcorp.com) changed: [fmov1n1.sn.dtcorp.com] => (item=fmov1n3.sn.dtcorp.com) TASK [Check if provided hostnames are valid] *********************************** ok: [fmov1n1.sn.dtcorp.com] => { "changed": false, "msg": "All assertions passed" } ok: [fmov1n2.sn.dtcorp.com] => { "changed": false, "msg": "All assertions passed" } ok: [fmov1n3.sn.dtcorp.com] => { "changed": false, "msg": "All assertions passed" } TASK [Check if /var/log has enough disk space] ********************************* skipping: [fmov1n1.sn.dtcorp.com] skipping: [fmov1n2.sn.dtcorp.com] skipping: [fmov1n3.sn.dtcorp.com] TASK [Check if the /var is greater than 15G] *********************************** skipping: [fmov1n1.sn.dtcorp.com] skipping: [fmov1n2.sn.dtcorp.com] skipping: [fmov1n3.sn.dtcorp.com] TASK [Check if disks have logical block size of 512B] ************************** skipping: [fmov1n1.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'}) skipping: [fmov1n1.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'}) skipping: [fmov1n1.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'}) skipping: [fmov1n2.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'}) skipping: [fmov1n2.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'}) skipping: [fmov1n2.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'}) skipping: [fmov1n3.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'}) skipping: [fmov1n3.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'}) skipping: [fmov1n3.sn.dtcorp.com] => (item={'vgname': 'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'}) TASK [Check if logical block size is 512 bytes] ******************************** skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n2.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n2.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n2.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n3.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n3.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n3.sn.dtcorp.com] => (item=Logical Block Size) TASK [Get logical block size of VDO devices] *********************************** skipping: [fmov1n1.sn.dtcorp.com] => (item={'name': 'vdo_nvme0n1', 'device': '/dev/nvme0n1', 'slabsize': '2G', 'logicalsize': '1000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) skipping: [fmov1n1.sn.dtcorp.com] => (item={'name': 'vdo_nvme2n1', 'device': '/dev/nvme2n1', 'slabsize': '32G', 'logicalsize': '5000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) skipping: [fmov1n1.sn.dtcorp.com] => (item={'name': 'vdo_nvme1n1', 'device': '/dev/nvme1n1', 'slabsize': '32G', 'logicalsize': '5000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) skipping: [fmov1n2.sn.dtcorp.com] => (item={'name': 'vdo_nvme0n1', 'device': '/dev/nvme0n1', 'slabsize': '2G', 'logicalsize': '1000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) skipping: [fmov1n2.sn.dtcorp.com] => (item={'name': 'vdo_nvme2n1', 'device': '/dev/nvme2n1', 'slabsize': '32G', 'logicalsize': '5000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) skipping: [fmov1n2.sn.dtcorp.com] => (item={'name': 'vdo_nvme1n1', 'device': '/dev/nvme1n1', 'slabsize': '32G', 'logicalsize': '5000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) skipping: [fmov1n3.sn.dtcorp.com] => (item={'name': 'vdo_nvme0n1', 'device': '/dev/nvme0n1', 'slabsize': '2G', 'logicalsize': '1000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) skipping: [fmov1n3.sn.dtcorp.com] => (item={'name': 'vdo_nvme2n1', 'device': '/dev/nvme2n1', 'slabsize': '32G', 'logicalsize': '5000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) skipping: [fmov1n3.sn.dtcorp.com] => (item={'name': 'vdo_nvme1n1', 'device': '/dev/nvme1n1', 'slabsize': '32G', 'logicalsize': '5000G', 'blockmapcachesize': '128M', 'emulate512': 'off', 'writepolicy': 'auto', 'maxDiscardSize': '16M'}) TASK [Check if logical block size is 512 bytes for VDO devices] **************** skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n1.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n2.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n2.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n2.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n3.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n3.sn.dtcorp.com] => (item=Logical Block Size) skipping: [fmov1n3.sn.dtcorp.com] => (item=Logical Block Size) TASK [gluster.infra/roles/firewall_config : Start firewalld if not already started] *** ok: [fmov1n3.sn.dtcorp.com] ok: [fmov1n2.sn.dtcorp.com] ok: [fmov1n1.sn.dtcorp.com] TASK [gluster.infra/roles/firewall_config : check if required variables are set] *** skipping: [fmov1n1.sn.dtcorp.com] skipping: [fmov1n2.sn.dtcorp.com] skipping: [fmov1n3.sn.dtcorp.com] TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports] ******** changed: [fmov1n3.sn.dtcorp.com] => (item=2049/tcp) changed: [fmov1n2.sn.dtcorp.com] => (item=2049/tcp) changed: [fmov1n1.sn.dtcorp.com] => (item=2049/tcp) changed: [fmov1n3.sn.dtcorp.com] => (item=54321/tcp) changed: [fmov1n2.sn.dtcorp.com] => (item=54321/tcp) changed: [fmov1n1.sn.dtcorp.com] => (item=54321/tcp) changed: [fmov1n3.sn.dtcorp.com] => (item=5900/tcp) changed: [fmov1n2.sn.dtcorp.com] => (item=5900/tcp) changed: [fmov1n1.sn.dtcorp.com] => (item=5900/tcp) changed: [fmov1n3.sn.dtcorp.com] => (item=5900-6923/tcp) changed: [fmov1n2.sn.dtcorp.com] => (item=5900-6923/tcp) changed: [fmov1n1.sn.dtcorp.com] => (item=5900-6923/tcp) changed: [fmov1n3.sn.dtcorp.com] => (item=5666/tcp) changed: [fmov1n2.sn.dtcorp.com] => (item=5666/tcp) changed: [fmov1n1.sn.dtcorp.com] => (item=5666/tcp) changed: [fmov1n3.sn.dtcorp.com] => (item=16514/tcp) changed: [fmov1n2.sn.dtcorp.com] => (item=16514/tcp) changed: [fmov1n1.sn.dtcorp.com] => (item=16514/tcp) TASK [gluster.infra/roles/firewall_config : Add/Delete services to firewalld rules] *** ok: [fmov1n3.sn.dtcorp.com] => (item=glusterfs) ok: [fmov1n2.sn.dtcorp.com] => (item=glusterfs) ok: [fmov1n1.sn.dtcorp.com] => (item=glusterfs) TASK [gluster.infra/roles/backend_setup : Check that the multipath.conf exists] *** ok: [fmov1n3.sn.dtcorp.com] ok: [fmov1n2.sn.dtcorp.com] ok: [fmov1n1.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is enabled if not] *** skipping: [fmov1n3.sn.dtcorp.com] changed: [fmov1n1.sn.dtcorp.com] changed: [fmov1n2.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is running] *** ok: [fmov1n3.sn.dtcorp.com] changed: [fmov1n1.sn.dtcorp.com] changed: [fmov1n2.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Create /etc/multipath/conf.d if doesn't exists] *** changed: [fmov1n3.sn.dtcorp.com] changed: [fmov1n2.sn.dtcorp.com] changed: [fmov1n1.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Get the UUID of the devices] ********* changed: [fmov1n3.sn.dtcorp.com] => (item=nvme0n1) changed: [fmov1n2.sn.dtcorp.com] => (item=nvme0n1) changed: [fmov1n1.sn.dtcorp.com] => (item=nvme0n1) changed: [fmov1n3.sn.dtcorp.com] => (item=nvme2n1) changed: [fmov1n2.sn.dtcorp.com] => (item=nvme2n1) changed: [fmov1n1.sn.dtcorp.com] => (item=nvme2n1) changed: [fmov1n3.sn.dtcorp.com] => (item=nvme1n1) changed: [fmov1n2.sn.dtcorp.com] => (item=nvme1n1) changed: [fmov1n1.sn.dtcorp.com] => (item=nvme1n1) TASK [gluster.infra/roles/backend_setup : Check that the blacklist.conf exists] *** ok: [fmov1n3.sn.dtcorp.com] ok: [fmov1n2.sn.dtcorp.com] ok: [fmov1n1.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Create blacklist template content] *** changed: [fmov1n3.sn.dtcorp.com] changed: [fmov1n2.sn.dtcorp.com] changed: [fmov1n1.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Add wwid to blacklist in blacklist.conf file] *** changed: [fmov1n3.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme0n1', 'stdout': "wwid 'eui.343756304d7020750025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:36.612051', 'end': '2020-07-14 21:06:36.623511', 'delta': '0:00:00.011460', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme0n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7020750025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme0n1', 'ansible_loop_var': 'item'}) changed: [fmov1n2.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme0n1', 'stdout': "wwid 'eui.343756304d7020530025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:36.674961', 'end': '2020-07-14 21:06:36.687875', 'delta': '0:00:00.012914', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme0n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7020530025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme0n1', 'ansible_loop_var': 'item'}) changed: [fmov1n1.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme0n1', 'stdout': "wwid 'eui.343756304d7020220025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:36.732721', 'end': '2020-07-14 21:06:36.744468', 'delta': '0:00:00.011747', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme0n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7020220025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme0n1', 'ansible_loop_var': 'item'}) changed: [fmov1n3.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme2n1', 'stdout': "wwid 'eui.343756304d7020730025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:41.411729', 'end': '2020-07-14 21:06:41.423305', 'delta': '0:00:00.011576', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme2n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7020730025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme2n1', 'ansible_loop_var': 'item'}) changed: [fmov1n2.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme2n1', 'stdout': "wwid 'eui.343756304d7020190025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:41.683414', 'end': '2020-07-14 21:06:41.695115', 'delta': '0:00:00.011701', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme2n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7020190025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme2n1', 'ansible_loop_var': 'item'}) changed: [fmov1n1.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme2n1', 'stdout': "wwid 'eui.343756304d7007630025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:41.832021', 'end': '2020-07-14 21:06:41.844162', 'delta': '0:00:00.012141', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme2n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7007630025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme2n1', 'ansible_loop_var': 'item'}) changed: [fmov1n3.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme1n1', 'stdout': "wwid 'eui.343756304d7020760025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:46.242072', 'end': '2020-07-14 21:06:46.253191', 'delta': '0:00:00.011119', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme1n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7020760025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme1n1', 'ansible_loop_var': 'item'}) changed: [fmov1n2.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme1n1', 'stdout': "wwid 'eui.343756304d7020690025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:46.697920', 'end': '2020-07-14 21:06:46.708944', 'delta': '0:00:00.011024', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme1n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7020690025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme1n1', 'ansible_loop_var': 'item'}) changed: [fmov1n1.sn.dtcorp.com] => (item={'cmd': 'multipath -a /dev/nvme1n1', 'stdout': "wwid 'eui.343756304d7020540025385800000004' added", 'stderr': '', 'rc': 0, 'start': '2020-07-14 21:06:46.860257', 'end': '2020-07-14 21:06:46.871208', 'delta': '0:00:00.010951', 'changed': True, 'invocation': {'module_args': {'_raw_params': 'multipath -a /dev/nvme1n1', '_uses_shell': True, 'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ["wwid 'eui.343756304d7020540025385800000004' added"], 'stderr_lines': [], 'failed': False, 'item': 'nvme1n1', 'ansible_loop_var': 'item'}) TASK [gluster.infra/roles/backend_setup : Reload multipathd] ******************* changed: [fmov1n3.sn.dtcorp.com] changed: [fmov1n2.sn.dtcorp.com] changed: [fmov1n1.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS distribution] *** ok: [fmov1n1.sn.dtcorp.com] ok: [fmov1n2.sn.dtcorp.com] ok: [fmov1n3.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for debian systems.] *** skipping: [fmov1n1.sn.dtcorp.com] skipping: [fmov1n2.sn.dtcorp.com] skipping: [fmov1n3.sn.dtcorp.com] TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.] *** fatal: [fmov1n3.sn.dtcorp.com]: FAILED! => {"changed": false, "msg": "The Python 2 yum module is needed for this module. If you require Python 3 support use the `dnf` Ansible module instead."} fatal: [fmov1n1.sn.dtcorp.com]: FAILED! => {"changed": false, "msg": "The Python 2 yum module is needed for this module. If you require Python 3 support use the `dnf` Ansible module instead."} fatal: [fmov1n2.sn.dtcorp.com]: FAILED! => {"changed": false, "msg": "The Python 2 yum module is needed for this module. If you require Python 3 support use the `dnf` Ansible module instead."} NO MORE HOSTS LEFT ************************************************************* NO MORE HOSTS LEFT ************************************************************* PLAY RECAP ********************************************************************* fmov1n1.sn.dtcorp.com : ok=16 changed=9 unreachable=0 failed=1 skipped=8 rescued=0 ignored=0 fmov1n2.sn.dtcorp.com : ok=15 changed=8 unreachable=0 failed=1 skipped=8 rescued=0 ignored=0 fmov1n3.sn.dtcorp.com : ok=14 changed=6 unreachable=0 failed=1 skipped=9 rescued=0 ignored=0 _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/VERCFGCDUOMSAV563MMCE4ZBH7LNJJVE/