Thank you Strahil. I think I edited the oVirt Node Cockpit Hyperconverged Wizard Gluster
Deployment Ansible playbook as detailed in your post and received the following new
failure:
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.]
***
fatal: [
fmov1n3.sn.dtcorp.com]: FAILED! => {"changed": false,
"msg": "The Python 2 yum module is needed for this module. If you require
Python 3 support use the `dnf` Ansible module instead."}
fatal: [
fmov1n1.sn.dtcorp.com]: FAILED! => {"changed": false,
"msg": "The Python 2 yum module is needed for this module. If you require
Python 3 support use the `dnf` Ansible module instead."}
fatal: [
fmov1n2.sn.dtcorp.com]: FAILED! => {"changed": false,
"msg": "The Python 2 yum module is needed for this module. If you require
Python 3 support use the `dnf` Ansible module instead."}
Any further assistance is most appreciated!!!
Respectfully,
Charles
---
Gluster Deployment Ansible Playbook
hc_nodes:
hosts:
fmov1n1.sn.dtcorp.com:
gluster_infra_volume_groups:
- vgname: gluster_vg_nvme0n1
pvname: /dev/mapper/vdo_nvme0n1
- vgname: gluster_vg_nvme2n1
pvname: /dev/mapper/vdo_nvme2n1
- vgname: gluster_vg_nvme1n1
pvname: /dev/mapper/vdo_nvme1n1
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_nvme0n1
- path: /gluster_bricks/data
lvname: gluster_lv_data
vgname: gluster_vg_nvme2n1
- path: /gluster_bricks/vmstore
lvname: gluster_lv_vmstore
vgname: gluster_vg_nvme1n1
gluster_infra_vdo:
- name: vdo_nvme0n1
device: /dev/nvme0n1
slabsize: 2G
logicalsize: 1000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
- name: vdo_nvme2n1
device: /dev/nvme2n1
slabsize: 32G
logicalsize: 5000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
- name: vdo_nvme1n1
device: /dev/nvme1n1
slabsize: 32G
logicalsize: 5000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
blacklist_mpath_devices:
- nvme0n1
- nvme2n1
- nvme1n1
gluster_infra_thick_lvs:
- vgname: gluster_vg_nvme0n1
lvname: gluster_lv_engine
size: '100%PVS'
gluster_infra_thinpools:
- vgname: gluster_vg_nvme2n1
thinpoolname: gluster_thinpool_gluster_vg_nvme2n1
poolmetadatasize: 3G
- vgname: gluster_vg_nvme1n1
thinpoolname: gluster_thinpool_gluster_vg_nvme1n1
poolmetadatasize: 3G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_nvme2n1
thinpool: gluster_thinpool_gluster_vg_nvme2n1
lvname: gluster_lv_data
lvsize: '100%PVS'
- vgname: gluster_vg_nvme1n1
thinpool: gluster_thinpool_gluster_vg_nvme1n1
lvname: gluster_lv_vmstore
lvsize: '100%PVS'
fmov1n2.sn.dtcorp.com:
gluster_infra_volume_groups:
- vgname: gluster_vg_nvme0n1
pvname: /dev/mapper/vdo_nvme0n1
- vgname: gluster_vg_nvme2n1
pvname: /dev/mapper/vdo_nvme2n1
- vgname: gluster_vg_nvme1n1
pvname: /dev/mapper/vdo_nvme1n1
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_nvme0n1
- path: /gluster_bricks/data
lvname: gluster_lv_data
vgname: gluster_vg_nvme2n1
- path: /gluster_bricks/vmstore
lvname: gluster_lv_vmstore
vgname: gluster_vg_nvme1n1
gluster_infra_vdo:
- name: vdo_nvme0n1
device: /dev/nvme0n1
slabsize: 2G
logicalsize: 1000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
- name: vdo_nvme2n1
device: /dev/nvme2n1
slabsize: 32G
logicalsize: 5000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
- name: vdo_nvme1n1
device: /dev/nvme1n1
slabsize: 32G
logicalsize: 5000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
blacklist_mpath_devices:
- nvme0n1
- nvme2n1
- nvme1n1
gluster_infra_thick_lvs:
- vgname: gluster_vg_nvme0n1
lvname: gluster_lv_engine
size: '100%PVS'
gluster_infra_thinpools:
- vgname: gluster_vg_nvme2n1
thinpoolname: gluster_thinpool_gluster_vg_nvme2n1
poolmetadatasize: 3G
- vgname: gluster_vg_nvme1n1
thinpoolname: gluster_thinpool_gluster_vg_nvme1n1
poolmetadatasize: 3G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_nvme2n1
thinpool: gluster_thinpool_gluster_vg_nvme2n1
lvname: gluster_lv_data
lvsize: '100%PVS'
- vgname: gluster_vg_nvme1n1
thinpool: gluster_thinpool_gluster_vg_nvme1n1
lvname: gluster_lv_vmstore
lvsize: '100%PVS'
fmov1n3.sn.dtcorp.com:
gluster_infra_volume_groups:
- vgname: gluster_vg_nvme0n1
pvname: /dev/mapper/vdo_nvme0n1
- vgname: gluster_vg_nvme2n1
pvname: /dev/mapper/vdo_nvme2n1
- vgname: gluster_vg_nvme1n1
pvname: /dev/mapper/vdo_nvme1n1
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_nvme0n1
- path: /gluster_bricks/data
lvname: gluster_lv_data
vgname: gluster_vg_nvme2n1
- path: /gluster_bricks/vmstore
lvname: gluster_lv_vmstore
vgname: gluster_vg_nvme1n1
gluster_infra_vdo:
- name: vdo_nvme0n1
device: /dev/nvme0n1
slabsize: 2G
logicalsize: 1000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
- name: vdo_nvme2n1
device: /dev/nvme2n1
slabsize: 32G
logicalsize: 5000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
- name: vdo_nvme1n1
device: /dev/nvme1n1
slabsize: 32G
logicalsize: 5000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
blacklist_mpath_devices:
- nvme0n1
- nvme2n1
- nvme1n1
gluster_infra_thick_lvs:
- vgname: gluster_vg_nvme0n1
lvname: gluster_lv_engine
size: '100%PVS'
gluster_infra_thinpools:
- vgname: gluster_vg_nvme2n1
thinpoolname: gluster_thinpool_gluster_vg_nvme2n1
poolmetadatasize: 3G
- vgname: gluster_vg_nvme1n1
thinpoolname: gluster_thinpool_gluster_vg_nvme1n1
poolmetadatasize: 3G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_nvme2n1
thinpool: gluster_thinpool_gluster_vg_nvme2n1
lvname: gluster_lv_data
lvsize: '100%PVS'
- vgname: gluster_vg_nvme1n1
thinpool: gluster_thinpool_gluster_vg_nvme1n1
lvname: gluster_lv_vmstore
lvsize: '100%PVS'
vars:
gluster_infra_disktype: JBOD
gluster_set_selinux_labels: true
gluster_infra_fw_ports:
- 2049/tcp
- 54321/tcp
- 5900/tcp
- 5900-6923/tcp
- 5666/tcp
- 16514/tcp
gluster_infra_fw_permanent: true
gluster_infra_fw_state: enabled
gluster_infra_fw_zone: public
gluster_infra_fw_services:
- glusterfs
gluster_features_force_varlogsizecheck: false
cluster_nodes:
-
fmov1n1.sn.dtcorp.com
-
fmov1n2.sn.dtcorp.com
-
fmov1n3.sn.dtcorp.com
gluster_features_hci_cluster: '{{ cluster_nodes }}'
gluster_features_hci_volumes:
- volname: engine
brick: /gluster_bricks/engine/engine
arbiter: 0
- volname: data
brick: /gluster_bricks/data/data
arbiter: 0
- volname: vmstore
brick: /gluster_bricks/vmstore/vmstore
arbiter: 0
---
/var/log/cockpit/ovirt-dashboard/gluster-deployment.log
PLAY [Setup backend] ***********************************************************
TASK [Gathering Facts] *********************************************************
ok: [
fmov1n1.sn.dtcorp.com]
ok: [
fmov1n2.sn.dtcorp.com]
ok: [
fmov1n3.sn.dtcorp.com]
TASK [Check if valid hostnames are provided] ***********************************
changed: [
fmov1n1.sn.dtcorp.com] => (
item=fmov1n1.sn.dtcorp.com)
changed: [
fmov1n1.sn.dtcorp.com] => (
item=fmov1n2.sn.dtcorp.com)
changed: [
fmov1n1.sn.dtcorp.com] => (
item=fmov1n3.sn.dtcorp.com)
TASK [Check if provided hostnames are valid] ***********************************
ok: [
fmov1n1.sn.dtcorp.com] => {
"changed": false,
"msg": "All assertions passed"
}
ok: [
fmov1n2.sn.dtcorp.com] => {
"changed": false,
"msg": "All assertions passed"
}
ok: [
fmov1n3.sn.dtcorp.com] => {
"changed": false,
"msg": "All assertions passed"
}
TASK [Check if /var/log has enough disk space] *********************************
skipping: [
fmov1n1.sn.dtcorp.com]
skipping: [
fmov1n2.sn.dtcorp.com]
skipping: [
fmov1n3.sn.dtcorp.com]
TASK [Check if the /var is greater than 15G] ***********************************
skipping: [
fmov1n1.sn.dtcorp.com]
skipping: [
fmov1n2.sn.dtcorp.com]
skipping: [
fmov1n3.sn.dtcorp.com]
TASK [Check if disks have logical block size of 512B] **************************
skipping: [
fmov1n1.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'})
skipping: [
fmov1n1.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'})
skipping: [
fmov1n1.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'})
skipping: [
fmov1n2.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'})
skipping: [
fmov1n2.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'})
skipping: [
fmov1n2.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'})
skipping: [
fmov1n3.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme0n1', 'pvname': '/dev/mapper/vdo_nvme0n1'})
skipping: [
fmov1n3.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme2n1', 'pvname': '/dev/mapper/vdo_nvme2n1'})
skipping: [
fmov1n3.sn.dtcorp.com] => (item={'vgname':
'gluster_vg_nvme1n1', 'pvname': '/dev/mapper/vdo_nvme1n1'})
TASK [Check if logical block size is 512 bytes] ********************************
skipping: [
fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n2.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n2.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n2.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n3.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n3.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n3.sn.dtcorp.com] => (item=Logical Block Size)
TASK [Get logical block size of VDO devices] ***********************************
skipping: [
fmov1n1.sn.dtcorp.com] => (item={'name': 'vdo_nvme0n1',
'device': '/dev/nvme0n1', 'slabsize': '2G',
'logicalsize': '1000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
skipping: [
fmov1n1.sn.dtcorp.com] => (item={'name': 'vdo_nvme2n1',
'device': '/dev/nvme2n1', 'slabsize': '32G',
'logicalsize': '5000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
skipping: [
fmov1n1.sn.dtcorp.com] => (item={'name': 'vdo_nvme1n1',
'device': '/dev/nvme1n1', 'slabsize': '32G',
'logicalsize': '5000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
skipping: [
fmov1n2.sn.dtcorp.com] => (item={'name': 'vdo_nvme0n1',
'device': '/dev/nvme0n1', 'slabsize': '2G',
'logicalsize': '1000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
skipping: [
fmov1n2.sn.dtcorp.com] => (item={'name': 'vdo_nvme2n1',
'device': '/dev/nvme2n1', 'slabsize': '32G',
'logicalsize': '5000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
skipping: [
fmov1n2.sn.dtcorp.com] => (item={'name': 'vdo_nvme1n1',
'device': '/dev/nvme1n1', 'slabsize': '32G',
'logicalsize': '5000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
skipping: [
fmov1n3.sn.dtcorp.com] => (item={'name': 'vdo_nvme0n1',
'device': '/dev/nvme0n1', 'slabsize': '2G',
'logicalsize': '1000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
skipping: [
fmov1n3.sn.dtcorp.com] => (item={'name': 'vdo_nvme2n1',
'device': '/dev/nvme2n1', 'slabsize': '32G',
'logicalsize': '5000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
skipping: [
fmov1n3.sn.dtcorp.com] => (item={'name': 'vdo_nvme1n1',
'device': '/dev/nvme1n1', 'slabsize': '32G',
'logicalsize': '5000G', 'blockmapcachesize': '128M',
'emulate512': 'off', 'writepolicy': 'auto',
'maxDiscardSize': '16M'})
TASK [Check if logical block size is 512 bytes for VDO devices] ****************
skipping: [
fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n1.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n2.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n2.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n2.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n3.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n3.sn.dtcorp.com] => (item=Logical Block Size)
skipping: [
fmov1n3.sn.dtcorp.com] => (item=Logical Block Size)
TASK [gluster.infra/roles/firewall_config : Start firewalld if not already started] ***
ok: [
fmov1n3.sn.dtcorp.com]
ok: [
fmov1n2.sn.dtcorp.com]
ok: [
fmov1n1.sn.dtcorp.com]
TASK [gluster.infra/roles/firewall_config : check if required variables are set] ***
skipping: [
fmov1n1.sn.dtcorp.com]
skipping: [
fmov1n2.sn.dtcorp.com]
skipping: [
fmov1n3.sn.dtcorp.com]
TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports] ********
changed: [
fmov1n3.sn.dtcorp.com] => (item=2049/tcp)
changed: [
fmov1n2.sn.dtcorp.com] => (item=2049/tcp)
changed: [
fmov1n1.sn.dtcorp.com] => (item=2049/tcp)
changed: [
fmov1n3.sn.dtcorp.com] => (item=54321/tcp)
changed: [
fmov1n2.sn.dtcorp.com] => (item=54321/tcp)
changed: [
fmov1n1.sn.dtcorp.com] => (item=54321/tcp)
changed: [
fmov1n3.sn.dtcorp.com] => (item=5900/tcp)
changed: [
fmov1n2.sn.dtcorp.com] => (item=5900/tcp)
changed: [
fmov1n1.sn.dtcorp.com] => (item=5900/tcp)
changed: [
fmov1n3.sn.dtcorp.com] => (item=5900-6923/tcp)
changed: [
fmov1n2.sn.dtcorp.com] => (item=5900-6923/tcp)
changed: [
fmov1n1.sn.dtcorp.com] => (item=5900-6923/tcp)
changed: [
fmov1n3.sn.dtcorp.com] => (item=5666/tcp)
changed: [
fmov1n2.sn.dtcorp.com] => (item=5666/tcp)
changed: [
fmov1n1.sn.dtcorp.com] => (item=5666/tcp)
changed: [
fmov1n3.sn.dtcorp.com] => (item=16514/tcp)
changed: [
fmov1n2.sn.dtcorp.com] => (item=16514/tcp)
changed: [
fmov1n1.sn.dtcorp.com] => (item=16514/tcp)
TASK [gluster.infra/roles/firewall_config : Add/Delete services to firewalld rules] ***
ok: [
fmov1n3.sn.dtcorp.com] => (item=glusterfs)
ok: [
fmov1n2.sn.dtcorp.com] => (item=glusterfs)
ok: [
fmov1n1.sn.dtcorp.com] => (item=glusterfs)
TASK [gluster.infra/roles/backend_setup : Check that the multipath.conf exists] ***
ok: [
fmov1n3.sn.dtcorp.com]
ok: [
fmov1n2.sn.dtcorp.com]
ok: [
fmov1n1.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is enabled if
not] ***
skipping: [
fmov1n3.sn.dtcorp.com]
changed: [
fmov1n1.sn.dtcorp.com]
changed: [
fmov1n2.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Ensure that multipathd services is running] ***
ok: [
fmov1n3.sn.dtcorp.com]
changed: [
fmov1n1.sn.dtcorp.com]
changed: [
fmov1n2.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Create /etc/multipath/conf.d if doesn't
exists] ***
changed: [
fmov1n3.sn.dtcorp.com]
changed: [
fmov1n2.sn.dtcorp.com]
changed: [
fmov1n1.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Get the UUID of the devices] *********
changed: [
fmov1n3.sn.dtcorp.com] => (item=nvme0n1)
changed: [
fmov1n2.sn.dtcorp.com] => (item=nvme0n1)
changed: [
fmov1n1.sn.dtcorp.com] => (item=nvme0n1)
changed: [
fmov1n3.sn.dtcorp.com] => (item=nvme2n1)
changed: [
fmov1n2.sn.dtcorp.com] => (item=nvme2n1)
changed: [
fmov1n1.sn.dtcorp.com] => (item=nvme2n1)
changed: [
fmov1n3.sn.dtcorp.com] => (item=nvme1n1)
changed: [
fmov1n2.sn.dtcorp.com] => (item=nvme1n1)
changed: [
fmov1n1.sn.dtcorp.com] => (item=nvme1n1)
TASK [gluster.infra/roles/backend_setup : Check that the blacklist.conf exists] ***
ok: [
fmov1n3.sn.dtcorp.com]
ok: [
fmov1n2.sn.dtcorp.com]
ok: [
fmov1n1.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Create blacklist template content] ***
changed: [
fmov1n3.sn.dtcorp.com]
changed: [
fmov1n2.sn.dtcorp.com]
changed: [
fmov1n1.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Add wwid to blacklist in blacklist.conf file]
***
changed: [
fmov1n3.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme0n1', 'stdout': "wwid
'eui.343756304d7020750025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:36.612051', 'end':
'2020-07-14 21:06:36.623511', 'delta': '0:00:00.011460',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme0n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7020750025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme0n1', 'ansible_loop_var': 'item'})
changed: [
fmov1n2.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme0n1', 'stdout': "wwid
'eui.343756304d7020530025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:36.674961', 'end':
'2020-07-14 21:06:36.687875', 'delta': '0:00:00.012914',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme0n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7020530025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme0n1', 'ansible_loop_var': 'item'})
changed: [
fmov1n1.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme0n1', 'stdout': "wwid
'eui.343756304d7020220025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:36.732721', 'end':
'2020-07-14 21:06:36.744468', 'delta': '0:00:00.011747',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme0n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7020220025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme0n1', 'ansible_loop_var': 'item'})
changed: [
fmov1n3.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme2n1', 'stdout': "wwid
'eui.343756304d7020730025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:41.411729', 'end':
'2020-07-14 21:06:41.423305', 'delta': '0:00:00.011576',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme2n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7020730025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme2n1', 'ansible_loop_var': 'item'})
changed: [
fmov1n2.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme2n1', 'stdout': "wwid
'eui.343756304d7020190025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:41.683414', 'end':
'2020-07-14 21:06:41.695115', 'delta': '0:00:00.011701',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme2n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7020190025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme2n1', 'ansible_loop_var': 'item'})
changed: [
fmov1n1.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme2n1', 'stdout': "wwid
'eui.343756304d7007630025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:41.832021', 'end':
'2020-07-14 21:06:41.844162', 'delta': '0:00:00.012141',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme2n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7007630025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme2n1', 'ansible_loop_var': 'item'})
changed: [
fmov1n3.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme1n1', 'stdout': "wwid
'eui.343756304d7020760025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:46.242072', 'end':
'2020-07-14 21:06:46.253191', 'delta': '0:00:00.011119',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme1n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7020760025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme1n1', 'ansible_loop_var': 'item'})
changed: [
fmov1n2.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme1n1', 'stdout': "wwid
'eui.343756304d7020690025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:46.697920', 'end':
'2020-07-14 21:06:46.708944', 'delta': '0:00:00.011024',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme1n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7020690025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme1n1', 'ansible_loop_var': 'item'})
changed: [
fmov1n1.sn.dtcorp.com] => (item={'cmd': 'multipath -a
/dev/nvme1n1', 'stdout': "wwid
'eui.343756304d7020540025385800000004' added", 'stderr': '',
'rc': 0, 'start': '2020-07-14 21:06:46.860257', 'end':
'2020-07-14 21:06:46.871208', 'delta': '0:00:00.010951',
'changed': True, 'invocation': {'module_args':
{'_raw_params': 'multipath -a /dev/nvme1n1', '_uses_shell': True,
'warn': True, 'stdin_add_newline': True, 'strip_empty_ends': True,
'argv': None, 'chdir': None, 'executable': None,
'creates': None, 'removes': None, 'stdin': None}},
'stdout_lines': ["wwid 'eui.343756304d7020540025385800000004'
added"], 'stderr_lines': [], 'failed': False, 'item':
'nvme1n1', 'ansible_loop_var': 'item'})
TASK [gluster.infra/roles/backend_setup : Reload multipathd] *******************
changed: [
fmov1n3.sn.dtcorp.com]
changed: [
fmov1n2.sn.dtcorp.com]
changed: [
fmov1n1.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS distribution]
***
ok: [
fmov1n1.sn.dtcorp.com]
ok: [
fmov1n2.sn.dtcorp.com]
ok: [
fmov1n3.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for debian systems.]
***
skipping: [
fmov1n1.sn.dtcorp.com]
skipping: [
fmov1n2.sn.dtcorp.com]
skipping: [
fmov1n3.sn.dtcorp.com]
TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.]
***
fatal: [
fmov1n3.sn.dtcorp.com]: FAILED! => {"changed": false,
"msg": "The Python 2 yum module is needed for this module. If you require
Python 3 support use the `dnf` Ansible module instead."}
fatal: [
fmov1n1.sn.dtcorp.com]: FAILED! => {"changed": false,
"msg": "The Python 2 yum module is needed for this module. If you require
Python 3 support use the `dnf` Ansible module instead."}
fatal: [
fmov1n2.sn.dtcorp.com]: FAILED! => {"changed": false,
"msg": "The Python 2 yum module is needed for this module. If you require
Python 3 support use the `dnf` Ansible module instead."}
NO MORE HOSTS LEFT *************************************************************
NO MORE HOSTS LEFT *************************************************************
PLAY RECAP *********************************************************************
fmov1n1.sn.dtcorp.com : ok=16 changed=9 unreachable=0 failed=1 skipped=8
rescued=0 ignored=0
fmov1n2.sn.dtcorp.com : ok=15 changed=8 unreachable=0 failed=1 skipped=8
rescued=0 ignored=0
fmov1n3.sn.dtcorp.com : ok=14 changed=6 unreachable=0 failed=1 skipped=9
rescued=0 ignored=0