I noted this as an independent bug.
Three separate issues here:
1) Wizard cannot take UUID names as names too long but does not note
character limit that would guide one not to do that :)
2) Wizard , as it should, allows you to assign devices independently for
each server, but does not retain this table on creation of ansible table.
3) Wizard, even when ansible file modified to reflect /dev/sd# for each
server, ignores this and runs against drive NOT called for in ansible file.
See attached Logs from where I took wizard output.. corrected first server
which is to use /dev/sdc vs servers 2,3 to use /dev/sdb and deployment
is against /dev/sdb on ALL THREE nodes.
# Ansible file used per bug where it defines drives incorrectly to be just
ID of first node
hc_nodes:
hosts:
thorst.penguinpages.local:
gluster_infra_volume_groups:
- vgname: gluster_vg_sdc
pvname: /dev/mapper/vdo_sdc
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_sdc
- path: /gluster_bricks/data
lvname: gluster_lv_data
vgname: gluster_vg_sdc
- path: /gluster_bricks/vmstore
lvname: gluster_lv_vmstore
vgname: gluster_vg_sdc
gluster_infra_vdo:
- name: vdo_sdc
device: /dev/sdc
slabsize: 32G
logicalsize: 11000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
blacklist_mpath_devices:
- sdc
gluster_infra_thick_lvs:
- vgname: gluster_vg_sdc
lvname: gluster_lv_engine
size: 1000G
gluster_infra_thinpools:
- vgname: gluster_vg_sdc
thinpoolname: gluster_thinpool_gluster_vg_sdc
poolmetadatasize: 3G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_sdc
thinpool: gluster_thinpool_gluster_vg_sdc
lvname: gluster_lv_data
lvsize: 5000G
- vgname: gluster_vg_sdc
thinpool: gluster_thinpool_gluster_vg_sdc
lvname: gluster_lv_vmstore
lvsize: 5000G
odinst.penguinpages.local:
gluster_infra_volume_groups:
- vgname: gluster_vg_sdb
pvname: /dev/mapper/vdo_sdb
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_sdb
- path: /gluster_bricks/data
lvname: gluster_lv_data
vgname: gluster_vg_sdb
- path: /gluster_bricks/vmstore
lvname: gluster_lv_vmstore
vgname: gluster_vg_sdb
gluster_infra_vdo:
- name: vdo_sdb
device: /dev/sdb
slabsize: 32G
logicalsize: 11000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
blacklist_mpath_devices:
- sdb
gluster_infra_thick_lvs:
- vgname: gluster_vg_sdb
lvname: gluster_lv_engine
size: 1000G
gluster_infra_thinpools:
- vgname: gluster_vg_sdb
thinpoolname: gluster_thinpool_gluster_vg_sdb
poolmetadatasize: 3G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_data
lvsize: 5000G
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_vmstore
lvsize: 5000G
medusast.penguinpages.local:
gluster_infra_volume_groups:
- vgname: gluster_vg_sdb
pvname: /dev/mapper/vdo_sdb
gluster_infra_mount_devices:
- path: /gluster_bricks/engine
lvname: gluster_lv_engine
vgname: gluster_vg_sdb
- path: /gluster_bricks/data
lvname: gluster_lv_data
vgname: gluster_vg_sdb
- path: /gluster_bricks/vmstore
lvname: gluster_lv_vmstore
vgname: gluster_vg_sdb
gluster_infra_vdo:
- name: vdo_sdb
device: /dev/sdb
slabsize: 32G
logicalsize: 11000G
blockmapcachesize: 128M
emulate512: 'off'
writepolicy: auto
maxDiscardSize: 16M
blacklist_mpath_devices:
- sdb
gluster_infra_thick_lvs:
- vgname: gluster_vg_sdb
lvname: gluster_lv_engine
size: 1000G
gluster_infra_thinpools:
- vgname: gluster_vg_sdb
thinpoolname: gluster_thinpool_gluster_vg_sdb
poolmetadatasize: 3G
gluster_infra_lv_logicalvols:
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_data
lvsize: 5000G
- vgname: gluster_vg_sdb
thinpool: gluster_thinpool_gluster_vg_sdb
lvname: gluster_lv_vmstore
lvsize: 5000G
vars:
gluster_infra_disktype: JBOD
gluster_set_selinux_labels: true
gluster_infra_fw_ports:
- 2049/tcp
- 54321/tcp
- 5900/tcp
- 5900-6923/tcp
- 5666/tcp
- 16514/tcp
gluster_infra_fw_permanent: true
gluster_infra_fw_state: enabled
gluster_infra_fw_zone: public
gluster_infra_fw_services:
- glusterfs
gluster_features_force_varlogsizecheck: false
cluster_nodes:
- thorst.penguinpages.local
- odinst.penguinpages.local
- medusast.penguinpages.local
gluster_features_hci_cluster: '{{ cluster_nodes }}'
gluster_features_hci_volumes:
- volname: engine
brick: /gluster_bricks/engine/engine
arbiter: 0
- volname: data
brick: /gluster_bricks/data/data
arbiter: 0
- volname: vmstore
brick: /gluster_bricks/vmstore/vmstore
arbiter: 0
I have three drives in each server. The first is a 512GB SSD and this is
how I want to do initial HCI setup. The other six are bricks with existing
data VDO volumes from old Glutster volume. I wanted to import those. But
lafter a few attempts it became apparent the HCI wizard just can't figure
out how to stop VDO / Create partitions etc... and ignore drives not within
its script. I have backups so I wiped all the data back to nakid
drives..well.. then had other issues with CentOS that I am working out...
so delay until I root cause that Hmmmm.. issue.
Thanks for input/help
On Sun, Sep 13, 2020 at 4:24 AM Parth Dhanjal <dparth(a)redhat.com> wrote:
Hey!
I am guessing that the deployment is failing due to the long lv name.
Can you try to edit the ansible-playbook and shorten the lv name to 55
characters and try again?
Regards
Parth Dhanjal
On Sun, Sep 13, 2020 at 11:16 AM Jeremey Wise <jeremey.wise(a)gmail.com>
wrote:
>
> Greetings:
>
> 3 x servers each server has 1 x 512GB ssd 2 x 1TB SSD JBOD
>
> Goal: use HCI disk setup wizard to deploy initial structure
>
> Each server has disk scanning in as different /dev/sd# and so trying to
> use more clear /dev/mapper/<disk ID>
>
> As such I set this per table below:
>
>
>
> # Select each server and set each drive
<<<<<<<<<<Double Check drive
> device ID as they do NOT match up per host
>
> # I transitioned to /dev/mapper object to avoid unclearness of /dev/sd#
>
> thor
>
> /dev/sdc
>
> /dev/mapper/Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
>
> odin
>
> /dev/sdb
>
> /dev/mapper/Micron_1100_MTFDDAV512TBN_17401F699137
>
> medusa
>
> /dev/sdb
>
> /dev/mapper/SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306
>
> # Note that drives need to be completely clear of any partition or file
> system
>
> [root@thor /]# gdisk /dev/sdc
>
> GPT fdisk (gdisk) version 1.0.3
>
>
>
> Partition table scan:
>
> MBR: protective
>
> BSD: not present
>
> APM: not present
>
> GPT: present
>
>
>
> Found valid GPT with protective MBR; using GPT.
>
>
>
> Command (? for help): x
>
>
>
> Expert command (? for help): z
>
> About to wipe out GPT on /dev/sdc. Proceed? (Y/N): y
>
> GPT data structures destroyed! You may now partition the disk using fdisk
> or
>
> other utilities.
>
> Blank out MBR? (Y/N): y
> [image: image.png]
>
>
>
>
> [image: image.png]
>
>
> But deployment fails with Error:
> <snip>
>
> TASK [gluster.infra/roles/backend_setup : Filter none-existing devices]
> ********
> task path:
> /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/get_vg_groupings.yml:38
> ok: [thorst.penguinpages.local] => {"ansible_facts":
> {"gluster_volumes_by_groupname": {}}, "changed": false}
> ok: [odinst.penguinpages.local] => {"ansible_facts":
> {"gluster_volumes_by_groupname": {}}, "changed": false}
> ok: [medusast.penguinpages.local] => {"ansible_facts":
> {"gluster_volumes_by_groupname": {}}, "changed": false}
>
> TASK [gluster.infra/roles/backend_setup : Make sure thick pvs exists in
> volume group] ***
> task path:
> /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thick_lv_create.yml:37
>
> TASK [gluster.infra/roles/backend_setup : update LVM fact's]
> *******************
> task path:
> /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thick_lv_create.yml:83
> skipping: [thorst.penguinpages.local] => {"changed": false,
> "skip_reason": "Conditional result was False"}
> skipping: [odinst.penguinpages.local] => {"changed": false,
> "skip_reason": "Conditional result was False"}
> skipping: [medusast.penguinpages.local] => {"changed": false,
> "skip_reason": "Conditional result was False"}
>
> TASK [gluster.infra/roles/backend_setup : Create thick logical volume]
> *********
> task path:
> /etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thick_lv_create.yml:90
> failed: [medusast.penguinpages.local] (item={'vgname': '
> gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L', 'lvname':
> 'gluster_lv_engine', 'size': '1000G'}) =>
{"ansible_index_var": "index",
> "ansible_loop_var": "item", "changed": false,
"err": " Volume group
> \"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L\" not found.\n
> Cannot process volume group
> gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L\n", "index": 0,
> "item": {"lvname": "gluster_lv_engine",
"size": "1000G", "vgname":
> "gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L"}, "msg":
"Volume
> group gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L does not
> exist.", "rc": 5}
> changed: [thorst.penguinpages.local] => (item={'vgname': '
> gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L', 'lvname':
> 'gluster_lv_engine', 'size': '1000G'}) =>
{"ansible_index_var": "index",
> "ansible_loop_var": "item", "changed": true,
"index": 0, "item": {"lvname":
> "gluster_lv_engine", "size": "1000G",
"vgname":
> "gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L"}, "msg":
""}
> failed: [odinst.penguinpages.local] (item={'vgname':
> 'gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L', 'lvname':
> 'gluster_lv_engine', 'size': '1000G'}) =>
{"ansible_index_var": "index",
> "ansible_loop_var": "item", "changed": false,
"err": " Volume group
> \"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L\" not found.\n
> Cannot process volume group
> gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L\n", "index": 0,
> "item": {"lvname": "gluster_lv_engine",
"size": "1000G", "vgname":
> "gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L"}, "msg":
"Volume
> group gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L does not
> exist.", "rc": 5}
>
> NO MORE HOSTS LEFT
> *************************************************************
>
> NO MORE HOSTS LEFT
> *************************************************************
>
> PLAY RECAP
> *********************************************************************
> medusast.penguinpages.local : ok=23 changed=5 unreachable=0
> failed=1 skipped=34 rescued=0 ignored=0
> odinst.penguinpages.local : ok=23 changed=5 unreachable=0
> failed=1 skipped=34 rescued=0 ignored=0
> thorst.penguinpages.local : ok=30 changed=9 unreachable=0
> failed=0 skipped=29 rescued=0 ignored=0
>
> Please check /var/log/cockpit/ovirt-dashboard/gluster-deployment.log for
> more informations.
>
> ############
>
> Why is oVirt ignoring when I set (and double check explicite device call
> for deployment?
>
> Attached is the ansible file it creates and then one I had to edit to
> correct what wizard should have built it as.
>
>
>
> --
> p <jeremey.wise(a)gmail.com>enguinpages
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>