Hey!
Are you running over CentOS?
Either you have to uncheck the "Blacklist gluster devices" option on the
bricks page and try again
Or you can add a filter to /etc/lvm/lvm.conf something like this
- a|^/dev/sda2$|",
On Mon, Nov 23, 2020 at 6:17 PM <rcpoling(a)gmail.com> wrote:
Trying to deploy a 3 Node Hyperconverged Ovirt Cluster with Gluster
as the
backend storage. I have tried this against the three nodes that I have as
well as with just a single node to get a working base line. The failure
that I keep getting stuck on is:
TASK [gluster.infra/roles/backend_setup : Create volume groups]
****************
task path:
/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vg_create.yml:59
failed: [ovirt01-storage.poling.local] (item={'key': 'gluster_vg_sdb',
'value': [{'vgname': 'gluster_vg_sdb', 'pvname':
'/dev/sdb'}]}) =>
{"ansible_loop_var": "item", "changed": false,
"err": " Device /dev/sdb
excluded by a filter.\n", "item": {"key":
"gluster_vg_sdb", "value":
[{"pvname": "/dev/sdb", "vgname":
"gluster_vg_sdb"}]}, "msg": "Creating
physical volume '/dev/sdb' failed", "rc": 5}
I have verified my dns records and have reverse dns set up. The Front End
Network and Storage Networks are physical separated and are 10GB
connections. In the reading i have done this seems to point to possibly
being a multipath issue, but i do see multipath configs being set in the
Gluster Wizard and when I check after the wizard fails out - it does look
like the mpath is set correctly.
[root@ovirt01 ~]# lsblk
NAME MAJ:MIN RM SIZE RO
TYPE MOUNTPOINT
sda 8:0 0 446.1G 0
disk
├─sda1 8:1 0 1G 0
part /boot
└─sda2 8:2 0 445.1G 0
part
├─onn-pool00_tmeta 253:0 0 1G 0 lvm
│ └─onn-pool00-tpool 253:2 0 351.7G 0 lvm
│ ├─onn-ovirt--node--ng--4.4.3--0.20201110.0+1 253:3 0 314.7G 0
lvm /
│ ├─onn-pool00 253:5 0 351.7G 1 lvm
│ ├─onn-var_log_audit 253:6 0 2G 0
lvm /var/log/audit
│ ├─onn-var_log 253:7 0 8G 0
lvm /var/log
│ ├─onn-var_crash 253:8 0 10G 0
lvm /var/crash
│ ├─onn-var 253:9 0 15G 0
lvm /var
│ ├─onn-tmp 253:10 0 1G 0
lvm /tmp
│ ├─onn-home 253:11 0 1G 0
lvm /home
│ └─onn-ovirt--node--ng--4.4.2--0.20200918.0+1 253:12 0 314.7G 0 lvm
├─onn-pool00_tdata 253:1 0 351.7G 0 lvm
│ └─onn-pool00-tpool 253:2 0 351.7G 0 lvm
│ ├─onn-ovirt--node--ng--4.4.3--0.20201110.0+1 253:3 0 314.7G 0
lvm /
│ ├─onn-pool00 253:5 0 351.7G 1 lvm
│ ├─onn-var_log_audit 253:6 0 2G 0
lvm /var/log/audit
│ ├─onn-var_log 253:7 0 8G 0
lvm /var/log
│ ├─onn-var_crash 253:8 0 10G 0
lvm /var/crash
│ ├─onn-var 253:9 0 15G 0
lvm /var
│ ├─onn-tmp 253:10 0 1G 0
lvm /tmp
│ ├─onn-home 253:11 0 1G 0
lvm /home
│ └─onn-ovirt--node--ng--4.4.2--0.20200918.0+1 253:12 0 314.7G 0 lvm
└─onn-swap 253:4 0 4G 0
lvm [SWAP]
sdb 8:16 0 5.5T 0
disk
└─sdb1 8:17 0 5.5T 0
part /sdb
Looking for any pointers on what else I should be looking at to get
gluster to deploy successfully. Thanks ~ R
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5DB2ENDLVT7...