
Sac, *To answer some of your questions:* *fdisk -l:* [root@host1 ~]# fdisk -l /dev/sdb Disk /dev/sde: 480.1 GB, 480070426624 bytes, 937637552 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 262144 bytes / 262144 bytes [root@host1 ~]# fdisk -l /dev/sdc Disk /dev/sdc: 3000.6 GB, 3000559427584 bytes, 5860467632 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 262144 bytes / 262144 bytes [root@host1 ~]# fdisk -l /dev/sdd Disk /dev/sdd: 3000.6 GB, 3000559427584 bytes, 5860467632 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 262144 bytes / 262144 bytes *1) i did wipefs to all /dev/sdb,c,d,e* *2) I did not zero out the disks as I had done it thru the controller.* *3) cat /proc/partitions:* [root@host1 ~]# cat /proc/partitions major minor #blocks name 8 0 586029016 sda 8 1 1048576 sda1 8 2 584978432 sda2 8 16 2930233816 sdb 8 32 2930233816 sdc 8 48 2930233816 sdd 8 64 468818776 sde *4) grep filter /etc/lvm/lvm.conf (I did not modify the lvm.conf file)* [root@host1 ~]# grep "filter =" /etc/lvm/lvm.conf # filter = [ "a|.*/|" ] # filter = [ "r|/dev/cdrom|" ] # filter = [ "a|loop|", "r|.*|" ] # filter = [ "a|loop|", "r|/dev/hdc|", "a|/dev/ide|", "r|.*|" ] # filter = [ "a|^/dev/hda8$|", "r|.*/|" ] # filter = [ "a|.*/|" ] # global_filter = [ "a|.*/|" ] # mlock_filter = [ "locale/locale-archive", "gconv/gconv-modules.cache" ] *What I did to get it working:* I re-installed my first 3 hosts using "ovirt-node-ng-installer-4.3.3-2019041712.el7.iso"and made sure I zeroed the disks from within the controller, then I performed the following steps: 1.- modifed the blacklist section on /etc/multipath.conf to this: blacklist { # protocol "(scsi:adt|scsi:sbp)" devnode "*" } 2.-Made sure the second line of /etc/multipath.conf has: # VDSM PRIVATE 3.-Increased /var/log to 15GB 4.-Rebuilt initramfs, rebooted 5.-wipefs -a /dev/sdb /dev/sdc /dev/sdd /dev/sde 6.-started the hyperconverged setup wizard and added* "gluster_features_force_varlogsizecheck: false"* to the "vars:" section on the Generated Ansible inventory : */etc/ansible/hc_wizard_inventory.yml* file as it was complaining about /var/log messages LV. *EUREKA: *After doing the above I was able to get past the filter issues, however I am still concerned if during a reboot the disks might come up differently. For example /dev/sdb might come up as /dev/sdx... I am trying to make sure this setup is always the same as we want to move this to production, however seems I still don't have the full hang of it and the RHV 4.1 course is way to old :) Thanks again for helping out with this. -AQ On Tue, May 21, 2019 at 3:29 AM Sachidananda URS <surs@redhat.com> wrote:
On Tue, May 21, 2019 at 12:16 PM Sahina Bose <sabose@redhat.com> wrote:
On Mon, May 20, 2019 at 9:55 PM Adrian Quintero <adrianquintero@gmail.com> wrote:
Sahina, Yesterday I started with a fresh install, I completely wiped clean all the disks, recreated the arrays from within my controller of our DL380 Gen 9's.
OS: RAID 1 (2x600GB HDDs): /dev/sda // Using ovirt node 4.3.3.1 iso. engine and VMSTORE1: JBOD (1x3TB HDD):/dev/sdb DATA1: JBOD (1x3TB HDD): /dev/sdc DATA2: JBOD (1x3TB HDD): /dev/sdd Caching disk: JOBD (1x440GB SDD): /dev/sde
*After the OS install on the first 3 servers and setting up ssh keys, I started the Hyperconverged deploy process:* 1.-Logged int to the first server http://host1.example.com:9090 2.-Selected Hyperconverged, clicked on "Run Gluster Wizard" 3.-Followed the wizard steps (Hosts, FQDNs, Packages, Volumes, Bricks, Review) *Hosts/FQDNs:* host1.example.com host2.example.com host3.example.com *Packages:* *Volumes:* engine:replicate:/gluster_bricks/engine/engine vmstore1:replicate:/gluster_bricks/vmstore1/vmstore1 data1:replicate:/gluster_bricks/data1/data1 data2:replicate:/gluster_bricks/data2/data2 *Bricks:* engine:/dev/sdb:100GB:/gluster_bricks/engine vmstore1:/dev/sdb:2600GB:/gluster_bricks/vmstrore1 data1:/dev/sdc:2700GB:/gluster_bricks/data1 data2:/dev/sdd:2700GB:/gluster_bricks/data2 LV Cache: /dev/sde:400GB:writethrough 4.-After I hit deploy on the last step of the "Wizard" that is when I get the disk filter error. TASK [gluster.infra/roles/backend_setup : Create volume groups] **************** failed: [vmm10.virt.iad3p] (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) => {"changed": false, "err": " Device /dev/sdb excluded by a filter.\n", "item": {"pvname": "/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg": "Creating physical volume '/dev/sdb' failed", "rc": 5} failed: [vmm12.virt.iad3p] (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) => {"changed": false, "err": " Device /dev/sdb excluded by a filter.\n", "item": {"pvname": "/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg": "Creating physical volume '/dev/sdb' failed", "rc": 5} failed: [vmm11.virt.iad3p] (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) => {"changed": false, "err": " Device /dev/sdb excluded by a filter.\n", "item": {"pvname": "/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg": "Creating physical volume '/dev/sdb' failed", "rc": 5} failed: [vmm12.virt.iad3p] (item={u'vgname': u'gluster_vg_sdc', u'pvname': u'/dev/sdc'}) => {"changed": false, "err": " Device /dev/sdc excluded by a filter.\n", "item": {"pvname": "/dev/sdc", "vgname": "gluster_vg_sdc"}, "msg": "Creating physical volume '/dev/sdc' failed", "rc": 5} failed: [vmm10.virt.iad3p] (item={u'vgname': u'gluster_vg_sdc', u'pvname': u'/dev/sdc'}) => {"changed": false, "err": " Device /dev/sdc excluded by a filter.\n", "item": {"pvname": "/dev/sdc", "vgname": "gluster_vg_sdc"}, "msg": "Creating physical volume '/dev/sdc' failed", "rc": 5} failed: [vmm11.virt.iad3p] (item={u'vgname': u'gluster_vg_sdc', u'pvname': u'/dev/sdc'}) => {"changed": false, "err": " Device /dev/sdc excluded by a filter.\n", "item": {"pvname": "/dev/sdc", "vgname": "gluster_vg_sdc"}, "msg": "Creating physical volume '/dev/sdc' failed", "rc": 5} failed: [vmm10.virt.iad3p] (item={u'vgname': u'gluster_vg_sdd', u'pvname': u'/dev/sdd'}) => {"changed": false, "err": " Device /dev/sdd excluded by a filter.\n", "item": {"pvname": "/dev/sdd", "vgname": "gluster_vg_sdd"}, "msg": "Creating physical volume '/dev/sdd' failed", "rc": 5} failed: [vmm12.virt.iad3p] (item={u'vgname': u'gluster_vg_sdd', u'pvname': u'/dev/sdd'}) => {"changed": false, "err": " Device /dev/sdd excluded by a filter.\n", "item": {"pvname": "/dev/sdd", "vgname": "gluster_vg_sdd"}, "msg": "Creating physical volume '/dev/sdd' failed", "rc": 5} failed: [vmm11.virt.iad3p] (item={u'vgname': u'gluster_vg_sdd', u'pvname': u'/dev/sdd'}) => {"changed": false, "err": " Device /dev/sdd excluded by a filter.\n", "item": {"pvname": "/dev/sdd", "vgname": "gluster_vg_sdd"}, "msg": "Creating physical volume '/dev/sdd' failed", "rc": 5}
Attached is the generated yml file ( /etc/ansible/hc_wizard_inventory.yml) and the "Deployment Failed" file
Also wondering if I hit this bug? https://bugzilla.redhat.com/show_bug.cgi?id=1635614
+Sachidananda URS <surs@redhat.com> +Gobinda Das <godas@redhat.com> to review the inventory file and failures
Hello Adrian,
Can you please provide the output of: # fdisk -l /dev/sdd # fdisk -l /dev/sdb
I think there could be stale signature on the disk causing this error. Some of the possible solutions to try: 1) # wipefs -a /dev/sdb # wipefs -a /dev/sdd
2) You can zero out first few sectors of disk by:
# dd if=/dev/zero of=/dev/sdb bs=1M count=10
3) Check if partition is visible in /proc/partitions If not: # partprobe /dev/sdb
4) Check if filtering is configured wrongly in /etc/lvm/lvm.conf grep for 'filter ='
-sac
-- Adrian Quintero