On Tue, Apr 7, 2020 at 12:22 PM Strahil Nikolov <hunter86_bg@yahoo.com> wrote:


The simplest way would be to say that 'blacklisting everything in multipath.conf' will solve your problems.
In reality it is a little bit more complicated.


Interesting your arguments Strahil. To be digged more at my part.
In the mean time this approach below seems to have solved all the problems.

Preamble: I was able to put the new disk in a new PCI slot so that the /dev/nvmeXX names remained consistent with previous setup, but anyway LVM seemed to complain and I was not able again  to create VG on PV
So it was confirmed my suspect that nvme disks were not filtered out by LVM and created confusion..

- under /etc/lvm
[root@ovirt lvm]# diff lvm.conf lvm.conf.orig
142d141
< filter = [ "r|nvme|", "a|.*/|" ]
153d151
< global_filter = [ "r|nvme|", "a|.*/|" ]
[root@ovirt lvm]#

NOTE: it was not sufficient the "filter" directive alone, even if in theory from what I read, the global_filter one should come in place only when lvm metad is active, while in ovirt node it is not.. To be understood better...

- under /etc
I noticed that also the OS disk was recomprised into multipath, so I blacklisted it, making the file private at the end...
 
[root@ovirt etc]# diff -u3 multipath.conf.orig multipath.conf
--- multipath.conf.orig 2020-04-07 16:25:12.148044435 +0200
+++ multipath.conf 2020-04-07 10:55:44.728734050 +0200
@@ -1,4 +1,5 @@
 # VDSM REVISION 1.8
+# VDSM PRIVATE
 
 # This file is managed by vdsm.
 #
@@ -164,6 +165,7 @@
 
 blacklist {
         protocol "(scsi:adt|scsi:sbp)"
+        wwid INTEL_SSDSCKKI256G8_PHLA835602TE256J
 }
 
 # Remove devices entries when overrides section is available.
[root@ovirt etc]#

- rebuild initramfs

cp /boot/$(imgbase layer --current)/initramfs-$(uname -r).img /root/
dracut -f /boot/$(imgbase layer --current)/initramfs-$(uname -r).img
cp -p  /boot/$(imgbase layer --current)/initramfs-$(uname -r).img /boot/

After reboot I see configured as multipath devices the disks to be used for gluster

[root@ovirt etc]# multipath -l
nvme.8086-50484b53373530353031325233373541474e-494e54454c205353 dm-2 NVME,INTEL SSDPED1K375GA                    
size=349G features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=0 status=active
  `- 0:0:1:0 nvme0n1 259:0 active undef running
eui.01000000010000005cd2e4b5e7db4d51 dm-0 NVME,INTEL SSDPEDKX040T7                    
size=3.6T features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=0 status=active
  `- 2:0:1:0 nvme2n1 259:1 active undef running
eui.01000000010000005cd2e4e359284f51 dm-1 NVME,INTEL SSDPE2KX010T7                    
size=932G features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=0 status=active
  `- 1:0:1:0 nvme1n1 259:2 active undef running
[root@ovirt etc]#

And LVM commands don't complain any more with duplicates:

[root@ovirt etc]# pvs
  PV                                                                                                         VG                 Fmt  Attr PSize    PFree  
  /dev/mapper/eui.01000000010000005cd2e4b5e7db4d51                                                           gluster_vg_4t      lvm2 a--    <3.64t      0
  /dev/mapper/eui.01000000010000005cd2e4e359284f51                                                           gluster_vg_nvme1n1 lvm2 a--   931.51g      0
  /dev/mapper/nvme.8086-50484b53373530353031325233373541474e-494e54454c20535344504544314b3337354741-00000001 gluster_vg_nvme0n1 lvm2 a--   349.32g      0
  /dev/sda2                                                                                                  onn                lvm2 a--  <228.40g <43.87g
[root@ovirt etc]#

And as you see I was able to create now the VG on top of the new PV on the 4Tb disk:

[root@ovirt etc]# vgs
  VG                 #PV #LV #SN Attr   VSize    VFree  
  gluster_vg_4t        1   2   0 wz--n-   <3.64t      0
  gluster_vg_nvme0n1   1   3   0 wz--n-  349.32g      0
  gluster_vg_nvme1n1   1   2   0 wz--n-  931.51g      0
  onn                  1  11   0 wz--n- <228.40g <43.87g
[root@ovirt etc]#

[root@ovirt etc]# lvs gluster_vg_4t
  LV             VG            Attr       LSize  Pool    Origin Data%  Meta%  Move Log Cpy%Sync Convert
  gluster_lv_big gluster_vg_4t Vwi-aot--- <4.35t my_pool        0.05                                  
  my_pool        gluster_vg_4t twi-aot--- <3.61t                0.05   0.14                            
[root@ovirt etc]#

Let's go hunting for the next problem... ;-)

Gianluca