Hi Nir,
Thanks for solution. I didn't notice the guest /dev/backupvg01/backuplv01
on all hypervisors. It seems that I've got this issue with 2 additionals
volumes, but no one noticed because they were only few gb.
[root@wrops2 BLUE/WRO ~]# ls -l /sys/block/$(basename $(readlink
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8))/holders
total 0
lrwxrwxrwx. 1 root root 0 Jun 13 10:48 dm-43 -> ../../dm-43
[root@wrops2 BLUE/WRO ~]# pvscan --cache
[root@wrops2 BLUE/WRO ~]# vgs -o pv_name,vg_name
PV
VG
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
backupvg01
/dev/sda2
centos_wrops2
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/99a1c067-9728-484a-a0cb-cb6689d5724c
deployvg
/dev/mapper/3600000e00d0000000024057200000000
e69d1c16-36d1-4375-aaee-69f5a5ce1616
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/86a6d83f-2661-4fe3-8874-ce4d8a111c0d
jenkins
/dev/sda3
w2vg1
[root@wrops2 BLUE/WRO ~]# dmsetup info
Name: backupvg01-backuplv01
State: ACTIVE
Read Ahead: 8192
Tables present: LIVE
Open count: 0
Event number: 0
Major, minor: 253, 43
Number of targets: 1
UUID: LVM-ubxOH5R2h6B8JwLGfhpiNjnAKlPxMPy6KfkeLBxXajoT3gxU0yC5JvOQQVkixrTA
[root@wrops2 BLUE/WRO ~]# lvchange -an /dev/backupvg01/backuplv01
[root@wrops2 BLUE/WRO ~]# lvremove
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
Do you really want to remove active logical volume
ee53af81-820d-4916-b766-5236ca99daf8? [y/n]: y
Logical volume "ee53af81-820d-4916-b766-5236ca99daf8" successfully removed
Would this configuration in lvm.conf:
filter = [ "r|/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/.*|" ]
on all hypervisors solve problem of scanning guest volumes?
2016-06-11 23:16 GMT+02:00 Nir Soffer <nsoffer(a)redhat.com>:
On Thu, Jun 9, 2016 at 11:46 AM, Krzysztof Dajka
<alteriks(a)gmail.com>
wrote:
> Hi,
>
> Recently I tried to delete 1TB disk created on top ~3TB LUN from
> ovirtengine.
> Disk is preallocated and I backuped data to other disk so I could
recreate
> it once again as thin volume. I couldn't remove this disk when it was
> attached to a VM. But once I detached it I could remove it permanently.
The
> thing is it only disappeared from ovirtengine GUI.
>
> I've got 4 hosts with FC HBA attached to storage array and all of them
are
> saying that this 1TB disk which should be gone is opened by all hosts
> simultaneously.
>
> [root@wrops1 BLUE ~]# lvdisplay -m
>
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
> --- Logical volume ---
> LV Path
>
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
> LV Name ee53af81-820d-4916-b766-5236ca99daf8
> VG Name e69d1c16-36d1-4375-aaee-69f5a5ce1616
> LV UUID sBdBRk-tNyZ-Rval-F4lw-ka6X-wOe8-AQenTb
> LV Write Access read/write
> LV Creation host, time wrops1.blue, 2015-07-31 10:40:57 +0200
> LV Status available
> # open 1
> LV Size 1.00 TiB
> Current LE 8192
> Segments 1
> Allocation inherit
> Read ahead sectors auto
> - currently set to 8192
> Block device 253:29
>
> --- Segments ---
> Logical extents 0 to 8191:
> Type linear
> Physical volume /dev/mapper/3600000e00d0000000024057200000000
> Physical extents 8145 to 16336
>
> Deactivating LV doesn't work:
> [root@wrops1 BLUE ~]# lvchange -an
>
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
> Logical volume
>
e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 is
> used by another device.
Looks like your lv is used as a physical volume on another vg - probably
a vg created on a guest. Lvm and systemd are trying hard to discover
stuff on multipath devices and expose anything to the hypervisor.
Can you share the output of:
ls -l /sys/block/$(basename $(readlink
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8))/holders
And:
pvscan --cache
vgs -o pv_name,vg_name
Nir
> Removing from hypervisor doesn't work either.
> [root@wrops1 BLUE ~]# lvremove --force
>
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
> Logical volume
>
e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 is
> used by another device.
>
> I tried and rebooted one host and as soon as it booted the volume became
> opened once again. Lsof on all hosts doesn't give anything meaningful
> regarding this LV. As opposed to other LV which are used by qemu-kvm.
>
> Has anyone encountered similar problem? How can I remove this LV?
>
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users
>