[ovirt-users] Disk from FC storage not removed

Krzysztof Dajka alteriks at gmail.com
Mon Jun 13 09:06:13 UTC 2016


Hi Nir,

Thanks for solution. I didn't notice the guest /dev/backupvg01/backuplv01
on all hypervisors. It seems that I've got this issue with 2 additionals
volumes, but no one noticed because they were only few gb.


[root at wrops2 BLUE/WRO ~]# ls -l /sys/block/$(basename $(readlink
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8))/holders
total 0
lrwxrwxrwx. 1 root root 0 Jun 13 10:48 dm-43 -> ../../dm-43

[root at wrops2 BLUE/WRO ~]# pvscan --cache
[root at wrops2 BLUE/WRO ~]# vgs -o pv_name,vg_name
  PV
      VG

/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
backupvg01
  /dev/sda2
     centos_wrops2

/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/99a1c067-9728-484a-a0cb-cb6689d5724c
deployvg
  /dev/mapper/3600000e00d0000000024057200000000
     e69d1c16-36d1-4375-aaee-69f5a5ce1616

/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/86a6d83f-2661-4fe3-8874-ce4d8a111c0d
jenkins
  /dev/sda3
     w2vg1

[root at wrops2 BLUE/WRO ~]# dmsetup info
Name:              backupvg01-backuplv01
State:             ACTIVE
Read Ahead:        8192
Tables present:    LIVE
Open count:        0
Event number:      0
Major, minor:      253, 43
Number of targets: 1
UUID: LVM-ubxOH5R2h6B8JwLGfhpiNjnAKlPxMPy6KfkeLBxXajoT3gxU0yC5JvOQQVkixrTA

[root at wrops2 BLUE/WRO ~]# lvchange -an /dev/backupvg01/backuplv01
[root at wrops2 BLUE/WRO ~]# lvremove
/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
Do you really want to remove active logical volume
ee53af81-820d-4916-b766-5236ca99daf8? [y/n]: y
  Logical volume "ee53af81-820d-4916-b766-5236ca99daf8" successfully removed


Would this configuration in lvm.conf:
filter = [ "r|/dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/.*|" ]
on all hypervisors solve problem of scanning guest volumes?

2016-06-11 23:16 GMT+02:00 Nir Soffer <nsoffer at redhat.com>:

> On Thu, Jun 9, 2016 at 11:46 AM, Krzysztof Dajka <alteriks at gmail.com>
> wrote:
> > Hi,
> >
> > Recently I tried to delete 1TB disk created on top ~3TB LUN from
> > ovirtengine.
> > Disk is preallocated and I backuped data to other disk so I could
> recreate
> > it once again as thin volume. I couldn't remove this disk when it was
> > attached to a VM. But once I detached it I could remove it permanently.
> The
> > thing is it only disappeared from ovirtengine GUI.
> >
> > I've got 4 hosts with FC HBA attached to storage array and all of them
> are
> > saying that this 1TB disk which should be gone is opened by all hosts
> > simultaneously.
> >
> > [root at wrops1 BLUE ~]# lvdisplay -m
> >
> /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
> >   --- Logical volume ---
> >   LV Path
> >
> /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
> >   LV Name                ee53af81-820d-4916-b766-5236ca99daf8
> >   VG Name                e69d1c16-36d1-4375-aaee-69f5a5ce1616
> >   LV UUID                sBdBRk-tNyZ-Rval-F4lw-ka6X-wOe8-AQenTb
> >   LV Write Access        read/write
> >   LV Creation host, time wrops1.blue, 2015-07-31 10:40:57 +0200
> >   LV Status              available
> >   # open                 1
> >   LV Size                1.00 TiB
> >   Current LE             8192
> >   Segments               1
> >   Allocation             inherit
> >   Read ahead sectors     auto
> >   - currently set to     8192
> >   Block device           253:29
> >
> >   --- Segments ---
> >   Logical extents 0 to 8191:
> >     Type                linear
> >     Physical volume     /dev/mapper/3600000e00d0000000024057200000000
> >     Physical extents    8145 to 16336
> >
> > Deactivating LV doesn't work:
> > [root at wrops1 BLUE ~]# lvchange -an
> >
> /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
> >   Logical volume
> >
> e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 is
> > used by another device.
>
> Looks like your lv is used as a physical volume on another vg - probably
> a vg created on a guest. Lvm and systemd are trying hard to discover
> stuff on multipath devices and expose anything to the hypervisor.
>
> Can you share the output of:
>
>     ls -l /sys/block/$(basename $(readlink
>
> /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8))/holders
>
> And:
>
>     pvscan --cache
>     vgs -o pv_name,vg_name
>
> Nir
>
> > Removing from hypervisor doesn't work either.
> > [root at wrops1 BLUE ~]# lvremove --force
> >
> /dev/e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8
> >   Logical volume
> >
> e69d1c16-36d1-4375-aaee-69f5a5ce1616/ee53af81-820d-4916-b766-5236ca99daf8 is
> > used by another device.
> >
> > I tried and rebooted one host and as soon as it booted the volume became
> > opened once again. Lsof on all hosts doesn't give anything meaningful
> > regarding this LV. As opposed to other LV which are used by qemu-kvm.
> >
> > Has anyone encountered similar problem? How can I remove this LV?
> >
> > _______________________________________________
> > Users mailing list
> > Users at ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20160613/f4f124cc/attachment-0001.html>


More information about the Users mailing list