
pod deletion should invoke unpublish the PVC which detaches it from the node which is seen in the engine log: 2021-08-20 17:40:35,664+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] START, HotUnPlugDiskVDSCommand(HostName = host, HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094', vmId='59a7461c-72fe-4e01-86a7-c70243f31596', diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4 2021-08-20 17:40:35,678+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Disk hot-unplug: <?xml version="1.0" encoding="UTF-8"?><hotunplug> <devices> <disk> <alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/> </disk> </devices> </hotunplug> 2021-08-20 17:40:35,749+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] FINISH, HotUnPlugDiskVDSCommand, return: , log id: 506ff4a4 2021-08-20 17:40:35,842+09 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID: USER_DETACH_DISK_FROM_VM(2,018), Disk pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 was successfully detached from VM centos by admin@internal-authz. I suspect something keeps the volume busy, can run: $ rbd status <pool_name>/volume-63a64445-1659-4d5f-8847-e7266e64b09e On Mon, Aug 23, 2021 at 3:56 AM <ssarang520@gmail.com> wrote:
Hi all,
I deployed ovirt-csi in the k8s by applying yaml manually. I used the latest version of the container image. (https://github.com/openshift/ovirt-csi-driver-operator/tree/master/assets)
After successfully creating pvc and pod, I tried to delete it. And the pod is deleted, but the pvc is not deleted. This is because deleting a pod does not unmap /dev/rbd0 attached to the ovirt vm.
How can I delete the pvc successfully?
oVirt engine version is 4.4.7.6-1.el8. Here is the engine log when deleting the pod:
2021-08-20 17:40:35,385+09 INFO [org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149) [] User admin@internal-authz with profile [internal] successfully logged in with scopes: ovirt-app-api ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate ovirt-ext=token:password-access 2021-08-20 17:40:35,403+09 INFO [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149) [68ee3182] Running command: CreateUserSessionCommand internal: false. 2021-08-20 17:40:35,517+09 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default task-149) [68ee3182] EVENT_ID: USER_VDC_LOGIN(30), User admin@internal-authz connecting from '192.168.7.169' using session 'XfDgNkmAGnPiZahK5itLhHQTCNHZ3JwXMMzOiZrYL3C32+1TTys3xcjrAmCIKPu02hgN1sdVpfZXWd0FznaPCQ==' logged in. 2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:35,663+09 INFO [org.ovirt.engine.core.bll.storage.disk.DetachDiskFromVmCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Running command: DetachDiskFromVmCommand internal: false. Entities affected : ID: 59a7461c-72fe-4e01-86a7-c70243f31596 Type: VMAction group CONFIGURE_VM_STORAGE with role type USER 2021-08-20 17:40:35,664+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] START, HotUnPlugDiskVDSCommand(HostName = host, HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094', vmId='59a7461c-72fe-4e01-86a7-c70243f31596', diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4 2021-08-20 17:40:35,678+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Disk hot-unplug: <?xml version="1.0" encoding="UTF-8"?><hotunplug> <devices> <disk> <alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/> </disk> </devices> </hotunplug>
2021-08-20 17:40:35,749+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] FINISH, HotUnPlugDiskVDSCommand, return: , log id: 506ff4a4 2021-08-20 17:40:35,842+09 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID: USER_DETACH_DISK_FROM_VM(2,018), Disk pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 was successfully detached from VM centos by admin@internal-authz. 2021-08-20 17:40:35,916+09 ERROR [org.ovirt.engine.core.sso.service.SsoService] (default task-150) [] OAuthException invalid_grant: The provided authorization grant for the auth code has expired. 2021-08-20 17:40:35,917+09 ERROR [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-149) [] Cannot authenticate using authentication Headers: invalid_grant: The provided authorization grant for the auth code has expired. 2021-08-20 17:40:36,029+09 INFO [org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149) [] User admin@internal-authz with profile [internal] successfully logged in with scopes: ovirt-app-api ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate ovirt-ext=token:password-access 2021-08-20 17:40:36,046+09 INFO [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149) [4c4bf441] Running command: CreateUserSessionCommand internal: false. 2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null 2021-08-20 17:40:49,241+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand] (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) [] START, DumpXmlsVDSCommand(HostName = host, Params:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094', vmIds='[59a7461c-72fe-4e01-86a7-c70243f31596]'}), log id: 7eb54202 2021-08-20 17:40:49,244+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand] (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) [] FINISH, DumpXmlsVDSCommand, return: {59a7461c-72fe-4e01-86a7-c70243f31596=<domain type='kvm' id='1' xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'> <name>centos</name> <uuid>59a7461c-72fe-4e01-86a7-c70243f31596</uuid> <metadata xmlns:ns1="http://ovirt.org/vm/tune/1.0" xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> <ns1:qos/> <ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> <ovirt-vm:balloonTarget type="int">4194304</ovirt-vm:balloonTarget> <ovirt-vm:ballooningEnabled>true</ovirt-vm:ballooningEnabled> <ovirt-vm:clusterVersion>4.6</ovirt-vm:clusterVersion> <ovirt-vm:destroy_on_reboot type="bool">False</ovirt-vm:destroy_on_reboot> <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused> <ovirt-vm:memGuaranteedSize type="int">4096</ovirt-vm:memGuaranteedSize> <ovirt-vm:minGuaranteedMemoryMb type="int">4096</ovirt-vm:minGuaranteedMemoryMb> <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior> <ovirt-vm:startTime type="float">1628558564.8754532</ovirt-vm:startTime> <ovirt-vm:device alias="ua-7c9f38e9-8889-46c8-83bb-92efb9272de9" mac_address="56:6f:16:a8:00:07"> <ovirt-vm:network>ovirtmgmt</ovirt-vm:network> <ovirt-vm:custom> <ovirt-vm:queues>2</ovirt-vm:queues> </ovirt-vm:custom> </ovirt-vm:device> <ovirt-vm:device devtype="disk" name="sda"> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID> <ovirt-vm:guestName>/dev/sda</ovirt-vm:guestName> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID> <ovirt-vm:managed type="bool">False</ovirt-vm:managed> <ovirt-vm:poolID>4ca6e0e8-e3a4-11eb-8830-480fcf63834f</ovirt-vm:poolID> <ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID> <ovirt-vm:volumeChain> <ovirt-vm:volumeChainNode> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID> <ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset> <ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003.lease</ovirt-vm:leasePath> <ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:path> <ovirt-vm:volumeID>4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:volumeID> </ovirt-vm:volumeChainNode> <ovirt-vm:volumeChainNode> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID> <ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset> <ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4.lease</ovirt-vm:leasePath> <ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:path> <ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID> </ovirt-vm:volumeChainNode> </ovirt-vm:volumeChain> </ovirt-vm:device> <ovirt-vm:device devtype="disk" name="sdc"> <ovirt-vm:managed type="bool">False</ovirt-vm:managed> </ovirt-vm:device> </ovirt-vm:vm> </metadata> <maxMemory slots='16' unit='KiB'>16777216</maxMemory> <memory unit='KiB'>4194304</memory> <currentMemory unit='KiB'>4194304</currentMemory> <vcpu placement='static' current='2'>16</vcpu> <resource> <partition>/machine</partition> </resource> <sysinfo type='smbios'> <system> <entry name='manufacturer'>oVirt</entry> <entry name='product'>RHEL</entry> <entry name='version'>8.4-1.2105.el8</entry> <entry name='serial'>83e66af8-0500-11e6-9c43-bc00007c0000</entry> <entry name='uuid'>59a7461c-72fe-4e01-86a7-c70243f31596</entry> <entry name='family'>oVirt</entry> </system> </sysinfo> <os> <type arch='x86_64' machine='pc-q35-rhel8.4.0'>hvm</type> <boot dev='hd'/> <smbios mode='sysinfo'/> </os> <features> <acpi/> </features> <cpu mode='custom' match='exact' check='full'> <model fallback='forbid'>Skylake-Client-noTSX-IBRS</model> <topology sockets='16' dies='1' cores='1' threads='1'/> <feature policy='require' name='ssbd'/> <feature policy='require' name='md-clear'/> <feature policy='disable' name='mpx'/> <feature policy='require' name='hypervisor'/> <numa> <cell id='0' cpus='0-15' memory='4194304' unit='KiB'/> </numa> </cpu> <clock offset='variable' adjustment='0' basis='utc'> <timer name='rtc' tickpolicy='catchup'/> <timer name='pit' tickpolicy='delay'/> <timer name='hpet' present='no'/> </clock> <on_poweroff>destroy</on_poweroff> <on_reboot>restart</on_reboot> <on_crash>destroy</on_crash> <pm> <suspend-to-mem enabled='no'/> <suspend-to-disk enabled='no'/> </pm> <devices> <emulator>/usr/libexec/qemu-kvm</emulator> <disk type='file' device='cdrom'> <driver name='qemu' error_policy='report'/> <source startupPolicy='optional'/> <target dev='sdc' bus='sata'/> <readonly/> <alias name='ua-df0ac774-3623-4868-8bd3-45c8f2aa3dc4'/> <address type='drive' controller='0' bus='0' target='0' unit='2'/> </disk> <disk type='file' device='disk' snapshot='no'> <driver name='qemu' type='qcow2' cache='none' error_policy='stop' io='threads'/> <source file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4' index='8'> <seclabel model='dac' relabel='no'/> </source> <backingStore type='file' index='1'> <format type='qcow2'/> <source file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003'> <seclabel model='dac' relabel='no'/> </source> <backingStore/> </backingStore> <target dev='sda' bus='scsi'/> <serial>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</serial> <alias name='ua-bee44276-234f-4ed7-8a8a-d90a5e3cb5b3'/> <address type='drive' controller='0' bus='0' target='0' unit='0'/> </disk> <controller type='pci' index='1' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='1' port='0x10'/> <alias name='pci.1'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0' multifunction='on'/> </controller> <controller type='pci' index='2' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='2' port='0x11'/> <alias name='pci.2'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x1'/> </controller> <controller type='pci' index='3' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='3' port='0x12'/> <alias name='pci.3'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x2'/> </controller> <controller type='pci' index='4' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='4' port='0x13'/> <alias name='pci.4'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x3'/> </controller> <controller type='pci' index='5' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='5' port='0x14'/> <alias name='pci.5'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x4'/> </controller> <controller type='pci' index='6' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='6' port='0x15'/> <alias name='pci.6'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x5'/> </controller> <controller type='pci' index='7' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='7' port='0x16'/> <alias name='pci.7'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x6'/> </controller> <controller type='pci' index='8' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='8' port='0x17'/> <alias name='pci.8'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x7'/> </controller> <controller type='pci' index='9' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='9' port='0x18'/> <alias name='pci.9'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x0' multifunction='on'/> </controller> <controller type='pci' index='10' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='10' port='0x19'/> <alias name='pci.10'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x1'/> </controller> <controller type='pci' index='11' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='11' port='0x1a'/> <alias name='pci.11'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x2'/> </controller> <controller type='pci' index='12' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='12' port='0x1b'/> <alias name='pci.12'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x3'/> </controller> <controller type='pci' index='13' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='13' port='0x1c'/> <alias name='pci.13'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x4'/> </controller> <controller type='pci' index='14' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='14' port='0x1d'/> <alias name='pci.14'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x5'/> </controller> <controller type='pci' index='15' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='15' port='0x1e'/> <alias name='pci.15'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x6'/> </controller> <controller type='pci' index='16' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='16' port='0x1f'/> <alias name='pci.16'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x7'/> </controller> <controller type='pci' index='0' model='pcie-root'> <alias name='pcie.0'/> </controller> <controller type='sata' index='0'> <alias name='ide'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x1f' function='0x2'/> </controller> <controller type='virtio-serial' index='0' ports='16'> <alias name='virtio-serial0'/> <address type='pci' domain='0x0000' bus='0x04' slot='0x00' function='0x0'/> </controller> <controller type='scsi' index='0' model='virtio-scsi'> <alias name='ua-82c49f93-c4e8-460b-bb7d-95db0e9d87a0'/> <address type='pci' domain='0x0000' bus='0x02' slot='0x00' function='0x0'/> </controller> <controller type='usb' index='0' model='qemu-xhci' ports='8'> <alias name='ua-ad56daea-edb1-45c7-a1ab-2a7db3aaeee2'/> <address type='pci' domain='0x0000' bus='0x03' slot='0x00' function='0x0'/> </controller> <interface type='bridge'> <mac address='56:6f:16:a8:00:07'/> <source bridge='ovirtmgmt'/> <target dev='vnet0'/> <model type='virtio'/> <driver name='vhost' queues='2'/> <filterref filter='vdsm-no-mac-spoofing'/> <link state='up'/> <mtu size='1500'/> <alias name='ua-7c9f38e9-8889-46c8-83bb-92efb9272de9'/> <address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/> </interface> <channel type='unix'> <source mode='bind' path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.ovirt-guest-agent.0'/> <target type='virtio' name='ovirt-guest-agent.0' state='disconnected'/> <alias name='channel0'/> <address type='virtio-serial' controller='0' bus='0' port='1'/> </channel> <channel type='unix'> <source mode='bind' path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.org.qemu.guest_agent.0'/> <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/> <alias name='channel1'/> <address type='virtio-serial' controller='0' bus='0' port='2'/> </channel> <channel type='spicevmc'> <target type='virtio' name='com.redhat.spice.0' state='disconnected'/> <alias name='channel2'/> <address type='virtio-serial' controller='0' bus='0' port='3'/> </channel> <input type='tablet' bus='usb'> <alias name='input0'/> <address type='usb' bus='0' port='1'/> </input> <input type='mouse' bus='ps2'> <alias name='input1'/> </input> <input type='keyboard' bus='ps2'> <alias name='input2'/> </input> <graphics type='vnc' port='5900' autoport='yes' listen='192.168.7.18' keymap='en-us' passwdValidTo='1970-01-01T00:00:01'> <listen type='network' address='192.168.7.18' network='vdsm-ovirtmgmt'/> </graphics> <graphics type='spice' port='5901' tlsPort='5902' autoport='yes' listen='192.168.7.18' passwdValidTo='1970-01-01T00:00:01'> <listen type='network' address='192.168.7.18' network='vdsm-ovirtmgmt'/> <channel name='main' mode='secure'/> <channel name='display' mode='secure'/> <channel name='inputs' mode='secure'/> <channel name='cursor' mode='secure'/> <channel name='playback' mode='secure'/> <channel name='record' mode='secure'/> <channel name='smartcard' mode='secure'/> <channel name='usbredir' mode='secure'/> </graphics> <video> <model type='qxl' ram='65536' vram='8192' vgamem='16384' heads='1' primary='yes'/> <alias name='ua-799f065a-b2b9-4e37-a502-f86c7cc8dc51'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x0'/> </video> <memballoon model='virtio'> <stats period='5'/> <alias name='ua-c2bfe0b9-065a-46b7-9b0b-ef7e0f699611'/> <address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x0'/> </memballoon> <rng model='virtio'> <backend model='random'>/dev/urandom</backend> <alias name='ua-fa110e6b-5eed-4b4b-93d8-0ac5de08aa2e'/> <address type='pci' domain='0x0000' bus='0x06' slot='0x00' function='0x0'/> </rng> </devices> <seclabel type='dynamic' model='selinux' relabel='yes'> <label>system_u:system_r:svirt_t:s0:c437,c650</label> <imagelabel>system_u:object_r:svirt_image_t:s0:c437,c650</imagelabel> </seclabel> <seclabel type='dynamic' model='dac' relabel='yes'> <label>+107:+107</label> <imagelabel>+107:+107</imagelabel> </seclabel> <qemu:capabilities> <qemu:add capability='blockdev'/> <qemu:add capability='incremental-backup'/> </qemu:capabilities> </domain> }, log id: 7eb54202
Here is the engine log when deleting the pvc:
2021-08-20 17:43:12,964 - cinder.volume.driver - INFO - Driver hasn't implemented _init_vendor_properties() 2021-08-20 17:43:12,990 - cinderlib-client - INFO - Deleting volume '63a64445-1659-4d5f-8847-e7266e64b09e' [feefc62f-e7cb-435d-ae21-4b52b53fbdfa] 2021-08-20 17:43:28,856 - cinder.volume.drivers.rbd - WARNING - ImageBusy error raised while deleting rbd volume. This may have been caused by a connection from a client that has crashed and, if so, may be resolved by retrying the delete after 30 seconds has elapsed. 2021-08-20 17:43:28,900 - cinderlib-client - ERROR - Failure occurred when trying to run command 'delete_volume': ImageBusy error raised while deleting rbd volume. This may have been caused by a connection from a client that has crashed and, if so, may be resolved by retrying the delete after 30 seconds has elapsed. [feefc62f-e7cb-435d-ae21-4b52b53fbdfa] 2021-08-20 17:43:28,901 - cinder - CRITICAL - Unhandled error Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 1195, in delete_volume _try_remove_volume(client, volume_name) File "/usr/lib/python3.6/site-packages/cinder/utils.py", line 683, in _wrapper return r.call(f, *args, **kwargs) File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 409, in call do = self.iter(retry_state=retry_state) File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 368, in iter raise retry_exc.reraise() File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 186, in reraise raise self.last_attempt.result() File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 425, in result return self.__get_result() File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 384, in __get_result raise self._exception File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 412, in call result = fn(*args, **kwargs) File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 1190, in _try_remove_volume self.RBDProxy().remove(client.ioctx, volume_name) File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in doit result = proxy_call(self._autowrap, f, *args, **kwargs) File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in proxy_call rv = execute(f, *args, **kwargs) File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 129, in execute six.reraise(c, e, tb) File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise raise value File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 83, in tworker rv = meth(*args, **kwargs) File "rbd.pyx", line 767, in rbd.RBD.remove rbd.ImageBusy: [errno 16] RBD image is busy (error removing image)
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "./cinderlib-client.py", line 170, in main args.command(args) File "./cinderlib-client.py", line 218, in delete_volume vol.delete() File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 494, in delete self._raise_with_resource() File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 222, in _raise_with_resource six.reraise(*exc_info) File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise raise value File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 487, in delete self.backend.driver.delete_volume(self._ovo) File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 1205, in delete_volume raise exception.VolumeIsBusy(msg, volume_name=volume_name) cinder.exception.VolumeIsBusy: ImageBusy error raised while deleting rbd volume. This may have been caused by a connection from a client that has crashed and, if so, may be resolved by retrying the delete after 30 seconds has elapsed.
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "./cinderlib-client.py", line 390, in <module> sys.exit(main(sys.argv[1:])) File "./cinderlib-client.py", line 176, in main sys.stderr.write(traceback.format_exc(e)) File "/usr/lib64/python3.6/traceback.py", line 167, in format_exc return "".join(format_exception(*sys.exc_info(), limit=limit, chain=chain)) File "/usr/lib64/python3.6/traceback.py", line 121, in format_exception type(value), value, tb, limit=limit).format(chain=chain)) File "/usr/lib64/python3.6/traceback.py", line 498, in __init__ _seen=_seen) File "/usr/lib64/python3.6/traceback.py", line 509, in __init__ capture_locals=capture_locals) File "/usr/lib64/python3.6/traceback.py", line 338, in extract if limit >= 0: TypeError: '>=' not supported between instances of 'VolumeIsBusy' and 'int' _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/AFYZX2HJZ3RBJS...