Hi all,
I deployed ovirt-csi in the k8s by applying yaml manually. I used the latest version of the container image.
(https://github.com/openshift/ovirt-csi-driver-operator/tree/master/assets)
After successfully creating pvc and pod, I tried to delete it.
And the pod is deleted, but the pvc is not deleted. This is because deleting a pod does not unmap /dev/rbd0 attached to the ovirt vm.
How can I delete the pvc successfully?
oVirt engine version is 4.4.7.6-1.el8.
Here is the engine log when deleting the pod:
2021-08-20 17:40:35,385+09 INFO [org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149) [] User admin@internal-authz with profile [internal] successfully logged in with scopes: ovirt-app-api ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate ovirt-ext=token:password-access
2021-08-20 17:40:35,403+09 INFO [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149) [68ee3182] Running command: CreateUserSessionCommand internal: false.
2021-08-20 17:40:35,517+09 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default task-149) [68ee3182] EVENT_ID: USER_VDC_LOGIN(30), User admin@internal-authz connecting from '192.168.7.169' using session 'XfDgNkmAGnPiZahK5itLhHQTCNHZ3JwXMMzOiZrYL3C32+1TTys3xcjrAmCIKPu02hgN1sdVpfZXWd0FznaPCQ==' logged in.
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:35,663+09 INFO [org.ovirt.engine.core.bll.storage.disk.DetachDiskFromVmCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Running command: DetachDiskFromVmCommand internal: false. Entities affected : ID: 59a7461c-72fe-4e01-86a7-c70243f31596 Type: VMAction group CONFIGURE_VM_STORAGE with role type USER
2021-08-20 17:40:35,664+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] START, HotUnPlugDiskVDSCommand(HostName = host, HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094', vmId='59a7461c-72fe-4e01-86a7-c70243f31596', diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4
2021-08-20 17:40:35,678+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Disk hot-unplug: <?xml version="1.0" encoding="UTF-8"?><hotunplug>
<devices>
<disk>
<alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/>
</disk>
</devices>
</hotunplug>
2021-08-20 17:40:35,749+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] FINISH, HotUnPlugDiskVDSCommand, return: , log id: 506ff4a4
2021-08-20 17:40:35,842+09 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID: USER_DETACH_DISK_FROM_VM(2,018), Disk pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 was successfully detached from VM centos by admin@internal-authz.
2021-08-20 17:40:35,916+09 ERROR [org.ovirt.engine.core.sso.service.SsoService] (default task-150) [] OAuthException invalid_grant: The provided authorization grant for the auth code has expired.
2021-08-20 17:40:35,917+09 ERROR [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-149) [] Cannot authenticate using authentication Headers: invalid_grant: The provided authorization grant for the auth code has expired.
2021-08-20 17:40:36,029+09 INFO [org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149) [] User admin@internal-authz with profile [internal] successfully logged in with scopes: ovirt-app-api ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate ovirt-ext=token:password-access
2021-08-20 17:40:36,046+09 INFO [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149) [4c4bf441] Running command: CreateUserSessionCommand internal: false.
2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't find relative path for class "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will return null
2021-08-20 17:40:49,241+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand] (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) [] START, DumpXmlsVDSCommand(HostName = host, Params:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094', vmIds='[59a7461c-72fe-4e01-86a7-c70243f31596]'}), log id: 7eb54202
2021-08-20 17:40:49,244+09 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand] (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) [] FINISH, DumpXmlsVDSCommand, return: {59a7461c-72fe-4e01-86a7-c70243f31596=<domain type='kvm' id='1' xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'>
<name>centos</name>
<uuid>59a7461c-72fe-4e01-86a7-c70243f31596</uuid>
<metadata xmlns:ns1="http://ovirt.org/vm/tune/1.0" xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
<ns1:qos/>
<ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
<ovirt-vm:balloonTarget type="int">4194304</ovirt-vm:balloonTarget>
<ovirt-vm:ballooningEnabled>true</ovirt-vm:ballooningEnabled>
<ovirt-vm:clusterVersion>4.6</ovirt-vm:clusterVersion>
<ovirt-vm:destroy_on_reboot type="bool">False</ovirt-vm:destroy_on_reboot>
<ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
<ovirt-vm:memGuaranteedSize type="int">4096</ovirt-vm:memGuaranteedSize>
<ovirt-vm:minGuaranteedMemoryMb type="int">4096</ovirt-vm:minGuaranteedMemoryMb>
<ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
<ovirt-vm:startTime type="float">1628558564.8754532</ovirt-vm:startTime>
<ovirt-vm:device alias="ua-7c9f38e9-8889-46c8-83bb-92efb9272de9" mac_address="56:6f:16:a8:00:07">
<ovirt-vm:network>ovirtmgmt</ovirt-vm:network>
<ovirt-vm:custom>
<ovirt-vm:queues>2</ovirt-vm:queues>
</ovirt-vm:custom>
</ovirt-vm:device>
<ovirt-vm:device devtype="disk" name="sda">
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
<ovirt-vm:guestName>/dev/sda</ovirt-vm:guestName>
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
<ovirt-vm:managed type="bool">False</ovirt-vm:managed>
<ovirt-vm:poolID>4ca6e0e8-e3a4-11eb-8830-480fcf63834f</ovirt-vm:poolID>
<ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
<ovirt-vm:volumeChain>
<ovirt-vm:volumeChainNode>
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
<ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset>
<ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003.lease</ovirt-vm:leasePath>
<ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:path>
<ovirt-vm:volumeID>4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:volumeID>
</ovirt-vm:volumeChainNode>
<ovirt-vm:volumeChainNode>
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
<ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset>
<ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4.lease</ovirt-vm:leasePath>
<ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:path>
<ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
</ovirt-vm:volumeChainNode>
</ovirt-vm:volumeChain>
</ovirt-vm:device>
<ovirt-vm:device devtype="disk" name="sdc">
<ovirt-vm:managed type="bool">False</ovirt-vm:managed>
</ovirt-vm:device>
</ovirt-vm:vm>
</metadata>
<maxMemory slots='16' unit='KiB'>16777216</maxMemory>
<memory unit='KiB'>4194304</memory>
<currentMemory unit='KiB'>4194304</currentMemory>
<vcpu placement='static' current='2'>16</vcpu>
<resource>
<partition>/machine</partition>
</resource>
<sysinfo type='smbios'>
<system>
<entry name='manufacturer'>oVirt</entry>
<entry name='product'>RHEL</entry>
<entry name='version'>8.4-1.2105.el8</entry>
<entry name='serial'>83e66af8-0500-11e6-9c43-bc00007c0000</entry>
<entry name='uuid'>59a7461c-72fe-4e01-86a7-c70243f31596</entry>
<entry name='family'>oVirt</entry>
</system>
</sysinfo>
<os>
<type arch='x86_64' machine='pc-q35-rhel8.4.0'>hvm</type>
<boot dev='hd'/>
<smbios mode='sysinfo'/>
</os>
<features>
<acpi/>
</features>
<cpu mode='custom' match='exact' check='full'>
<model fallback='forbid'>Skylake-Client-noTSX-IBRS</model>
<topology sockets='16' dies='1' cores='1' threads='1'/>
<feature policy='require' name='ssbd'/>
<feature policy='require' name='md-clear'/>
<feature policy='disable' name='mpx'/>
<feature policy='require' name='hypervisor'/>
<numa>
<cell id='0' cpus='0-15' memory='4194304' unit='KiB'/>
</numa>
</cpu>
<clock offset='variable' adjustment='0' basis='utc'>
<timer name='rtc' tickpolicy='catchup'/>
<timer name='pit' tickpolicy='delay'/>
<timer name='hpet' present='no'/>
</clock>
<on_poweroff>destroy</on_poweroff>
<on_reboot>restart</on_reboot>
<on_crash>destroy</on_crash>
<pm>
<suspend-to-mem enabled='no'/>
<suspend-to-disk enabled='no'/>
</pm>
<devices>
<emulator>/usr/libexec/qemu-kvm</emulator>
<disk type='file' device='cdrom'>
<driver name='qemu' error_policy='report'/>
<source startupPolicy='optional'/>
<target dev='sdc' bus='sata'/>
<readonly/>
<alias name='ua-df0ac774-3623-4868-8bd3-45c8f2aa3dc4'/>
<address type='drive' controller='0' bus='0' target='0' unit='2'/>
</disk>
<disk type='file' device='disk' snapshot='no'>
<driver name='qemu' type='qcow2' cache='none' error_policy='stop' io='threads'/>
<source file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4' index='8'>
<seclabel model='dac' relabel='no'/>
</source>
<backingStore type='file' index='1'>
<format type='qcow2'/>
<source file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003'>
<seclabel model='dac' relabel='no'/>
</source>
<backingStore/>
</backingStore>
<target dev='sda' bus='scsi'/>
<serial>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</serial>
<alias name='ua-bee44276-234f-4ed7-8a8a-d90a5e3cb5b3'/>
<address type='drive' controller='0' bus='0' target='0' unit='0'/>
</disk>
<controller type='pci' index='1' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='1' port='0x10'/>
<alias name='pci.1'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0' multifunction='on'/>
</controller>
<controller type='pci' index='2' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='2' port='0x11'/>
<alias name='pci.2'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x1'/>
</controller>
<controller type='pci' index='3' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='3' port='0x12'/>
<alias name='pci.3'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x2'/>
</controller>
<controller type='pci' index='4' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='4' port='0x13'/>
<alias name='pci.4'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x3'/>
</controller>
<controller type='pci' index='5' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='5' port='0x14'/>
<alias name='pci.5'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x4'/>
</controller>
<controller type='pci' index='6' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='6' port='0x15'/>
<alias name='pci.6'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x5'/>
</controller>
<controller type='pci' index='7' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='7' port='0x16'/>
<alias name='pci.7'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x6'/>
</controller>
<controller type='pci' index='8' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='8' port='0x17'/>
<alias name='pci.8'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x7'/>
</controller>
<controller type='pci' index='9' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='9' port='0x18'/>
<alias name='pci.9'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x0' multifunction='on'/>
</controller>
<controller type='pci' index='10' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='10' port='0x19'/>
<alias name='pci.10'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x1'/>
</controller>
<controller type='pci' index='11' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='11' port='0x1a'/>
<alias name='pci.11'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x2'/>
</controller>
<controller type='pci' index='12' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='12' port='0x1b'/>
<alias name='pci.12'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x3'/>
</controller>
<controller type='pci' index='13' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='13' port='0x1c'/>
<alias name='pci.13'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x4'/>
</controller>
<controller type='pci' index='14' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='14' port='0x1d'/>
<alias name='pci.14'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x5'/>
</controller>
<controller type='pci' index='15' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='15' port='0x1e'/>
<alias name='pci.15'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x6'/>
</controller>
<controller type='pci' index='16' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='16' port='0x1f'/>
<alias name='pci.16'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x7'/>
</controller>
<controller type='pci' index='0' model='pcie-root'>
<alias name='pcie.0'/>
</controller>
<controller type='sata' index='0'>
<alias name='ide'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x1f' function='0x2'/>
</controller>
<controller type='virtio-serial' index='0' ports='16'>
<alias name='virtio-serial0'/>
<address type='pci' domain='0x0000' bus='0x04' slot='0x00' function='0x0'/>
</controller>
<controller type='scsi' index='0' model='virtio-scsi'>
<alias name='ua-82c49f93-c4e8-460b-bb7d-95db0e9d87a0'/>
<address type='pci' domain='0x0000' bus='0x02' slot='0x00' function='0x0'/>
</controller>
<controller type='usb' index='0' model='qemu-xhci' ports='8'>
<alias name='ua-ad56daea-edb1-45c7-a1ab-2a7db3aaeee2'/>
<address type='pci' domain='0x0000' bus='0x03' slot='0x00' function='0x0'/>
</controller>
<interface type='bridge'>
<mac address='56:6f:16:a8:00:07'/>
<source bridge='ovirtmgmt'/>
<target dev='vnet0'/>
<model type='virtio'/>
<driver name='vhost' queues='2'/>
<filterref filter='vdsm-no-mac-spoofing'/>
<link state='up'/>
<mtu size='1500'/>
<alias name='ua-7c9f38e9-8889-46c8-83bb-92efb9272de9'/>
<address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/>
</interface>
<channel type='unix'>
<source mode='bind' path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.ovirt-guest-agent.0'/>
<target type='virtio' name='ovirt-guest-agent.0' state='disconnected'/>
<alias name='channel0'/>
<address type='virtio-serial' controller='0' bus='0' port='1'/>
</channel>
<channel type='unix'>
<source mode='bind' path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.org.qemu.guest_agent.0'/>
<target type='virtio' name='org.qemu.guest_agent.0' state='connected'/>
<alias name='channel1'/>
<address type='virtio-serial' controller='0' bus='0' port='2'/>
</channel>
<channel type='spicevmc'>
<target type='virtio' name='com.redhat.spice.0' state='disconnected'/>
<alias name='channel2'/>
<address type='virtio-serial' controller='0' bus='0' port='3'/>
</channel>
<input type='tablet' bus='usb'>
<alias name='input0'/>
<address type='usb' bus='0' port='1'/>
</input>
<input type='mouse' bus='ps2'>
<alias name='input1'/>
</input>
<input type='keyboard' bus='ps2'>
<alias name='input2'/>
</input>
<graphics type='vnc' port='5900' autoport='yes' listen='192.168.7.18' keymap='en-us' passwdValidTo='1970-01-01T00:00:01'>
<listen type='network' address='192.168.7.18' network='vdsm-ovirtmgmt'/>
</graphics>
<graphics type='spice' port='5901' tlsPort='5902' autoport='yes' listen='192.168.7.18' passwdValidTo='1970-01-01T00:00:01'>
<listen type='network' address='192.168.7.18' network='vdsm-ovirtmgmt'/>
<channel name='main' mode='secure'/>
<channel name='display' mode='secure'/>
<channel name='inputs' mode='secure'/>
<channel name='cursor' mode='secure'/>
<channel name='playback' mode='secure'/>
<channel name='record' mode='secure'/>
<channel name='smartcard' mode='secure'/>
<channel name='usbredir' mode='secure'/>
</graphics>
<video>
<model type='qxl' ram='65536' vram='8192' vgamem='16384' heads='1' primary='yes'/>
<alias name='ua-799f065a-b2b9-4e37-a502-f86c7cc8dc51'/>
<address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x0'/>
</video>
<memballoon model='virtio'>
<stats period='5'/>
<alias name='ua-c2bfe0b9-065a-46b7-9b0b-ef7e0f699611'/>
<address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x0'/>
</memballoon>
<rng model='virtio'>
<backend model='random'>/dev/urandom</backend>
<alias name='ua-fa110e6b-5eed-4b4b-93d8-0ac5de08aa2e'/>
<address type='pci' domain='0x0000' bus='0x06' slot='0x00' function='0x0'/>
</rng>
</devices>
<seclabel type='dynamic' model='selinux' relabel='yes'>
<label>system_u:system_r:svirt_t:s0:c437,c650</label>
<imagelabel>system_u:object_r:svirt_image_t:s0:c437,c650</imagelabel>
</seclabel>
<seclabel type='dynamic' model='dac' relabel='yes'>
<label>+107:+107</label>
<imagelabel>+107:+107</imagelabel>
</seclabel>
<qemu:capabilities>
<qemu:add capability='blockdev'/>
<qemu:add capability='incremental-backup'/>
</qemu:capabilities>
</domain>
}, log id: 7eb54202
Here is the engine log when deleting the pvc:
2021-08-20 17:43:12,964 - cinder.volume.driver - INFO - Driver hasn't implemented _init_vendor_properties()
2021-08-20 17:43:12,990 - cinderlib-client - INFO - Deleting volume '63a64445-1659-4d5f-8847-e7266e64b09e' [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
2021-08-20 17:43:28,856 - cinder.volume.drivers.rbd - WARNING - ImageBusy error raised while deleting rbd volume. This may have been caused by a connection from a client that has crashed and, if so, may be resolved by retrying the delete after 30 seconds has elapsed.
2021-08-20 17:43:28,900 - cinderlib-client - ERROR - Failure occurred when trying to run command 'delete_volume': ImageBusy error raised while deleting rbd volume. This may have been caused by a connection from a client that has crashed and, if so, may be resolved by retrying the delete after 30 seconds has elapsed. [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
2021-08-20 17:43:28,901 - cinder - CRITICAL - Unhandled error
Traceback (most recent call last):
File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 1195, in delete_volume
_try_remove_volume(client, volume_name)
File "/usr/lib/python3.6/site-packages/cinder/utils.py", line 683, in _wrapper
return r.call(f, *args, **kwargs)
File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 409, in call
do = self.iter(retry_state=retry_state)
File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 368, in iter
raise retry_exc.reraise()
File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 186, in reraise
raise self.last_attempt.result()
File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 425, in result
return self.__get_result()
File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 384, in __get_result
raise self._exception
File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 412, in call
result = fn(*args, **kwargs)
File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 1190, in _try_remove_volume
self.RBDProxy().remove(client.ioctx, volume_name)
File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in doit
result = proxy_call(self._autowrap, f, *args, **kwargs)
File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in proxy_call
rv = execute(f, *args, **kwargs)
File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 129, in execute
six.reraise(c, e, tb)
File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
raise value
File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 83, in tworker
rv = meth(*args, **kwargs)
File "rbd.pyx", line 767, in rbd.RBD.remove
rbd.ImageBusy: [errno 16] RBD image is busy (error removing image)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "./cinderlib-client.py", line 170, in main
args.command(args)
File "./cinderlib-client.py", line 218, in delete_volume
vol.delete()
File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 494, in delete
self._raise_with_resource()
File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 222, in _raise_with_resource
six.reraise(*exc_info)
File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
raise value
File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 487, in delete
self.backend.driver.delete_volume(self._ovo)
File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 1205, in delete_volume
raise exception.VolumeIsBusy(msg, volume_name=volume_name)
cinder.exception.VolumeIsBusy: ImageBusy error raised while deleting rbd volume. This may have been caused by a connection from a client that has crashed and, if so, may be resolved by retrying the delete after 30 seconds has elapsed.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "./cinderlib-client.py", line 390, in <module>
sys.exit(main(sys.argv[1:]))
File "./cinderlib-client.py", line 176, in main
sys.stderr.write(traceback.format_exc(e))
File "/usr/lib64/python3.6/traceback.py", line 167, in format_exc
return "".join(format_exception(*sys.exc_info(), limit=limit, chain=chain))
File "/usr/lib64/python3.6/traceback.py", line 121, in format_exception
type(value), value, tb, limit=limit).format(chain=chain))
File "/usr/lib64/python3.6/traceback.py", line 498, in __init__
_seen=_seen)
File "/usr/lib64/python3.6/traceback.py", line 509, in __init__
capture_locals=capture_locals)
File "/usr/lib64/python3.6/traceback.py", line 338, in extract
if limit >= 0:
TypeError: '>=' not supported between instances of 'VolumeIsBusy' and 'int'
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/AFYZX2HJZ3RBJS5P7ENIQ4UIAJGP5MGG/