pod deletion should invoke unpublish the PVC which detaches it from
the node which is seen in the engine log:
2021-08-20 17:40:35,664+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
(default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] START,
HotUnPlugDiskVDSCommand(HostName = host,
HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094',
vmId='59a7461c-72fe-4e01-86a7-c70243f31596',
diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4
2021-08-20 17:40:35,678+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
(default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Disk
hot-unplug: <?xml version="1.0"
encoding="UTF-8"?><hotunplug>
<devices>
<disk>
<alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/>
</disk>
</devices>
</hotunplug>
2021-08-20 17:40:35,749+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
(default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] FINISH,
HotUnPlugDiskVDSCommand, return: , log id: 506ff4a4
2021-08-20 17:40:35,842+09 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID:
USER_DETACH_DISK_FROM_VM(2,018), Disk
pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 was successfully detached
from VM centos by admin@internal-authz.
I suspect something keeps the volume busy, can run:
$ rbd status <pool_name>/volume-63a64445-1659-4d5f-8847-e7266e64b09e
On Mon, Aug 23, 2021 at 3:56 AM <ssarang520(a)gmail.com> wrote:
Hi all,
I deployed ovirt-csi in the k8s by applying yaml manually. I used the latest version of
the container image.
(
https://github.com/openshift/ovirt-csi-driver-operator/tree/master/assets)
After successfully creating pvc and pod, I tried to delete it.
And the pod is deleted, but the pvc is not deleted. This is because deleting a pod does
not unmap /dev/rbd0 attached to the ovirt vm.
How can I delete the pvc successfully?
oVirt engine version is 4.4.7.6-1.el8.
Here is the engine log when deleting the pod:
2021-08-20 17:40:35,385+09 INFO
[org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149) [] User
admin@internal-authz with profile [internal] successfully logged in with scopes:
ovirt-app-api ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search
ovirt-ext=token-info:validate ovirt-ext=token:password-access
2021-08-20 17:40:35,403+09 INFO [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand]
(default task-149) [68ee3182] Running command: CreateUserSessionCommand internal: false.
2021-08-20 17:40:35,517+09 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default task-149)
[68ee3182] EVENT_ID: USER_VDC_LOGIN(30), User admin@internal-authz connecting from
'192.168.7.169' using session
'XfDgNkmAGnPiZahK5itLhHQTCNHZ3JwXMMzOiZrYL3C32+1TTys3xcjrAmCIKPu02hgN1sdVpfZXWd0FznaPCQ=='
logged in.
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:35,520+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:35,663+09 INFO
[org.ovirt.engine.core.bll.storage.disk.DetachDiskFromVmCommand] (default task-149)
[198e2dc9-b908-474e-a395-0fe682c29af0] Running command: DetachDiskFromVmCommand internal:
false. Entities affected : ID: 59a7461c-72fe-4e01-86a7-c70243f31596 Type: VMAction group
CONFIGURE_VM_STORAGE with role type USER
2021-08-20 17:40:35,664+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149)
[198e2dc9-b908-474e-a395-0fe682c29af0] START, HotUnPlugDiskVDSCommand(HostName = host,
HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094',
vmId='59a7461c-72fe-4e01-86a7-c70243f31596',
diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4
2021-08-20 17:40:35,678+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149)
[198e2dc9-b908-474e-a395-0fe682c29af0] Disk hot-unplug: <?xml version="1.0"
encoding="UTF-8"?><hotunplug>
<devices>
<disk>
<alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/>
</disk>
</devices>
</hotunplug>
2021-08-20 17:40:35,749+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default task-149)
[198e2dc9-b908-474e-a395-0fe682c29af0] FINISH, HotUnPlugDiskVDSCommand, return: , log id:
506ff4a4
2021-08-20 17:40:35,842+09 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default task-149)
[198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID: USER_DETACH_DISK_FROM_VM(2,018), Disk
pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 was successfully detached from VM centos by
admin@internal-authz.
2021-08-20 17:40:35,916+09 ERROR [org.ovirt.engine.core.sso.service.SsoService] (default
task-150) [] OAuthException invalid_grant: The provided authorization grant for the auth
code has expired.
2021-08-20 17:40:35,917+09 ERROR [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter]
(default task-149) [] Cannot authenticate using authentication Headers: invalid_grant: The
provided authorization grant for the auth code has expired.
2021-08-20 17:40:36,029+09 INFO
[org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149) [] User
admin@internal-authz with profile [internal] successfully logged in with scopes:
ovirt-app-api ovirt-ext=token-info:authz-search ovirt-ext=token-info:public-authz-search
ovirt-ext=token-info:validate ovirt-ext=token:password-access
2021-08-20 17:40:36,046+09 INFO [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand]
(default task-149) [4c4bf441] Running command: CreateUserSessionCommand internal: false.
2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:36,114+09 WARN [org.ovirt.engine.api.restapi.util.LinkHelper] (default
task-149) [] Can't find relative path for class
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
return null
2021-08-20 17:40:49,241+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) [] START,
DumpXmlsVDSCommand(HostName = host,
Params:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094',
vmIds='[59a7461c-72fe-4e01-86a7-c70243f31596]'}), log id: 7eb54202
2021-08-20 17:40:49,244+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) [] FINISH,
DumpXmlsVDSCommand, return: {59a7461c-72fe-4e01-86a7-c70243f31596=<domain
type='kvm' id='1'
xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'>
<name>centos</name>
<uuid>59a7461c-72fe-4e01-86a7-c70243f31596</uuid>
<metadata
xmlns:ns1="http://ovirt.org/vm/tune/1.0"
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
<ns1:qos/>
<ovirt-vm:vm
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
<ovirt-vm:balloonTarget
type="int">4194304</ovirt-vm:balloonTarget>
<ovirt-vm:ballooningEnabled>true</ovirt-vm:ballooningEnabled>
<ovirt-vm:clusterVersion>4.6</ovirt-vm:clusterVersion>
<ovirt-vm:destroy_on_reboot
type="bool">False</ovirt-vm:destroy_on_reboot>
<ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
<ovirt-vm:memGuaranteedSize
type="int">4096</ovirt-vm:memGuaranteedSize>
<ovirt-vm:minGuaranteedMemoryMb
type="int">4096</ovirt-vm:minGuaranteedMemoryMb>
<ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
<ovirt-vm:startTime
type="float">1628558564.8754532</ovirt-vm:startTime>
<ovirt-vm:device alias="ua-7c9f38e9-8889-46c8-83bb-92efb9272de9"
mac_address="56:6f:16:a8:00:07">
<ovirt-vm:network>ovirtmgmt</ovirt-vm:network>
<ovirt-vm:custom>
<ovirt-vm:queues>2</ovirt-vm:queues>
</ovirt-vm:custom>
</ovirt-vm:device>
<ovirt-vm:device devtype="disk" name="sda">
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
<ovirt-vm:guestName>/dev/sda</ovirt-vm:guestName>
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
<ovirt-vm:managed type="bool">False</ovirt-vm:managed>
<ovirt-vm:poolID>4ca6e0e8-e3a4-11eb-8830-480fcf63834f</ovirt-vm:poolID>
<ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
<ovirt-vm:volumeChain>
<ovirt-vm:volumeChainNode>
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
<ovirt-vm:leaseOffset
type="int">0</ovirt-vm:leaseOffset>
<ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003.lease</ovirt-vm:leasePath>
<ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:path>
<ovirt-vm:volumeID>4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:volumeID>
</ovirt-vm:volumeChainNode>
<ovirt-vm:volumeChainNode>
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
<ovirt-vm:leaseOffset
type="int">0</ovirt-vm:leaseOffset>
<ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4.lease</ovirt-vm:leasePath>
<ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:path>
<ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
</ovirt-vm:volumeChainNode>
</ovirt-vm:volumeChain>
</ovirt-vm:device>
<ovirt-vm:device devtype="disk" name="sdc">
<ovirt-vm:managed type="bool">False</ovirt-vm:managed>
</ovirt-vm:device>
</ovirt-vm:vm>
</metadata>
<maxMemory slots='16' unit='KiB'>16777216</maxMemory>
<memory unit='KiB'>4194304</memory>
<currentMemory unit='KiB'>4194304</currentMemory>
<vcpu placement='static' current='2'>16</vcpu>
<resource>
<partition>/machine</partition>
</resource>
<sysinfo type='smbios'>
<system>
<entry name='manufacturer'>oVirt</entry>
<entry name='product'>RHEL</entry>
<entry name='version'>8.4-1.2105.el8</entry>
<entry
name='serial'>83e66af8-0500-11e6-9c43-bc00007c0000</entry>
<entry
name='uuid'>59a7461c-72fe-4e01-86a7-c70243f31596</entry>
<entry name='family'>oVirt</entry>
</system>
</sysinfo>
<os>
<type arch='x86_64'
machine='pc-q35-rhel8.4.0'>hvm</type>
<boot dev='hd'/>
<smbios mode='sysinfo'/>
</os>
<features>
<acpi/>
</features>
<cpu mode='custom' match='exact' check='full'>
<model fallback='forbid'>Skylake-Client-noTSX-IBRS</model>
<topology sockets='16' dies='1' cores='1'
threads='1'/>
<feature policy='require' name='ssbd'/>
<feature policy='require' name='md-clear'/>
<feature policy='disable' name='mpx'/>
<feature policy='require' name='hypervisor'/>
<numa>
<cell id='0' cpus='0-15' memory='4194304'
unit='KiB'/>
</numa>
</cpu>
<clock offset='variable' adjustment='0' basis='utc'>
<timer name='rtc' tickpolicy='catchup'/>
<timer name='pit' tickpolicy='delay'/>
<timer name='hpet' present='no'/>
</clock>
<on_poweroff>destroy</on_poweroff>
<on_reboot>restart</on_reboot>
<on_crash>destroy</on_crash>
<pm>
<suspend-to-mem enabled='no'/>
<suspend-to-disk enabled='no'/>
</pm>
<devices>
<emulator>/usr/libexec/qemu-kvm</emulator>
<disk type='file' device='cdrom'>
<driver name='qemu' error_policy='report'/>
<source startupPolicy='optional'/>
<target dev='sdc' bus='sata'/>
<readonly/>
<alias name='ua-df0ac774-3623-4868-8bd3-45c8f2aa3dc4'/>
<address type='drive' controller='0' bus='0'
target='0' unit='2'/>
</disk>
<disk type='file' device='disk' snapshot='no'>
<driver name='qemu' type='qcow2' cache='none'
error_policy='stop' io='threads'/>
<source
file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4'
index='8'>
<seclabel model='dac' relabel='no'/>
</source>
<backingStore type='file' index='1'>
<format type='qcow2'/>
<source
file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003'>
<seclabel model='dac' relabel='no'/>
</source>
<backingStore/>
</backingStore>
<target dev='sda' bus='scsi'/>
<serial>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</serial>
<alias name='ua-bee44276-234f-4ed7-8a8a-d90a5e3cb5b3'/>
<address type='drive' controller='0' bus='0'
target='0' unit='0'/>
</disk>
<controller type='pci' index='1'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='1' port='0x10'/>
<alias name='pci.1'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x0' multifunction='on'/>
</controller>
<controller type='pci' index='2'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='2' port='0x11'/>
<alias name='pci.2'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x1'/>
</controller>
<controller type='pci' index='3'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='3' port='0x12'/>
<alias name='pci.3'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x2'/>
</controller>
<controller type='pci' index='4'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='4' port='0x13'/>
<alias name='pci.4'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x3'/>
</controller>
<controller type='pci' index='5'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='5' port='0x14'/>
<alias name='pci.5'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x4'/>
</controller>
<controller type='pci' index='6'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='6' port='0x15'/>
<alias name='pci.6'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x5'/>
</controller>
<controller type='pci' index='7'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='7' port='0x16'/>
<alias name='pci.7'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x6'/>
</controller>
<controller type='pci' index='8'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='8' port='0x17'/>
<alias name='pci.8'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x7'/>
</controller>
<controller type='pci' index='9'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='9' port='0x18'/>
<alias name='pci.9'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x0' multifunction='on'/>
</controller>
<controller type='pci' index='10'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='10' port='0x19'/>
<alias name='pci.10'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x1'/>
</controller>
<controller type='pci' index='11'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='11' port='0x1a'/>
<alias name='pci.11'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x2'/>
</controller>
<controller type='pci' index='12'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='12' port='0x1b'/>
<alias name='pci.12'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x3'/>
</controller>
<controller type='pci' index='13'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='13' port='0x1c'/>
<alias name='pci.13'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x4'/>
</controller>
<controller type='pci' index='14'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='14' port='0x1d'/>
<alias name='pci.14'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x5'/>
</controller>
<controller type='pci' index='15'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='15' port='0x1e'/>
<alias name='pci.15'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x6'/>
</controller>
<controller type='pci' index='16'
model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='16' port='0x1f'/>
<alias name='pci.16'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x7'/>
</controller>
<controller type='pci' index='0' model='pcie-root'>
<alias name='pcie.0'/>
</controller>
<controller type='sata' index='0'>
<alias name='ide'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x1f' function='0x2'/>
</controller>
<controller type='virtio-serial' index='0' ports='16'>
<alias name='virtio-serial0'/>
<address type='pci' domain='0x0000' bus='0x04'
slot='0x00' function='0x0'/>
</controller>
<controller type='scsi' index='0' model='virtio-scsi'>
<alias name='ua-82c49f93-c4e8-460b-bb7d-95db0e9d87a0'/>
<address type='pci' domain='0x0000' bus='0x02'
slot='0x00' function='0x0'/>
</controller>
<controller type='usb' index='0' model='qemu-xhci'
ports='8'>
<alias name='ua-ad56daea-edb1-45c7-a1ab-2a7db3aaeee2'/>
<address type='pci' domain='0x0000' bus='0x03'
slot='0x00' function='0x0'/>
</controller>
<interface type='bridge'>
<mac address='56:6f:16:a8:00:07'/>
<source bridge='ovirtmgmt'/>
<target dev='vnet0'/>
<model type='virtio'/>
<driver name='vhost' queues='2'/>
<filterref filter='vdsm-no-mac-spoofing'/>
<link state='up'/>
<mtu size='1500'/>
<alias name='ua-7c9f38e9-8889-46c8-83bb-92efb9272de9'/>
<address type='pci' domain='0x0000' bus='0x01'
slot='0x00' function='0x0'/>
</interface>
<channel type='unix'>
<source mode='bind'
path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.ovirt-guest-agent.0'/>
<target type='virtio' name='ovirt-guest-agent.0'
state='disconnected'/>
<alias name='channel0'/>
<address type='virtio-serial' controller='0' bus='0'
port='1'/>
</channel>
<channel type='unix'>
<source mode='bind'
path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.org.qemu.guest_agent.0'/>
<target type='virtio' name='org.qemu.guest_agent.0'
state='connected'/>
<alias name='channel1'/>
<address type='virtio-serial' controller='0' bus='0'
port='2'/>
</channel>
<channel type='spicevmc'>
<target type='virtio' name='com.redhat.spice.0'
state='disconnected'/>
<alias name='channel2'/>
<address type='virtio-serial' controller='0' bus='0'
port='3'/>
</channel>
<input type='tablet' bus='usb'>
<alias name='input0'/>
<address type='usb' bus='0' port='1'/>
</input>
<input type='mouse' bus='ps2'>
<alias name='input1'/>
</input>
<input type='keyboard' bus='ps2'>
<alias name='input2'/>
</input>
<graphics type='vnc' port='5900' autoport='yes'
listen='192.168.7.18' keymap='en-us'
passwdValidTo='1970-01-01T00:00:01'>
<listen type='network' address='192.168.7.18'
network='vdsm-ovirtmgmt'/>
</graphics>
<graphics type='spice' port='5901' tlsPort='5902'
autoport='yes' listen='192.168.7.18'
passwdValidTo='1970-01-01T00:00:01'>
<listen type='network' address='192.168.7.18'
network='vdsm-ovirtmgmt'/>
<channel name='main' mode='secure'/>
<channel name='display' mode='secure'/>
<channel name='inputs' mode='secure'/>
<channel name='cursor' mode='secure'/>
<channel name='playback' mode='secure'/>
<channel name='record' mode='secure'/>
<channel name='smartcard' mode='secure'/>
<channel name='usbredir' mode='secure'/>
</graphics>
<video>
<model type='qxl' ram='65536' vram='8192'
vgamem='16384' heads='1' primary='yes'/>
<alias name='ua-799f065a-b2b9-4e37-a502-f86c7cc8dc51'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x01' function='0x0'/>
</video>
<memballoon model='virtio'>
<stats period='5'/>
<alias name='ua-c2bfe0b9-065a-46b7-9b0b-ef7e0f699611'/>
<address type='pci' domain='0x0000' bus='0x05'
slot='0x00' function='0x0'/>
</memballoon>
<rng model='virtio'>
<backend model='random'>/dev/urandom</backend>
<alias name='ua-fa110e6b-5eed-4b4b-93d8-0ac5de08aa2e'/>
<address type='pci' domain='0x0000' bus='0x06'
slot='0x00' function='0x0'/>
</rng>
</devices>
<seclabel type='dynamic' model='selinux' relabel='yes'>
<label>system_u:system_r:svirt_t:s0:c437,c650</label>
<imagelabel>system_u:object_r:svirt_image_t:s0:c437,c650</imagelabel>
</seclabel>
<seclabel type='dynamic' model='dac' relabel='yes'>
<label>+107:+107</label>
<imagelabel>+107:+107</imagelabel>
</seclabel>
<qemu:capabilities>
<qemu:add capability='blockdev'/>
<qemu:add capability='incremental-backup'/>
</qemu:capabilities>
</domain>
}, log id: 7eb54202
Here is the engine log when deleting the pvc:
2021-08-20 17:43:12,964 - cinder.volume.driver - INFO - Driver hasn't implemented
_init_vendor_properties()
2021-08-20 17:43:12,990 - cinderlib-client - INFO - Deleting volume
'63a64445-1659-4d5f-8847-e7266e64b09e' [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
2021-08-20 17:43:28,856 - cinder.volume.drivers.rbd - WARNING - ImageBusy error raised
while deleting rbd volume. This may have been caused by a connection from a client that
has crashed and, if so, may be resolved by retrying the delete after 30 seconds has
elapsed.
2021-08-20 17:43:28,900 - cinderlib-client - ERROR - Failure occurred when trying to run
command 'delete_volume': ImageBusy error raised while deleting rbd volume. This
may have been caused by a connection from a client that has crashed and, if so, may be
resolved by retrying the delete after 30 seconds has elapsed.
[feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
2021-08-20 17:43:28,901 - cinder - CRITICAL - Unhandled error
Traceback (most recent call last):
File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line
1195, in delete_volume
_try_remove_volume(client, volume_name)
File "/usr/lib/python3.6/site-packages/cinder/utils.py", line 683, in
_wrapper
return r.call(f, *args, **kwargs)
File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 409, in
call
do = self.iter(retry_state=retry_state)
File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 368, in
iter
raise retry_exc.reraise()
File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 186, in
reraise
raise self.last_attempt.result()
File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 425, in result
return self.__get_result()
File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 384, in
__get_result
raise self._exception
File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 412, in
call
result = fn(*args, **kwargs)
File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line
1190, in _try_remove_volume
self.RBDProxy().remove(client.ioctx, volume_name)
File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in doit
result = proxy_call(self._autowrap, f, *args, **kwargs)
File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in
proxy_call
rv = execute(f, *args, **kwargs)
File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 129, in
execute
six.reraise(c, e, tb)
File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
raise value
File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 83, in
tworker
rv = meth(*args, **kwargs)
File "rbd.pyx", line 767, in rbd.RBD.remove
rbd.ImageBusy: [errno 16] RBD image is busy (error removing image)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "./cinderlib-client.py", line 170, in main
args.command(args)
File "./cinderlib-client.py", line 218, in delete_volume
vol.delete()
File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 494,
in delete
self._raise_with_resource()
File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 222,
in _raise_with_resource
six.reraise(*exc_info)
File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
raise value
File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 487,
in delete
self.backend.driver.delete_volume(self._ovo)
File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line
1205, in delete_volume
raise exception.VolumeIsBusy(msg, volume_name=volume_name)
cinder.exception.VolumeIsBusy: ImageBusy error raised while deleting rbd volume. This may
have been caused by a connection from a client that has crashed and, if so, may be
resolved by retrying the delete after 30 seconds has elapsed.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "./cinderlib-client.py", line 390, in <module>
sys.exit(main(sys.argv[1:]))
File "./cinderlib-client.py", line 176, in main
sys.stderr.write(traceback.format_exc(e))
File "/usr/lib64/python3.6/traceback.py", line 167, in format_exc
return "".join(format_exception(*sys.exc_info(), limit=limit,
chain=chain))
File "/usr/lib64/python3.6/traceback.py", line 121, in format_exception
type(value), value, tb, limit=limit).format(chain=chain))
File "/usr/lib64/python3.6/traceback.py", line 498, in __init__
_seen=_seen)
File "/usr/lib64/python3.6/traceback.py", line 509, in __init__
capture_locals=capture_locals)
File "/usr/lib64/python3.6/traceback.py", line 338, in extract
if limit >= 0:
TypeError: '>=' not supported between instances of 'VolumeIsBusy' and
'int'
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AFYZX2HJZ3R...