Hi,
We setup a new iSCSI storage with 16TB, plenty of space.
Install a new VM (Ubuntu 24) with only one disk, 50GB, thin-provisioning
After some minutes it pauses.
Move the disk to another storage and we were able to start the VM. Don´t no for how long.
Running on CentOS 8, Version 4.5.4-1.el8
Did the same test in another environment:
CentOS 9, Version 4.5.6-1.el9, whitout any problem
# vdsm-tool dump-volume-chains 2ae7fdc6-d3e4-4b61-9cec-b7d819a4fb02
Images volume chains (base volume first)
image: dcd253fe-890d-4afd-9cb2-d45800f447f7
- 03c20fe2-1274-4873-a0b4-7d751e666806
status: OK, voltype: LEAF, format: RAW, legality: LEGAL, type: PREALLOCATED, capacity:
134217728, truesize: 134217728
image: 05f63799-db5e-424b-bb75-822803867b83
- 1ecda185-9639-49c2-ac0d-3acf47262025
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: SPARSE, capacity:
268435456000, truesize: 225620000768
image: a78ca2e3-0ea3-461b-ac1a-b589d082e914
- 37120c67-919f-4f6d-afa8-0fff300308c5
status: OK, voltype: INTERNAL, format: COW, legality: LEGAL, type: SPARSE, capacity:
21474836480, truesize: 20266876928
- 5879b376-f8ae-492b-b5c3-b6327abc0d4e
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: SPARSE, capacity:
21474836480, truesize: 8858370048
image: 4aaf8e07-077e-423b-a1ce-75c49a0af940
- 46fda312-d6b1-49ef-adbe-04f307419c36
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: SPARSE, capacity:
322122547200, truesize: 55431921664
image: fa96460a-6f43-497c-bf53-46544aebfe13
- 4ca5a23b-dc31-4102-bae5-e1241d3df8a6
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: SPARSE, capacity:
42949672960, truesize: 47110422528
image: f58f989f-24fe-4731-a27f-e5cd12350121
- 513e54c7-8087-4b56-ab0d-5c675b78288d
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: SPARSE, capacity:
26843545600, truesize: 14495514624
image: 054d77dc-9067-4a33-bac4-48e688a3c6c7
- 7f36baf0-41f5-46b9-b100-be9fabc956be
status: OK, voltype: LEAF, format: RAW, legality: LEGAL, type: PREALLOCATED, capacity:
134217728, truesize: 134217728
image: 81885b60-494a-4909-b479-9b40ccaeac45
- aa2dcff2-e5a9-45b5-a436-19f758ff83da
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: PREALLOCATED, capacity:
322122547200, truesize: 322122547200
image: 0254116f-edca-45b1-bfdb-5c9933b7fea6
- bbbe77e1-eab2-4a23-ae84-6d02a4f3824f
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: SPARSE, capacity:
268435456000, truesize: 162537668608
image: 62ed0783-33e2-4871-ad73-8ed66c7d076a
- c313d2db-7b4c-4d90-912a-175c0c724a6d
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: SPARSE, capacity:
268435456000, truesize: 171261820928
image: dc2ca30c-5c2d-4d34-9ed0-acfde4d87b2a
- db355623-8588-4c8c-97d1-43bffcea73b7
status: OK, voltype: LEAF, format: COW, legality: LEGAL, type: PREALLOCATED, capacity:
85899345920, truesize: 85899345920
# virsh --readonly dumpxml Teste_Disco
setlocale: No such file or directory
<domain type='kvm' id='99'
xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'>
<name>Teste_Disco</name>
<uuid>818a5c26-6581-4569-852c-23e94ebbd5e7</uuid>
<metadata
xmlns:ns1="http://ovirt.org/vm/tune/1.0"
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
<ns1:qos/>
<ovirt-vm:vm
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
<ovirt-vm:balloonTarget type="int">1048576</ovirt-vm:balloonTarget>
<ovirt-vm:ballooningEnabled>true</ovirt-vm:ballooningEnabled>
<ovirt-vm:clusterVersion>4.7</ovirt-vm:clusterVersion>
<ovirt-vm:cpuPolicy>none</ovirt-vm:cpuPolicy>
<ovirt-vm:destroy_on_reboot
type="bool">False</ovirt-vm:destroy_on_reboot>
<ovirt-vm:jobs>{}</ovirt-vm:jobs>
<ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
<ovirt-vm:memGuaranteedSize
type="int">1024</ovirt-vm:memGuaranteedSize>
<ovirt-vm:minGuaranteedMemoryMb
type="int">1024</ovirt-vm:minGuaranteedMemoryMb>
<ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
<ovirt-vm:startTime
type="float">1733399150.5427186</ovirt-vm:startTime>
<ovirt-vm:device alias="ua-01640387-95f2-452b-905f-1ab4c8840324"
mac_address="56:6f:fc:b2:00:1f">
<ovirt-vm:network>ZimbraOUT</ovirt-vm:network>
</ovirt-vm:device>
<ovirt-vm:device devtype="disk" name="sda">
<ovirt-vm:domainID>efebb0ca-83a5-40b0-8cf9-1ee1b50674c8</ovirt-vm:domainID>
<ovirt-vm:imageID>006e28f0-19c9-4457-b050-62ff257625e9</ovirt-vm:imageID>
<ovirt-vm:managed type="bool">False</ovirt-vm:managed>
<ovirt-vm:poolID>21979e6a-900f-11ed-80d3-000c29a8c80d</ovirt-vm:poolID>
<ovirt-vm:volumeID>f77445e9-fcd7-4a73-bb1f-b5e49b1d6a36</ovirt-vm:volumeID>
<ovirt-vm:volumeChain>
<ovirt-vm:volumeChainNode>
<ovirt-vm:domainID>efebb0ca-83a5-40b0-8cf9-1ee1b50674c8</ovirt-vm:domainID>
<ovirt-vm:imageID>006e28f0-19c9-4457-b050-62ff257625e9</ovirt-vm:imageID>
<ovirt-vm:leaseOffset type="int">111149056</ovirt-vm:leaseOffset>
<ovirt-vm:leasePath>/dev/efebb0ca-83a5-40b0-8cf9-1ee1b50674c8/leases</ovirt-vm:leasePath>
<ovirt-vm:path>/rhev/data-center/mnt/blockSD/efebb0ca-83a5-40b0-8cf9-1ee1b50674c8/images/006e28f0-19c9-4457-b050-62ff257625e9/aef15109-4c9f-45aa-b69a-7b6d95d85222</ovirt-vm:path>
<ovirt-vm:volumeID>aef15109-4c9f-45aa-b69a-7b6d95d85222</ovirt-vm:volumeID>
</ovirt-vm:volumeChainNode>
<ovirt-vm:volumeChainNode>
<ovirt-vm:domainID>efebb0ca-83a5-40b0-8cf9-1ee1b50674c8</ovirt-vm:domainID>
<ovirt-vm:imageID>006e28f0-19c9-4457-b050-62ff257625e9</ovirt-vm:imageID>
<ovirt-vm:leaseOffset type="int">130023424</ovirt-vm:leaseOffset>
<ovirt-vm:leasePath>/dev/efebb0ca-83a5-40b0-8cf9-1ee1b50674c8/leases</ovirt-vm:leasePath>
<ovirt-vm:path>/rhev/data-center/mnt/blockSD/efebb0ca-83a5-40b0-8cf9-1ee1b50674c8/images/006e28f0-19c9-4457-b050-62ff257625e9/f77445e9-fcd7-4a73-bb1f-b5e49b1d6a36</ovirt-vm:path>
<ovirt-vm:volumeID>f77445e9-fcd7-4a73-bb1f-b5e49b1d6a36</ovirt-vm:volumeID>
</ovirt-vm:volumeChainNode>
</ovirt-vm:volumeChain>
</ovirt-vm:device>
<ovirt-vm:device devtype="cdrom" name="sdc">
<ovirt-vm:volumeID>371245b9-cd8e-44eb-8adc-31e9fbcc44ff</ovirt-vm:volumeID>
<ovirt-vm:volumeChain>
<ovirt-vm:volumeChainNode>
<ovirt-vm:domainID>efebb0ca-83a5-40b0-8cf9-1ee1b50674c8</ovirt-vm:domainID>
<ovirt-vm:imageID>3994e0c3-6e77-4c93-9120-6d95bbf8c2a0</ovirt-vm:imageID>
<ovirt-vm:leaseOffset type="int">127926272</ovirt-vm:leaseOffset>
<ovirt-vm:leasePath>/dev/efebb0ca-83a5-40b0-8cf9-1ee1b50674c8/leases</ovirt-vm:leasePath>
<ovirt-vm:path>/rhev/data-center/mnt/blockSD/efebb0ca-83a5-40b0-8cf9-1ee1b50674c8/images/3994e0c3-6e77-4c93-9120-6d95bbf8c2a0/371245b9-cd8e-44eb-8adc-31e9fbcc44ff</ovirt-vm:path>
<ovirt-vm:volumeID>371245b9-cd8e-44eb-8adc-31e9fbcc44ff</ovirt-vm:volumeID>
</ovirt-vm:volumeChainNode>
</ovirt-vm:volumeChain>
</ovirt-vm:device>
</ovirt-vm:vm>
</metadata>
<maxMemory slots='16' unit='KiB'>4194304</maxMemory>
<memory unit='KiB'>1048576</memory>
<currentMemory unit='KiB'>1048576</currentMemory>
<vcpu placement='static' current='1'>16</vcpu>
<iothreads>1</iothreads>
<cputune>
<vcpupin vcpu='0' cpuset='0-47'/>
</cputune>
<resource>
<partition>/machine</partition>
</resource>
<sysinfo type='smbios'>
<system>
<entry name='manufacturer'>oVirt</entry>
<entry name='product'>RHEL</entry>
<entry name='version'>8.7.2206.0-1.el8</entry>
<entry name='serial'>32353550-3135-5a43-4a32-343530483757</entry>
<entry name='uuid'>818a5c26-6581-4569-852c-23e94ebbd5e7</entry>
<entry name='family'>oVirt</entry>
</system>
</sysinfo>
<os>
<type arch='x86_64' machine='pc-q35-rhel8.6.0'>hvm</type>
<loader readonly='yes' secure='no'
type='pflash'>/usr/share/OVMF/OVMF_CODE.secboot.fd</loader>
<nvram
template='/usr/share/OVMF/OVMF_VARS.fd'>/var/lib/libvirt/qemu/nvram/818a5c26-6581-4569-852c-23e94ebbd5e7.fd</nvram>
<bootmenu enable='yes' timeout='30000'/>
<smbios mode='sysinfo'/>
</os>
<features>
<acpi/>
<vmcoreinfo state='on'/>
</features>
<cpu mode='custom' match='exact' check='full'>
<model fallback='forbid'>EPYC</model>
<topology sockets='16' dies='1' cores='1'
threads='1'/>
<feature policy='require' name='ibpb'/>
<feature policy='require' name='virt-ssbd'/>
<feature policy='disable' name='monitor'/>
<feature policy='require' name='x2apic'/>
<feature policy='require' name='hypervisor'/>
<feature policy='disable' name='svm'/>
<feature policy='require' name='topoext'/>
<numa>
<cell id='0' cpus='0-15' memory='1048576'
unit='KiB'/>
</numa>
</cpu>
<clock offset='variable' adjustment='0' basis='utc'>
<timer name='rtc' tickpolicy='catchup'/>
<timer name='pit' tickpolicy='delay'/>
<timer name='hpet' present='no'/>
</clock>
<on_poweroff>destroy</on_poweroff>
<on_reboot>restart</on_reboot>
<on_crash>destroy</on_crash>
<pm>
<suspend-to-mem enabled='no'/>
<suspend-to-disk enabled='no'/>
</pm>
<devices>
<emulator>/usr/libexec/qemu-kvm</emulator>
<disk type='file' device='cdrom'>
<driver name='qemu' type='raw' error_policy='report'/>
<source index='3'/>
<target dev='sdc' bus='sata'/>
<readonly/>
<boot order='1'/>
<alias name='ua-c3c8ded0-7229-441e-a6ad-f2854c9104f3'/>
<address type='drive' controller='0' bus='0' target='0'
unit='2'/>
</disk>
<disk type='block' device='disk' snapshot='no'>
<driver name='qemu' type='qcow2' cache='none'
error_policy='stop' io='native'/>
<source
dev='/rhev/data-center/mnt/blockSD/efebb0ca-83a5-40b0-8cf9-1ee1b50674c8/images/006e28f0-19c9-4457-b050-62ff257625e9/f77445e9-fcd7-4a73-bb1f-b5e49b1d6a36'
index='5'>
<seclabel model='dac' relabel='no'/>
</source>
<backingStore type='block' index='6'>
<format type='qcow2'/>
<source
dev='/rhev/data-center/mnt/blockSD/efebb0ca-83a5-40b0-8cf9-1ee1b50674c8/images/006e28f0-19c9-4457-b050-62ff257625e9/aef15109-4c9f-45aa-b69a-7b6d95d85222'>
<seclabel model='dac' relabel='no'/>
</source>
<backingStore/>
</backingStore>
<target dev='sda' bus='scsi'/>
<serial>006e28f0-19c9-4457-b050-62ff257625e9</serial>
<boot order='2'/>
<alias name='ua-006e28f0-19c9-4457-b050-62ff257625e9'/>
<address type='drive' controller='0' bus='0' target='0'
unit='0'/>
</disk>
<controller type='usb' index='0' model='qemu-xhci'
ports='8'>
<alias name='ua-6ea483c0-d13d-43bb-9cdc-95c4219b58bd'/>
<address type='pci' domain='0x0000' bus='0x02'
slot='0x00' function='0x0'/>
</controller>
<controller type='virtio-serial' index='0' ports='16'>
<alias name='ua-8ac37f08-a78b-404c-9edb-b2cba37dfab4'/>
<address type='pci' domain='0x0000' bus='0x03'
slot='0x00' function='0x0'/>
</controller>
<controller type='scsi' index='0' model='virtio-scsi'>
<driver iothread='1'/>
<alias name='ua-9cf06f2f-96e0-47b1-977d-eaaab78b501b'/>
<address type='pci' domain='0x0000' bus='0x04'
slot='0x00' function='0x0'/>
</controller>
<controller type='pci' index='0' model='pcie-root'>
<alias name='pcie.0'/>
</controller>
<controller type='pci' index='1' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='1' port='0x10'/>
<alias name='pci.1'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x0' multifunction='on'/>
</controller>
<controller type='pci' index='2' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='2' port='0x11'/>
<alias name='pci.2'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x1'/>
</controller>
<controller type='pci' index='3' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='3' port='0x12'/>
<alias name='pci.3'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x2'/>
</controller>
<controller type='pci' index='4' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='4' port='0x13'/>
<alias name='pci.4'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x3'/>
</controller>
<controller type='pci' index='5' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='5' port='0x14'/>
<alias name='pci.5'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x4'/>
</controller>
<controller type='pci' index='6' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='6' port='0x15'/>
<alias name='pci.6'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x5'/>
</controller>
<controller type='pci' index='7' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='7' port='0x16'/>
<alias name='pci.7'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x6'/>
</controller>
<controller type='pci' index='8' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='8' port='0x17'/>
<alias name='pci.8'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x02' function='0x7'/>
</controller>
<controller type='pci' index='9' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='9' port='0x18'/>
<alias name='pci.9'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x0' multifunction='on'/>
</controller>
<controller type='pci' index='10' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='10' port='0x19'/>
<alias name='pci.10'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x1'/>
</controller>
<controller type='pci' index='11' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='11' port='0x1a'/>
<alias name='pci.11'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x2'/>
</controller>
<controller type='pci' index='12' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='12' port='0x1b'/>
<alias name='pci.12'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x3'/>
</controller>
<controller type='pci' index='13' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='13' port='0x1c'/>
<alias name='pci.13'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x4'/>
</controller>
<controller type='pci' index='14' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='14' port='0x1d'/>
<alias name='pci.14'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x5'/>
</controller>
<controller type='pci' index='15' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='15' port='0x1e'/>
<alias name='pci.15'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x6'/>
</controller>
<controller type='pci' index='16' model='pcie-root-port'>
<model name='pcie-root-port'/>
<target chassis='16' port='0x1f'/>
<alias name='pci.16'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x03' function='0x7'/>
</controller>
<controller type='sata' index='0'>
<alias name='ide'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x1f' function='0x2'/>
</controller>
<interface type='bridge'>
<mac address='56:6f:fc:b2:00:1f'/>
<source bridge='ZimbraOUT'/>
<target dev='vnet98'/>
<model type='virtio'/>
<filterref filter='vdsm-no-mac-spoofing'/>
<link state='up'/>
<mtu size='1500'/>
<boot order='3'/>
<alias name='ua-01640387-95f2-452b-905f-1ab4c8840324'/>
<address type='pci' domain='0x0000' bus='0x01'
slot='0x00' function='0x0'/>
</interface>
<channel type='unix'>
<source mode='bind'
path='/var/lib/libvirt/qemu/channels/818a5c26-6581-4569-852c-23e94ebbd5e7.ovirt-guest-agent.0'/>
<target type='virtio' name='ovirt-guest-agent.0'
state='disconnected'/>
<alias name='channel0'/>
<address type='virtio-serial' controller='0' bus='0'
port='1'/>
</channel>
<channel type='unix'>
<source mode='bind'
path='/var/lib/libvirt/qemu/channels/818a5c26-6581-4569-852c-23e94ebbd5e7.org.qemu.guest_agent.0'/>
<target type='virtio' name='org.qemu.guest_agent.0'
state='disconnected'/>
<alias name='channel1'/>
<address type='virtio-serial' controller='0' bus='0'
port='2'/>
</channel>
<input type='tablet' bus='usb'>
<alias name='input0'/>
<address type='usb' bus='0' port='1'/>
</input>
<input type='mouse' bus='ps2'>
<alias name='input1'/>
</input>
<input type='keyboard' bus='ps2'>
<alias name='input2'/>
</input>
<graphics type='vnc' port='5912' autoport='yes'
listen='192.168.6.112' keymap='pt'
passwdValidTo='2024-12-05T11:32:23'>
<listen type='network' address='192.168.6.112'
network='vdsm-ovirtmgmt'/>
</graphics>
<audio id='1' type='none'/>
<video>
<model type='virtio' vram='16384' heads='1'
primary='yes'/>
<alias name='ua-1a3762f1-ff7f-4fa9-9681-c6fea475a8dd'/>
<address type='pci' domain='0x0000' bus='0x00'
slot='0x01' function='0x0'/>
</video>
<memballoon model='virtio'>
<stats period='5'/>
<alias name='ua-8304854e-731b-4b52-a588-8bf38cd8b804'/>
<address type='pci' domain='0x0000' bus='0x05'
slot='0x00' function='0x0'/>
</memballoon>
<rng model='virtio'>
<backend model='random'>/dev/urandom</backend>
<alias name='ua-52a31dc2-d4bb-4315-8706-8200f80ac55c'/>
<address type='pci' domain='0x0000' bus='0x06'
slot='0x00' function='0x0'/>
</rng>
</devices>
<seclabel type='dynamic' model='selinux' relabel='yes'>
<label>system_u:system_r:svirt_t:s0:c146,c484</label>
<imagelabel>system_u:object_r:svirt_image_t:s0:c146,c484</imagelabel>
</seclabel>
<seclabel type='dynamic' model='dac' relabel='yes'>
<label>+107:+107</label>
<imagelabel>+107:+107</imagelabel>
</seclabel>
<qemu:capabilities>
<qemu:add capability='blockdev'/>
<qemu:add capability='incremental-backup'/>
</qemu:capabilities>
</domain>
Thanks
José
De: "Jean-Louis Dupond" <jean-louis(a)dupond.be>
Para: "suporte" <suporte(a)logicworks.pt>
Cc: "Colin Coe" <colin.coe(a)gmail.com>, "users"
<users(a)ovirt.org>
Itens enviados: Terça-feira, 5 de Novembro de 2024 11:04:12
Assunto: Re: [ovirt-users] Re: VM has been paused due to no Storage space error.
Hi,
The information is really unclear.
In previous emails you state you use preallocated disks, now you tell us it only happens
on thin provisioned disks.
Also the VM disk was 300GB, and now its 80GB.
How did you check the LV size? Can you show me the command + output.
What does: vdsm-tool dump-volume-chains storage-id give for the disk that has the issue?
The VM has 1 or more disks?
Can you also do a virsh --readonly dumpxml <vm> on the hypervisor where the vm runs
and give the output?
Thanks
Jean-Louis
On 5/11/2024 11:55, [ mailto:suporte@logicworks.pt | suporte(a)logicworks.pt ] wrote:
Hi,
Can't find any error on vdsm log. What should I look for?
The size of the Logical Volume for that disk is 80GB
We have plenty free space.
The VM cannot start again. it only happens with thin provision disks. If I move the disk
of the paused VM to another storage it works for a day and than it pauses again.
Thanks
José
De: "Jean-Louis Dupond" [ mailto:jean-louis@dupond.be |
<jean-louis(a)dupond.be> ]
Para: "suporte" [ mailto:suporte@logicworks.pt | <suporte(a)logicworks.pt> ]
, "Colin Coe" [ mailto:colin.coe@gmail.com | <colin.coe(a)gmail.com> ]
Cc: "users" [ mailto:users@ovirt.org | <users(a)ovirt.org> ]
Itens enviados: Segunda-feira, 4 de Novembro de 2024 12:14:30
Assunto: Re: [ovirt-users] Re: VM has been paused due to no Storage space error.
Hi,
I think you need to dig deeper into the issue.
- What does vdsm logs show for example?
- What is the size of the Logical Volume for that disk?
- The Volume Group (checked on that hypervisor) still gives free space?
- There is no way to start the VM again without getting it paused?
Jean-Louis
On 4/11/2024 13:09, [ mailto:suporte@logicworks.pt | suporte(a)logicworks.pt ] wrote:
BQ_BEGIN
But we are in Version 4.5.4-1.el8, so still don't know.
From: "suporte" [ mailto:suporte@logicworks.pt | <suporte(a)logicworks.pt> ]
To: "Colin Coe" [ mailto:colin.coe@gmail.com | <colin.coe(a)gmail.com> ]
Cc: "Jean-Louis Dupond" [ mailto:jean-louis@dupond.be |
<jean-louis(a)dupond.be> ] , "users" [ mailto:users@ovirt.org |
<users(a)ovirt.org> ]
Sent: Monday, November 4, 2024 12:08:11 PM
Subject: Re: [ovirt-users] Re: VM has been paused due to no Storage space error.
if I understand correctly, upgrading will solve the issue!
From: "Colin Coe" [ mailto:colin.coe@gmail.com | <colin.coe(a)gmail.com> ]
To: "suporte" [ mailto:suporte@logicworks.pt | <suporte(a)logicworks.pt> ]
Cc: "Jean-Louis Dupond" [ mailto:jean-louis@dupond.be |
<jean-louis(a)dupond.be> ] , "users" [ mailto:users@ovirt.org |
<users(a)ovirt.org> ]
Sent: Monday, November 4, 2024 12:04:19 PM
Subject: Re: [ovirt-users] Re: VM has been paused due to no Storage space error.
I don't think this article will help you. It specifically talks about upgrading to RHV
4.4SP1 (which is oVirt 4.5). There is a config workaround but it says don't do this in
RHV 4.4SP1.
On Mon, 4 Nov 2024 at 19:30, José Ferradeira via Users < [ mailto:users@ovirt.org |
users(a)ovirt.org ] > wrote:
BQ_BEGIN
Hi,
I found this: [
https://access.redhat.com/solutions/130843 |
https://access.redhat.com/solutions/130843 ]
But have no access.
Any help?
Thanks
José
From: "José Ferradeira via Users" < [ mailto:users@ovirt.org |
users(a)ovirt.org ] >
To: "Jean-Louis Dupond" < [ mailto:jean-louis@dupond.be |
jean-louis(a)dupond.be ] >
Cc: "users" < [ mailto:users@ovirt.org | users(a)ovirt.org ] >
Sent: Wednesday, August 21, 2024 11:27:47 AM
Subject: [ovirt-users] Re: VM has been paused due to no Storage space error.
Hi,
Ok, discard is not enabled on the disk.
Reading [
https://gitlab.com/qemu-project/qemu/-/issues/1621 |
https://gitlab.com/qemu-project/qemu/-/issues/1621 ]
I'm sure the image did not grow over 110% of its virtual size
Any idea?
Thanks
José
From: "Jean-Louis Dupond" < [ mailto:jean-louis@dupond.be |
jean-louis(a)dupond.be ] >
To: [ mailto:suporte@logicworks.pt | suporte(a)logicworks.pt ]
Cc: "users" < [ mailto:users@ovirt.org | users(a)ovirt.org ] >
Sent: Wednesday, August 21, 2024 10:17:22 AM
Subject: Re: [ovirt-users] VM has been paused due to no Storage space error.
Hi,
Not talking about discard on the storage domain, but on disk/vm level.
Is it enabled there?
Cause if that is the cause, it could be [
https://gitlab.com/qemu-project/qemu/-/issues/1621 |
https://gitlab.com/qemu-project/qemu/-/issues/1621 ]
On 21/08/2024 11:14, [ mailto:suporte@logicworks.pt | suporte(a)logicworks.pt ] wrote:
BQ_BEGIN
Hi Jean-Louis,
The VM disk is 300GB with 216GB available, preallocated on a iSCSI storage with 3532GB
total and 1932GB free.
Have enable incremental backup. Discard After Delete is not enabled.
Thanks
José
From: "Jean-Louis Dupond" [ mailto:jean-louis@dupond.be |
<jean-louis(a)dupond.be> ]
To: [ mailto:suporte@logicworks.pt | suporte(a)logicworks.pt ] , "users" [
mailto:users@ovirt.org | <users(a)ovirt.org> ]
Sent: Wednesday, August 21, 2024 9:12:53 AM
Subject: Re: [ovirt-users] VM has been paused due to no Storage space error.
Hi Jose,
How big are the disks? And what size are they on the storage?
I guess the images are qcow2?
Is discard enabled?
Jean-Louis
On 20/08/2024 13:42, José Ferradeira via Users wrote:
BQ_BEGIN
Hello,
running ovirt Version 4.5.4-1.el8 on Centos 8, randomly we have this error:
VM has been paused due to no Storage space error.
We have plenty of space on the iSCSI storage. This is a preallocated disk, VirtIO-SCSi.
No user interaction. It happens, so far, with 3 VM, Windows and Ubuntu.
This service was stopped: dnf-makecache.service
This is what I found on the engine log:
2024-08-19 01:04:35,522+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-25) [eb7e5f1] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'Up' -->
'Paused'
2024-08-19 01:04:35,665+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ForkJoinPool-1-worker-25) [eb7e5f1] EVENT_ID: VM_PAUSED_ENOSPC(138), VM Bravo has been
paused due to no Storage space error.
2024-08-19 09:26:35,855+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-29) [72482216] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'Paused' -->
'Down'
2024-08-19 09:26:48,114+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [72482216] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'WaitForLaunch'
--> 'PoweringUp'
2024-08-19 09:27:50,062+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-6) [] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'PoweringUp'
--> 'Up'
2024-08-19 09:29:25,145+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [72482216] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'Up' -->
'Paused'
2024-08-19 09:29:25,273+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ForkJoinPool-1-worker-15) [72482216] EVENT_ID: VM_PAUSED_ENOSPC(138), VM Bravo has been
paused due to no Storage space error.
2024-08-19 09:37:26,128+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [6d88f065] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'Paused' -->
'Down'
2024-08-19 09:41:43,300+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [6d88f065] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'WaitForLaunch'
--> 'PoweringUp'
2024-08-19 09:42:14,882+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-23) [6d88f065] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'PoweringUp'
--> 'Up'
2024-08-19 09:42:59,792+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [6d88f065] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'Up' -->
'Paused'
2024-08-19 09:42:59,894+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ForkJoinPool-1-worker-15) [6d88f065] EVENT_ID: VM_PAUSED_ENOSPC(138), VM Bravo has been
paused due to no Storage space error.
2024-08-19 09:45:30,334+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [6b3d8ee] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'Paused' -->
'Down'
2024-08-19 09:47:51,068+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [6b3d8ee] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'WaitForLaunch'
--> 'PoweringUp'
2024-08-19 09:48:50,710+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-80) [] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'PoweringUp'
--> 'Up'
2024-08-19 10:06:38,810+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [1dd98021] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'PoweringDown'
--> 'Down'
2024-08-19 10:08:11,606+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [1dd98021] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'WaitForLaunch'
--> 'PoweringUp'
2024-08-19 10:09:12,507+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-25) [] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'PoweringUp'
--> 'Up'
2024-08-19 10:21:13,835+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [63fa2421] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'Up' -->
'Down'
2024-08-19 10:25:19,302+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-15) [63fa2421] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'WaitForLaunch'
--> 'PoweringUp'
2024-08-19 10:26:05,456+01 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-3) [63fa2421] VM
'ccc65521-934d-4f77-adf3-9f9eeb83a4f8'(Bravo) moved from 'PoweringUp'
--> 'Up'
And we cannot start the VM anymore.
Any idea?
Thanks
--
Jose Ferradeira
[
http://www.logicworks.pt/ |
http://www.logicworks.pt ]
_______________________________________________
Users mailing list -- [ mailto:users@ovirt.org | users(a)ovirt.org ] To unsubscribe send an
email to [ mailto:users-leave@ovirt.org | users-leave(a)ovirt.org ] Privacy Statement: [
https://www.ovirt.org/privacy-policy.html |
https://www.ovirt.org/privacy-policy.html ]
oVirt Code of Conduct: [
https://www.ovirt.org/community/about/community-guidelines/ |
https://www.ovirt.org/community/about/community-guidelines/ ] List Archives: [
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YKELJXHKYLT...
|
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YKELJXHKYLT...
]
BQ_END
_______________________________________________
Users mailing list -- [ mailto:users@ovirt.org | users(a)ovirt.org ]
To unsubscribe send an email to [ mailto:users-leave@ovirt.org | users-leave(a)ovirt.org ]
Privacy Statement: [
https://www.ovirt.org/privacy-policy.html |
https://www.ovirt.org/privacy-policy.html ]
oVirt Code of Conduct: [
https://www.ovirt.org/community/about/community-guidelines/ |
https://www.ovirt.org/community/about/community-guidelines/ ]
List Archives: [
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OON4G6VUC6Z...
|
https://lists.ovirt.org/archives/list/users@ovirt.org/message/OON4G6VUC6Z...
]
_______________________________________________
Users mailing list -- [ mailto:users@ovirt.org | users(a)ovirt.org ]
To unsubscribe send an email to [ mailto:users-leave@ovirt.org | users-leave(a)ovirt.org ]
Privacy Statement: [
https://www.ovirt.org/privacy-policy.html |
https://www.ovirt.org/privacy-policy.html ]
oVirt Code of Conduct: [
https://www.ovirt.org/community/about/community-guidelines/ |
https://www.ovirt.org/community/about/community-guidelines/ ]
List Archives: [
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AWZVVV4TEB6...
|
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AWZVVV4TEB6...
]
BQ_END
BQ_END
BQ_END