The hosts will still mount the volume via FUSE, but you might double check you set the
storage up as Gluster and not NFS.
Then gluster used to need some config in glusterd.vol to set
option rpc-auth-allow-insecure on
I’m not sure if that got added to a hyper converged setup or not, but I’d check it.
On Feb 10, 2020, at 4:41 PM, Stephen Panicho
<s.panicho(a)gmail.com> wrote:
No, this was a relatively new cluster-- only a couple days old. Just a handful of VMs
including the engine.
On Mon, Feb 10, 2020 at 5:26 PM Jayme <jaymef(a)gmail.com
<mailto:jaymef@gmail.com>> wrote:
Curious do the vms have active snapshots?
On Mon, Feb 10, 2020 at 5:59 PM <s.panicho(a)gmail.com
<mailto:s.panicho@gmail.com>> wrote:
Hello, all. I have a 3-node Hyperconverged oVirt 4.3.8 cluster running on CentOS 7.7
hosts. I was investigating poor Gluster performance and heard about libgfapi, so I thought
I'd give it a shot. Looking through the documentation, followed by lots of threads and
BZ reports, I've done the following to enable it:
First, I shut down all VMs except the engine. Then...
On the hosts:
1. setsebool -P virt_use_glusterfs on
2. dynamic_ownership=0 in /etc/libvirt/qemu.conf
On the engine VM:
1. engine-config -s LibgfApiSupported=true --cver=4.3
2. systemctl restart ovirt-engine
VMs now fail to launch. Am I doing this correctly? I should also note that the hosts
still have the Gluster domain mounted via FUSE.
Here's a relevant bit from engine.log:
2020-02-06T16:38:32.573511Z qemu-kvm: -drive
file=gluster://node1.fs.trashnet.xyz:24007/vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78,file.debug=4,format=qcow2,if=none,id=drive-ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,serial=a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,werror=stop,rerror=stop,cache=none,discard=unmap,aio=native
<
http://node1.fs.trashnet.xyz:24007/vmstore/781717e5-1cff-43a1-b586-994150...;:
Could not read qcow2 header: Invalid argument.
The full engine.log from one of the attempts:
2020-02-06 16:38:24,909Z INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-12) [] add VM
'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache) to rerun treatment
2020-02-06 16:38:25,010Z ERROR [org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring]
(ForkJoinPool-1-worker-12) [] Rerun VM 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'.
Called from VDS 'node2.ovirt.trashnet.xyz
<
http://node2.ovirt.trashnet.xyz/>'
2020-02-06 16:38:25,091Z WARN
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engine-Thread-216) [] EVENT_ID:
USER_INITIATED_RUN_VM_FAILED(151), Failed to run VM yumcache on Host
node2.ovirt.trashnet.xyz <
http://node2.ovirt.trashnet.xyz/>.
2020-02-06 16:38:25,166Z INFO [org.ovirt.engine.core.bll.RunVmCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] Lock Acquired to object
'EngineLock:{exclusiveLocks='[df9dbac4-35c0-40ee-acd4-a1cfc959aa8b=VM]',
sharedLocks=''}'
2020-02-06 16:38:25,179Z INFO
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] START, IsVmDuringInitiatingVDSCommand(
IsVmDuringInitiatingVDSCommandParameters:{vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'}),
log id: 2107f52a
2020-02-06 16:38:25,181Z INFO
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, IsVmDuringInitiatingVDSCommand,
return: false, log id: 2107f52a
2020-02-06 16:38:25,298Z INFO [org.ovirt.engine.core.bll.RunVmCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] Running command: RunVmCommand internal:
false. Entities affected : ID: df9dbac4-35c0-40ee-acd4-a1cfc959aa8b Type: VMAction group
RUN_VM with role type USER
2020-02-06 16:38:25,313Z INFO [org.ovirt.engine.core.bll.utils.EmulatedMachineUtils]
(EE-ManagedThreadFactory-engine-Thread-216) [] Emulated machine 'pc-q35-rhel7.6.0'
which is different than that of the cluster is set for
'yumcache'(df9dbac4-35c0-40ee-acd4-a1cfc959aa8b)
2020-02-06 16:38:25,382Z INFO
[org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] START, UpdateVmDynamicDataVDSCommand(
UpdateVmDynamicDataVDSCommandParameters:{hostId='null',
vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b',
vmDynamic='org.ovirt.engine.core.common.businessentities.VmDynamic@9774a64'}), log
id: 4a83911f
2020-02-06 16:38:25,417Z INFO
[org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, UpdateVmDynamicDataVDSCommand,
return: , log id: 4a83911f
2020-02-06 16:38:25,418Z INFO [org.ovirt.engine.core.vdsbroker.CreateVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] START, CreateVDSCommand(
CreateVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452',
vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', vm='VM [yumcache]'}), log id:
5e07ba66
2020-02-06 16:38:25,420Z INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] START, CreateBrokerVDSCommand(HostName =
node1.ovirt.trashnet.xyz <
http://node1.ovirt.trashnet.xyz/>,
CreateVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452',
vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', vm='VM [yumcache]'}), log id:
1bfa03c4
2020-02-06 16:38:25,424Z INFO
[org.ovirt.engine.core.vdsbroker.builder.vminfo.VmInfoBuildUtils]
(EE-ManagedThreadFactory-engine-Thread-216) [] Kernel FIPS - Guid:
c3465ca2-395e-4c0c-b72e-b5b7153df452 fips: false
2020-02-06 16:38:25,435Z INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] VM <?xml version="1.0"
encoding="UTF-8"?><domain type="kvm"
xmlns:ovirt-tune="http://ovirt.org/vm/tune/1.0
<
http://ovirt.org/vm/tune/1.0>" xmlns:ovirt-vm="http://ovirt.org/vm/1.0
<
http://ovirt.org/vm/1.0>">
<name>yumcache</name>
<uuid>df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</uuid>
<memory>1048576</memory>
<currentMemory>1048576</currentMemory>
<iothreads>1</iothreads>
<maxMemory slots="16">4194304</maxMemory>
<vcpu current="1">16</vcpu>
<sysinfo type="smbios">
<system>
<entry name="manufacturer">oVirt</entry>
<entry name="product">OS-NAME:</entry>
<entry name="version">OS-VERSION:</entry>
<entry name="serial">HOST-SERIAL:</entry>
<entry
name="uuid">df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</entry>
</system>
</sysinfo>
<clock offset="variable" adjustment="0">
<timer name="rtc" tickpolicy="catchup"/>
<timer name="pit" tickpolicy="delay"/>
<timer name="hpet" present="no"/>
</clock>
<features>
<acpi/>
</features>
<cpu match="exact">
<model>EPYC</model>
<feature name="ibpb" policy="require"/>
<feature name="virt-ssbd" policy="require"/>
<topology cores="1" threads="1" sockets="16"/>
<numa>
<cell id="0" cpus="0" memory="1048576"/>
</numa>
</cpu>
<cputune/>
<devices>
<input type="tablet" bus="usb"/>
<channel type="unix">
<target type="virtio" name="ovirt-guest-agent.0"/>
<source mode="bind"
path="/var/lib/libvirt/qemu/channels/df9dbac4-35c0-40ee-acd4-a1cfc959aa8b.ovirt-guest-agent.0"/>
</channel>
<channel type="unix">
<target type="virtio" name="org.qemu.guest_agent.0"/>
<source mode="bind"
path="/var/lib/libvirt/qemu/channels/df9dbac4-35c0-40ee-acd4-a1cfc959aa8b.org.qemu.guest_agent.0"/>
</channel>
<controller type="pci" model="pcie-root-port"
index="1">
<address bus="0x00" domain="0x0000" function="0x0"
slot="0x02" type="pci" multifunction="on"/>
</controller>
<memballoon model="virtio">
<stats period="5"/>
<alias name="ua-27c77007-3a3c-4431-958d-90fd1c7257dd"/>
<address bus="0x05" domain="0x0000" function="0x0"
slot="0x00" type="pci"/>
</memballoon>
<controller type="pci" model="pcie-root-port"
index="2">
<address bus="0x00" domain="0x0000" function="0x1"
slot="0x02" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="9">
<address bus="0x00" domain="0x0000" function="0x0"
slot="0x03" type="pci" multifunction="on"/>
</controller>
<controller type="sata" index="0">
<address bus="0x00" domain="0x0000" function="0x2"
slot="0x1f" type="pci"/>
</controller>
<rng model="virtio">
<backend model="random">/dev/urandom</backend>
<alias name="ua-51960005-6b95-47e9-82a7-67d5e0d6cf8a"/>
</rng>
<controller type="pci" model="pcie-root-port"
index="6">
<address bus="0x00" domain="0x0000" function="0x5"
slot="0x02" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="15">
<address bus="0x00" domain="0x0000" function="0x6"
slot="0x03" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="13">
<address bus="0x00" domain="0x0000" function="0x4"
slot="0x03" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="7">
<address bus="0x00" domain="0x0000" function="0x6"
slot="0x02" type="pci"/>
</controller>
<graphics type="vnc" port="-1" autoport="yes"
passwd="*****" passwdValidTo="1970-01-01T00:00:01"
keymap="en-us">
<listen type="network" network="vdsm-ovirtmgmt"/>
</graphics>
<controller type="pci" model="pcie-root-port"
index="16">
<address bus="0x00" domain="0x0000" function="0x7"
slot="0x03" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="12">
<address bus="0x00" domain="0x0000" function="0x3"
slot="0x03" type="pci"/>
</controller>
<video>
<model type="qxl" vram="32768" heads="1"
ram="65536" vgamem="16384"/>
<alias name="ua-8a295e96-40c3-44de-a3b0-1c4a685a5473"/>
<address bus="0x00" domain="0x0000" function="0x0"
slot="0x01" type="pci"/>
</video>
<graphics type="spice" port="-1" autoport="yes"
passwd="*****" passwdValidTo="1970-01-01T00:00:01"
tlsPort="-1">
<channel name="main" mode="secure"/>
<channel name="inputs" mode="secure"/>
<channel name="cursor" mode="secure"/>
<channel name="playback" mode="secure"/>
<channel name="record" mode="secure"/>
<channel name="display" mode="secure"/>
<channel name="smartcard" mode="secure"/>
<channel name="usbredir" mode="secure"/>
<listen type="network" network="vdsm-ovirtmgmt"/>
</graphics>
<controller type="pci" model="pcie-root-port"
index="5">
<address bus="0x00" domain="0x0000" function="0x4"
slot="0x02" type="pci"/>
</controller>
<controller type="usb" model="qemu-xhci" index="0"
ports="8">
<address bus="0x02" domain="0x0000" function="0x0"
slot="0x00" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="4">
<address bus="0x00" domain="0x0000" function="0x3"
slot="0x02" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="3">
<address bus="0x00" domain="0x0000" function="0x2"
slot="0x02" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="11">
<address bus="0x00" domain="0x0000" function="0x2"
slot="0x03" type="pci"/>
</controller>
<controller type="scsi" model="virtio-scsi"
index="0">
<driver iothread="1"/>
<alias name="ua-d0bf6fcd-7aa2-4658-b7cc-3dac259b7ad2"/>
<address bus="0x03" domain="0x0000" function="0x0"
slot="0x00" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="8">
<address bus="0x00" domain="0x0000" function="0x7"
slot="0x02" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="14">
<address bus="0x00" domain="0x0000" function="0x5"
slot="0x03" type="pci"/>
</controller>
<controller type="pci" model="pcie-root-port"
index="10">
<address bus="0x00" domain="0x0000" function="0x1"
slot="0x03" type="pci"/>
</controller>
<controller type="virtio-serial" index="0"
ports="16">
<address bus="0x04" domain="0x0000" function="0x0"
slot="0x00" type="pci"/>
</controller>
<channel type="spicevmc">
<target type="virtio" name="com.redhat.spice.0"/>
</channel>
<controller type="pci" model="pcie-root"/>
<interface type="bridge">
<model type="virtio"/>
<link state="up"/>
<source bridge="vmnet"/>
<alias name="ua-ceda0ef6-9139-4e5c-8840-86fe344ecbd3"/>
<address bus="0x01" domain="0x0000" function="0x0"
slot="0x00" type="pci"/>
<mac address="56:6f:91:b9:00:05"/>
<mtu size="1500"/>
<filterref filter="vdsm-no-mac-spoofing"/>
<bandwidth/>
</interface>
<disk type="file" device="cdrom" snapshot="no">
<driver name="qemu" type="raw"
error_policy="report"/>
<source file="" startupPolicy="optional">
<seclabel model="dac" type="none"
relabel="no"/>
</source>
<target dev="sdc" bus="sata"/>
<readonly/>
<alias name="ua-bdf99844-2d02-411b-90bb-671ee26764cb"/>
<address bus="0" controller="0" unit="2"
type="drive" target="0"/>
</disk>
<disk snapshot="no" type="network"
device="disk">
<target dev="sda" bus="scsi"/>
<source protocol="gluster"
name="vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78">
<host name="node1.fs.trashnet.xyz
<
http://node1.fs.trashnet.xyz/>" port="0"/>
<seclabel model="dac" type="none"
relabel="no"/>
</source>
<driver name="qemu" discard="unmap" io="native"
type="qcow2" error_policy="stop" cache="none"/>
<alias name="ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4"/>
<address bus="0" controller="0" unit="0"
type="drive" target="0"/>
<boot order="1"/>
<serial>a1d56b14-6d72-4f46-a0aa-eb0870c36bc4</serial>
</disk>
<lease>
<key>df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</key>
<lockspace>781717e5-1cff-43a1-b586-9941503544e8</lockspace>
<target offset="6291456"
path="/rhev/data-center/mnt/glusterSD/node1.fs.trashnet.xyz:_vmstore/781717e5-1cff-43a1-b586-9941503544e8/dom_md/xleases"/>
</lease>
</devices>
<pm>
<suspend-to-disk enabled="no"/>
<suspend-to-mem enabled="no"/>
</pm>
<os>
<type arch="x86_64"
machine="pc-q35-rhel7.6.0">hvm</type>
<smbios mode="sysinfo"/>
</os>
<metadata>
<ovirt-tune:qos/>
<ovirt-vm:vm>
<ovirt-vm:minGuaranteedMemoryMb
type="int">512</ovirt-vm:minGuaranteedMemoryMb>
<ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion>
<ovirt-vm:custom/>
<ovirt-vm:device mac_address="56:6f:91:b9:00:05">
<ovirt-vm:custom/>
</ovirt-vm:device>
<ovirt-vm:device devtype="disk" name="sda">
<ovirt-vm:poolID>2ffaec76-462c-11ea-b155-00163e512202</ovirt-vm:poolID>
<ovirt-vm:volumeID>a2314816-7970-49ce-a80c-ab0d1cf17c78</ovirt-vm:volumeID>
<ovirt-vm:imageID>a1d56b14-6d72-4f46-a0aa-eb0870c36bc4</ovirt-vm:imageID>
<ovirt-vm:domainID>781717e5-1cff-43a1-b586-9941503544e8</ovirt-vm:domainID>
</ovirt-vm:device>
<ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
<ovirt-vm:resumeBehavior>kill</ovirt-vm:resumeBehavior>
</ovirt-vm:vm>
</metadata>
</domain>
2020-02-06 16:38:25,455Z INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, CreateBrokerVDSCommand, return: ,
log id: 1bfa03c4
2020-02-06 16:38:25,494Z INFO [org.ovirt.engine.core.vdsbroker.CreateVDSCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, CreateVDSCommand, return:
WaitForLaunch, log id: 5e07ba66
2020-02-06 16:38:25,495Z INFO [org.ovirt.engine.core.bll.RunVmCommand]
(EE-ManagedThreadFactory-engine-Thread-216) [] Lock freed to object
'EngineLock:{exclusiveLocks='[df9dbac4-35c0-40ee-acd4-a1cfc959aa8b=VM]',
sharedLocks=''}'
2020-02-06 16:38:25,533Z INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engine-Thread-216) [] EVENT_ID: USER_STARTED_VM(153), VM yumcache
was started by admin@internal-authz (Host: node1.ovirt.trashnet.xyz
<
http://node1.ovirt.trashnet.xyz/>).
2020-02-06 16:38:33,300Z INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-5) [] VM 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b' was
reported as Down on VDS
'c3465ca2-395e-4c0c-b72e-b5b7153df452'(node1.ovirt.trashnet.xyz
<
http://node1.ovirt.trashnet.xyz/>)
2020-02-06 16:38:33,301Z INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] (ForkJoinPool-1-worker-5) []
START, DestroyVDSCommand(HostName = node1.ovirt.trashnet.xyz
<
http://node1.ovirt.trashnet.xyz/>,
DestroyVmVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452',
vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', secondsToWait='0',
gracefully='false', reason='', ignoreNoVm='true'}), log id:
1f951ea9
2020-02-06 16:38:33,478Z INFO
[org.ovirt.engine.core.vdsbroker.monitoring.VmsStatisticsFetcher]
(EE-ManagedThreadFactory-engineScheduled-Thread-8) [] Fetched 2 VMs from VDS
'c3465ca2-395e-4c0c-b72e-b5b7153df452'
2020-02-06 16:38:33,545Z INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] (ForkJoinPool-1-worker-5) []
FINISH, DestroyVDSCommand, return: , log id: 1f951ea9
2020-02-06 16:38:33,546Z INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-5) [] VM 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache)
moved from 'WaitForLaunch' --> 'Down'
2020-02-06 16:38:33,623Z ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ForkJoinPool-1-worker-5) [] EVENT_ID: VM_DOWN_ERROR(119), VM yumcache is down with error.
Exit message: internal error: qemu unexpectedly closed the monitor: [2020-02-06
16:38:31.723977] E [MSGID: 108006] [afr-common.c:5323:__afr_handle_child_down_event]
0-vmstore-replicate-0: All subvolumes are down. Going offline until at least one of them
comes back up.
[2020-02-06 16:38:31.724765] I [io-stats.c:4027:fini] 0-vmstore: io-stats translator
unloaded
2020-02-06T16:38:32.573511Z qemu-kvm: -drive
file=gluster://node1.fs.trashnet.xyz:24007/vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78,file.debug=4,format=qcow2,if=none,id=drive-ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,serial=a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,werror=stop,rerror=stop,cache=none,discard=unmap,aio=native
<
http://node1.fs.trashnet.xyz:24007/vmstore/781717e5-1cff-43a1-b586-994150...;:
Could not read qcow2 header: Invalid argument.
2020-02-06 16:38:33,624Z INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-5) [] add VM
'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache) to rerun treatment
2020-02-06 16:38:33,796Z ERROR [org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring]
(ForkJoinPool-1-worker-5) [] Rerun VM 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'.
Called from VDS 'node1.ovirt.trashnet.xyz
<
http://node1.ovirt.trashnet.xyz/>'
2020-02-06 16:38:33,899Z WARN
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engine-Thread-223) [] EVENT_ID:
USER_INITIATED_RUN_VM_FAILED(151), Failed to run VM yumcache on Host
node1.ovirt.trashnet.xyz <
http://node1.ovirt.trashnet.xyz/>.
_______________________________________________
Users mailing list -- users(a)ovirt.org <mailto:users@ovirt.org>
To unsubscribe send an email to users-leave(a)ovirt.org
<mailto:users-leave@ovirt.org>
Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
<
https://www.ovirt.org/site/privacy-policy/>
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
<
https://www.ovirt.org/community/about/community-guidelines/>
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6GTBANZ4R44...
<
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6GTBANZ4R44...
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5EZYM4OADD6...