
On 06/21/14 16:57, Tiemen Ruiten wrote:
On 06/21/14 16:37, Tiemen Ruiten wrote:
Hello,
I've been struggling to set up an Ovirt cluster and am now bumping into this problem:
When I try to create a new (Gluster) storage domain, it fails to attach to the data center. The error on the node from vdsm.log:
Thread-13::DEBUG::2014-06-21 16:17:14,157::BindingXMLRPC::251::vds::(wrapper) client [192.168.10.119] flowID [6e44c0a3] Thread-13::DEBUG::2014-06-21 16:17:14,159::task::595::TaskManager.Task::(_updateState) Task=`97b78287-45d2-4d5a-8336-460987df3840`::moving from state init -> state preparing Thread-13::INFO::2014-06-21 16:17:14,160::logUtils::44::dispatcher::(wrapper) Run and protect: connectStorageServer(domType=7, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '', 'connection': '192.168.10.120:/vmimage', 'iqn': '', 'user': '', 'tpgt': '1', 'vfs_type': 'glusterfs', 'password': '******', 'id': '901b15ec-6b05-43c1-8a50-06b34c8ffdbd'}], options=None) Thread-13::DEBUG::2014-06-21 16:17:14,172::hsm::2340::Storage.HSM::(__prefetchDomains) glusterDomPath: glusterSD/* Thread-13::DEBUG::2014-06-21 16:17:14,185::hsm::2352::Storage.HSM::(__prefetchDomains) Found SD uuids: ('dc661957-c0c1-44ba-a5b9-e6558904207b',) Thread-13::DEBUG::2014-06-21 16:17:14,185::hsm::2408::Storage.HSM::(connectStorageServer) knownSDs: {dc661957-c0c1-44ba-a5b9-e6558904207b: storage.glusterSD.findDomain} Thread-13::INFO::2014-06-21 16:17:14,186::logUtils::47::dispatcher::(wrapper) Run and protect: connectStorageServer, Return response: {'statuslist': [{'status': 0, 'id': '901b15ec-6b05-43c1-8a50-06b34c8ffdbd'}]} Thread-13::DEBUG::2014-06-21 16:17:14,186::task::1185::TaskManager.Task::(prepare) Task=`97b78287-45d2-4d5a-8336-460987df3840`::finished: {'statuslist': [{'status': 0, 'id': '901b15ec-6b05-43c1-8a50-06b34c8ffdbd'}]} Thread-13::DEBUG::2014-06-21 16:17:14,187::task::595::TaskManager.Task::(_updateState) Task=`97b78287-45d2-4d5a-8336-460987df3840`::moving from state preparing -> state finished Thread-13::DEBUG::2014-06-21 16:17:14,187::resourceManager::940::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {} Thread-13::DEBUG::2014-06-21 16:17:14,187::resourceManager::977::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-13::DEBUG::2014-06-21 16:17:14,188::task::990::TaskManager.Task::(_decref) Task=`97b78287-45d2-4d5a-8336-460987df3840`::ref 0 aborting False Thread-13::DEBUG::2014-06-21 16:17:14,195::BindingXMLRPC::251::vds::(wrapper) client [192.168.10.119] flowID [6e44c0a3] Thread-13::DEBUG::2014-06-21 16:17:14,195::task::595::TaskManager.Task::(_updateState) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::moving from state init -> state preparing Thread-13::INFO::2014-06-21 16:17:14,196::logUtils::44::dispatcher::(wrapper) Run and protect: createStoragePool(poolType=None, spUUID='806d2356-12cf-437c-8917-dd13ee823e36', poolName='testing', masterDom='dc661957-c0c1-44ba-a5b9-e6558904207b', domList=['dc661957-c0c1-44ba-a5b9-e6558904207b'], masterVersion=2, lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, ioOpTimeoutSec=10, leaseRetries=3, options=None) Thread-13::DEBUG::2014-06-21 16:17:14,196::misc::756::SamplingMethod::(__call__) Trying to enter sampling method (storage.sdc.refreshStorage) Thread-13::DEBUG::2014-06-21 16:17:14,197::misc::758::SamplingMethod::(__call__) Got in to sampling method Thread-13::DEBUG::2014-06-21 16:17:14,197::misc::756::SamplingMethod::(__call__) Trying to enter sampling method (storage.iscsi.rescan) Thread-13::DEBUG::2014-06-21 16:17:14,198::misc::758::SamplingMethod::(__call__) Got in to sampling method Thread-13::DEBUG::2014-06-21 16:17:14,198::iscsi::407::Storage.ISCSI::(rescan) Performing SCSI scan, this will take up to 30 seconds Thread-13::DEBUG::2014-06-21 16:17:14,199::iscsiadm::92::Storage.Misc.excCmd::(_runCmd) '/usr/bin/sudo -n /sbin/iscsiadm -m session -R' (cwd None) Thread-13::DEBUG::2014-06-21 16:17:14,228::misc::766::SamplingMethod::(__call__) Returning last result Thread-13::DEBUG::2014-06-21 16:17:14,229::multipath::110::Storage.Misc.excCmd::(rescan) '/usr/bin/sudo -n /sbin/multipath -r' (cwd None) Thread-13::DEBUG::2014-06-21 16:17:14,294::multipath::110::Storage.Misc.excCmd::(rescan) SUCCESS: <err> = ''; <rc> = 0 Thread-13::DEBUG::2014-06-21 16:17:14,295::lvm::497::OperationMutex::(_invalidateAllPvs) Operation 'lvm invalidate operation' got the operation mutex Thread-13::DEBUG::2014-06-21 16:17:14,295::lvm::499::OperationMutex::(_invalidateAllPvs) Operation 'lvm invalidate operation' released the operation mutex Thread-13::DEBUG::2014-06-21 16:17:14,296::lvm::508::OperationMutex::(_invalidateAllVgs) Operation 'lvm invalidate operation' got the operation mutex Thread-13::DEBUG::2014-06-21 16:17:14,296::lvm::510::OperationMutex::(_invalidateAllVgs) Operation 'lvm invalidate operation' released the operation mutex Thread-13::DEBUG::2014-06-21 16:17:14,297::lvm::528::OperationMutex::(_invalidateAllLvs) Operation 'lvm invalidate operation' got the operation mutex Thread-13::DEBUG::2014-06-21 16:17:14,297::lvm::530::OperationMutex::(_invalidateAllLvs) Operation 'lvm invalidate operation' released the operation mutex Thread-13::DEBUG::2014-06-21 16:17:14,298::misc::766::SamplingMethod::(__call__) Returning last result Thread-13::DEBUG::2014-06-21 16:17:14,318::fileSD::150::Storage.StorageDomain::(__init__) Reading domain in path /rhev/data-center/mnt/glusterSD/192.168.10.120:_vmimage/dc661957-c0c1-44ba-a5b9-e6558904207b Thread-13::DEBUG::2014-06-21 16:17:14,322::persistentDict::192::Storage.PersistentDict::(__init__) Created a persistent dict with FileMetadataRW backend Thread-13::DEBUG::2014-06-21 16:17:14,328::persistentDict::234::Storage.PersistentDict::(refresh) read lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=vmimage', 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60', 'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', 'REMOTE_PATH=192.168.10.120:/vmimage', 'ROLE=Regular', 'SDUUID=dc661957-c0c1-44ba-a5b9-e6558904207b', 'TYPE=GLUSTERFS', 'VERSION=3', '_SHA_CKSUM=9fdc035c398d2cd8b5c31bf5eea2882c8782ed57'] Thread-13::DEBUG::2014-06-21 16:17:14,334::fileSD::609::Storage.StorageDomain::(imageGarbageCollector) Removing remnants of deleted images [] Thread-13::INFO::2014-06-21 16:17:14,335::sd::383::Storage.StorageDomain::(_registerResourceNamespaces) Resource namespace dc661957-c0c1-44ba-a5b9-e6558904207b_imageNS already registered Thread-13::INFO::2014-06-21 16:17:14,335::sd::391::Storage.StorageDomain::(_registerResourceNamespaces) Resource namespace dc661957-c0c1-44ba-a5b9-e6558904207b_volumeNS already registered Thread-13::INFO::2014-06-21 16:17:14,336::fileSD::350::Storage.StorageDomain::(validate) sdUUID=dc661957-c0c1-44ba-a5b9-e6558904207b Thread-13::DEBUG::2014-06-21 16:17:14,340::persistentDict::234::Storage.PersistentDict::(refresh) read lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=vmimage', 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60', 'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', 'REMOTE_PATH=192.168.10.120:/vmimage', 'ROLE=Regular', 'SDUUID=dc661957-c0c1-44ba-a5b9-e6558904207b', 'TYPE=GLUSTERFS', 'VERSION=3', '_SHA_CKSUM=9fdc035c398d2cd8b5c31bf5eea2882c8782ed57'] Thread-13::DEBUG::2014-06-21 16:17:14,341::resourceManager::198::ResourceManager.Request::(__init__) ResName=`Storage.806d2356-12cf-437c-8917-dd13ee823e36`ReqID=`de2ede47-22fa-43b8-9f3b-dc714a45b450`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '980' at 'createStoragePool' Thread-13::DEBUG::2014-06-21 16:17:14,342::resourceManager::542::ResourceManager::(registerResource) Trying to register resource 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' for lock type 'exclusive' Thread-13::DEBUG::2014-06-21 16:17:14,342::resourceManager::601::ResourceManager::(registerResource) Resource 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' is free. Now locking as 'exclusive' (1 active user) Thread-13::DEBUG::2014-06-21 16:17:14,343::resourceManager::238::ResourceManager.Request::(grant) ResName=`Storage.806d2356-12cf-437c-8917-dd13ee823e36`ReqID=`de2ede47-22fa-43b8-9f3b-dc714a45b450`::Granted request Thread-13::DEBUG::2014-06-21 16:17:14,343::task::827::TaskManager.Task::(resourceAcquired) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::_resourcesAcquired: Storage.806d2356-12cf-437c-8917-dd13ee823e36 (exclusive) Thread-13::DEBUG::2014-06-21 16:17:14,344::task::990::TaskManager.Task::(_decref) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::ref 1 aborting False Thread-13::DEBUG::2014-06-21 16:17:14,345::resourceManager::198::ResourceManager.Request::(__init__) ResName=`Storage.dc661957-c0c1-44ba-a5b9-e6558904207b`ReqID=`71bf6917-b501-4016-ad8e-8b84849da8cb`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '982' at 'createStoragePool' Thread-13::DEBUG::2014-06-21 16:17:14,345::resourceManager::542::ResourceManager::(registerResource) Trying to register resource 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' for lock type 'exclusive' Thread-13::DEBUG::2014-06-21 16:17:14,346::resourceManager::601::ResourceManager::(registerResource) Resource 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' is free. Now locking as 'exclusive' (1 active user) Thread-13::DEBUG::2014-06-21 16:17:14,346::resourceManager::238::ResourceManager.Request::(grant) ResName=`Storage.dc661957-c0c1-44ba-a5b9-e6558904207b`ReqID=`71bf6917-b501-4016-ad8e-8b84849da8cb`::Granted request Thread-13::DEBUG::2014-06-21 16:17:14,347::task::827::TaskManager.Task::(resourceAcquired) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::_resourcesAcquired: Storage.dc661957-c0c1-44ba-a5b9-e6558904207b (exclusive) Thread-13::DEBUG::2014-06-21 16:17:14,347::task::990::TaskManager.Task::(_decref) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::ref 1 aborting False Thread-13::INFO::2014-06-21 16:17:14,347::sp::133::Storage.StoragePool::(setBackend) updating pool 806d2356-12cf-437c-8917-dd13ee823e36 backend from type NoneType instance 0x39e278bf00 to type StoragePoolDiskBackend instance 0x7f764c093cb0 Thread-13::INFO::2014-06-21 16:17:14,348::sp::548::Storage.StoragePool::(create) spUUID=806d2356-12cf-437c-8917-dd13ee823e36 poolName=testing master_sd=dc661957-c0c1-44ba-a5b9-e6558904207b domList=['dc661957-c0c1-44ba-a5b9-e6558904207b'] masterVersion=2 {'LEASETIMESEC': 60, 'IOOPTIMEOUTSEC': 10, 'LEASERETRIES': 3, 'LOCKRENEWALINTERVALSEC': 5} Thread-13::INFO::2014-06-21 16:17:14,348::fileSD::350::Storage.StorageDomain::(validate) sdUUID=dc661957-c0c1-44ba-a5b9-e6558904207b Thread-13::DEBUG::2014-06-21 16:17:14,352::persistentDict::234::Storage.PersistentDict::(refresh) read lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=vmimage', 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60', 'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', 'REMOTE_PATH=192.168.10.120:/vmimage', 'ROLE=Regular', 'SDUUID=dc661957-c0c1-44ba-a5b9-e6558904207b', 'TYPE=GLUSTERFS', 'VERSION=3', '_SHA_CKSUM=9fdc035c398d2cd8b5c31bf5eea2882c8782ed57'] Thread-13::DEBUG::2014-06-21 16:17:14,357::persistentDict::234::Storage.PersistentDict::(refresh) read lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=vmimage', 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60', 'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', 'REMOTE_PATH=192.168.10.120:/vmimage', 'ROLE=Regular', 'SDUUID=dc661957-c0c1-44ba-a5b9-e6558904207b', 'TYPE=GLUSTERFS', 'VERSION=3', '_SHA_CKSUM=9fdc035c398d2cd8b5c31bf5eea2882c8782ed57'] Thread-13::WARNING::2014-06-21 16:17:14,358::fileUtils::167::Storage.fileUtils::(createdir) Dir /rhev/data-center/806d2356-12cf-437c-8917-dd13ee823e36 already exists Thread-13::DEBUG::2014-06-21 16:17:14,358::persistentDict::167::Storage.PersistentDict::(transaction) Starting transaction Thread-13::DEBUG::2014-06-21 16:17:14,359::persistentDict::175::Storage.PersistentDict::(transaction) Finished transaction Thread-13::INFO::2014-06-21 16:17:14,359::clusterlock::184::SANLock::(acquireHostId) Acquiring host id for domain dc661957-c0c1-44ba-a5b9-e6558904207b (id: 250) Thread-24::DEBUG::2014-06-21 16:17:14,394::task::595::TaskManager.Task::(_updateState) Task=`c4430b80-31d9-4a1d-bee8-fae01a438da6`::moving from state init -> state preparing Thread-24::INFO::2014-06-21 16:17:14,395::logUtils::44::dispatcher::(wrapper) Run and protect: repoStats(options=None) Thread-24::INFO::2014-06-21 16:17:14,395::logUtils::47::dispatcher::(wrapper) Run and protect: repoStats, Return response: {} Thread-24::DEBUG::2014-06-21 16:17:14,396::task::1185::TaskManager.Task::(prepare) Task=`c4430b80-31d9-4a1d-bee8-fae01a438da6`::finished: {} Thread-24::DEBUG::2014-06-21 16:17:14,396::task::595::TaskManager.Task::(_updateState) Task=`c4430b80-31d9-4a1d-bee8-fae01a438da6`::moving from state preparing -> state finished Thread-24::DEBUG::2014-06-21 16:17:14,396::resourceManager::940::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {} Thread-24::DEBUG::2014-06-21 16:17:14,396::resourceManager::977::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-24::DEBUG::2014-06-21 16:17:14,397::task::990::TaskManager.Task::(_decref) Task=`c4430b80-31d9-4a1d-bee8-fae01a438da6`::ref 0 aborting False Thread-13::ERROR::2014-06-21 16:17:15,361::task::866::TaskManager.Task::(_setError) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::Unexpected error Traceback (most recent call last): File "/usr/share/vdsm/storage/task.py", line 873, in _run return fn(*args, **kargs) File "/usr/share/vdsm/logUtils.py", line 45, in wrapper res = f(*args, **kwargs) File "/usr/share/vdsm/storage/hsm.py", line 988, in createStoragePool leaseParams) File "/usr/share/vdsm/storage/sp.py", line 573, in create self._acquireTemporaryClusterLock(msdUUID, leaseParams) File "/usr/share/vdsm/storage/sp.py", line 515, in _acquireTemporaryClusterLock msd.acquireHostId(self.id) File "/usr/share/vdsm/storage/sd.py", line 467, in acquireHostId self._clusterLock.acquireHostId(hostId, async) File "/usr/share/vdsm/storage/clusterlock.py", line 199, in acquireHostId raise se.AcquireHostIdFailure(self._sdUUID, e) AcquireHostIdFailure: Cannot acquire host id: ('dc661957-c0c1-44ba-a5b9-e6558904207b', SanlockException(90, 'Sanlock lockspace add failure', 'Message too long')) Thread-13::DEBUG::2014-06-21 16:17:15,363::task::885::TaskManager.Task::(_run) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::Task._run: d815e5e5-0202-4137-94be-21dc5e2b61c9 (None, '806d2356-12cf-437c-8917-dd13ee823e36', 'testing', 'dc661957-c0c1-44ba-a5b9-e6558904207b', ['dc661957-c0c1-44ba-a5b9-e6558904207b'], 2, None, 5, 60, 10, 3) {} failed - stopping task Thread-13::DEBUG::2014-06-21 16:17:15,364::task::1211::TaskManager.Task::(stop) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::stopping in state preparing (force False) Thread-13::DEBUG::2014-06-21 16:17:15,364::task::990::TaskManager.Task::(_decref) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::ref 1 aborting True Thread-13::INFO::2014-06-21 16:17:15,365::task::1168::TaskManager.Task::(prepare) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::aborting: Task is aborted: 'Cannot acquire host id' - code 661 Thread-13::DEBUG::2014-06-21 16:17:15,365::task::1173::TaskManager.Task::(prepare) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::Prepare: aborted: Cannot acquire host id Thread-13::DEBUG::2014-06-21 16:17:15,365::task::990::TaskManager.Task::(_decref) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::ref 0 aborting True Thread-13::DEBUG::2014-06-21 16:17:15,366::task::925::TaskManager.Task::(_doAbort) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::Task._doAbort: force False Thread-13::DEBUG::2014-06-21 16:17:15,366::resourceManager::977::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-13::DEBUG::2014-06-21 16:17:15,366::task::595::TaskManager.Task::(_updateState) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::moving from state preparing -> state aborting Thread-13::DEBUG::2014-06-21 16:17:15,366::task::550::TaskManager.Task::(__state_aborting) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::_aborting: recover policy none Thread-13::DEBUG::2014-06-21 16:17:15,367::task::595::TaskManager.Task::(_updateState) Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::moving from state aborting -> state failed Thread-13::DEBUG::2014-06-21 16:17:15,367::resourceManager::940::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b': < ResourceRef 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b', isValid: 'True' obj: 'None'>, 'Storage.806d2356-12cf-437c-8917-dd13ee823e36': < ResourceRef 'Storage.806d2356-12cf-437c-8917-dd13ee823e36', isValid: 'True' obj: 'None'>} Thread-13::DEBUG::2014-06-21 16:17:15,367::resourceManager::977::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-13::DEBUG::2014-06-21 16:17:15,368::resourceManager::616::ResourceManager::(releaseResource) Trying to release resource 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' Thread-13::DEBUG::2014-06-21 16:17:15,369::resourceManager::635::ResourceManager::(releaseResource) Released resource 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' (0 active users) Thread-13::DEBUG::2014-06-21 16:17:15,369::resourceManager::641::ResourceManager::(releaseResource) Resource 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' is free, finding out if anyone is waiting for it. Thread-13::DEBUG::2014-06-21 16:17:15,369::resourceManager::649::ResourceManager::(releaseResource) No one is waiting for resource 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b', Clearing records. Thread-13::DEBUG::2014-06-21 16:17:15,370::resourceManager::616::ResourceManager::(releaseResource) Trying to release resource 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' Thread-13::DEBUG::2014-06-21 16:17:15,370::resourceManager::635::ResourceManager::(releaseResource) Released resource 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' (0 active users) Thread-13::DEBUG::2014-06-21 16:17:15,370::resourceManager::641::ResourceManager::(releaseResource) Resource 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' is free, finding out if anyone is waiting for it. Thread-13::DEBUG::2014-06-21 16:17:15,371::resourceManager::649::ResourceManager::(releaseResource) No one is waiting for resource 'Storage.806d2356-12cf-437c-8917-dd13ee823e36', Clearing records. Thread-13::ERROR::2014-06-21 16:17:15,371::dispatcher::65::Storage.Dispatcher.Protect::(run) {'status': {'message': "Cannot acquire host id: ('dc661957-c0c1-44ba-a5b9-e6558904207b', SanlockException(90, 'Sanlock lockspace add failure', 'Message too long'))", 'code': 661}}
My oVirt version: 3.4.2-1.el6 (CentOS 6.5) The hypervisor hosts run GlusterFS 3.5.0-3.fc19.(Fedora 19) The two storage servers run GlusterFS 3.5.0-2.el6 (Centos 6.5)
So I am NOT using local storage of the hypervisor hosts for the GlusterFS bricks.
What can I do to solve this error?
By the way, the options on the GlusterFS volume are as follows:
Volume Name: vmimage Type: Replicate Volume ID: 348e1d45-1b80-420b-91c2-93f0d764f227 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: 192.168.10.120:/export/gluster01/brick Brick2: 192.168.10.149:/export/gluster01/brick Options Reconfigured: network.ping-timeout: 10 cluster.quorum-count: 1 cluster.quorum-type: auto server.allow-insecure: on storage.owner-uid: 36 storage.owner-gid: 36
OK, fixed it. For someone else's reference, I had to set the following options on the gluster volume: network.remote-dio: on performance.io-cache: off performance.read-ahead: off performance.quick-read: off cluster.eager-lock: enable Apparently that's done by the 'optimize for virt store' checkbox, but obviously not when the volume is created manually. Having this in the documentation on ovirt.org would have saved me a lot of time and frustration. -- Tiemen Ruiten Systems Engineer R&D Media