[ovirt-users] can't attach storage domain to data center
Tiemen Ruiten
t.ruiten at rdmedia.com
Tue Jun 24 08:53:03 UTC 2014
I found the relevant Redhat documentation:
https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.0/html/Quick_Start_Guide/chap-Quick_Start_Guide-Virtual_Preparation.html
So for Punit's suggestion to work, you should also execute this command:|
gluster volume set /|VOLNAME|/ group virt|
Hope this helps someone!
Tiemen
On 06/24/14 04:11, Punit Dambiwal wrote:
> Hi,
>
> You can also use the below method on every gluster node :-
>
> For Group-virt (optimize for virt store)
>
> 1. Create the file name /var/lib/glusterd/groups/virt
> 2. And paste all the contents from this location to this file :-
> https://raw.githubusercontent.com/gluster/glusterfs/master/extras/group-virt.example
> 3. service glusterd restart
> 4. service vdsmd restart
>
> --------------
> quick-read=off
> read-ahead=off
> io-cache=off
> stat-prefetch=off
> eager-lock=enable
> remote-dio=enable
> quorum-type=auto
> server-quorum-type=server
> --------------
>
> Thanks,
> Punit Dambiwal
>
>
>
> On Mon, Jun 23, 2014 at 4:35 PM, Itamar Heim <iheim at redhat.com
> <mailto:iheim at redhat.com>> wrote:
>
> On 06/22/2014 06:38 PM, Tiemen Ruiten wrote:
>
> On 06/21/14 16:57, Tiemen Ruiten wrote:
>
> On 06/21/14 16:37, Tiemen Ruiten wrote:
>
> Hello,
>
> I've been struggling to set up an Ovirt cluster and am
> now bumping into
> this problem:
>
> When I try to create a new (Gluster) storage domain,
> it fails to attach
> to the data center. The error on the node from vdsm.log:
>
> Thread-13::DEBUG::2014-06-21
> 16:17:14,157::BindingXMLRPC::251::vds::(wrapper)
> client [192.168.10.119]
> flowID [6e44c0a3]
> Thread-13::DEBUG::2014-06-21
> 16:17:14,159::task::595::TaskManager.Task::(_updateState)
> Task=`97b78287-45d2-4d5a-8336-460987df3840`::moving
> from state init ->
> state preparing
> Thread-13::INFO::2014-06-21
> 16:17:14,160::logUtils::44::dispatcher::(wrapper) Run
> and protect:
> connectStorageServer(domType=7,
> spUUID='00000000-0000-0000-0000-000000000000',
> conList=[{'port': '',
> 'connection': '192.168.10.120:/vmimage', 'iqn': '',
> 'user': '', 'tpgt':
> '1', 'vfs_type': 'glusterfs', 'password': '******', 'id':
> '901b15ec-6b05-43c1-8a50-06b34c8ffdbd'}], options=None)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,172::hsm::2340::Storage.HSM::(__prefetchDomains)
> glusterDomPath: glusterSD/*
> Thread-13::DEBUG::2014-06-21
> 16:17:14,185::hsm::2352::Storage.HSM::(__prefetchDomains)
> Found SD
> uuids: ('dc661957-c0c1-44ba-a5b9-e6558904207b',)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,185::hsm::2408::Storage.HSM::(connectStorageServer)
> knownSDs:
> {dc661957-c0c1-44ba-a5b9-e6558904207b:
> storage.glusterSD.findDomain}
> Thread-13::INFO::2014-06-21
> 16:17:14,186::logUtils::47::dispatcher::(wrapper) Run
> and protect:
> connectStorageServer, Return response: {'statuslist':
> [{'status': 0,
> 'id': '901b15ec-6b05-43c1-8a50-06b34c8ffdbd'}]}
> Thread-13::DEBUG::2014-06-21
> 16:17:14,186::task::1185::TaskManager.Task::(prepare)
> Task=`97b78287-45d2-4d5a-8336-460987df3840`::finished:
> {'statuslist':
> [{'status': 0, 'id':
> '901b15ec-6b05-43c1-8a50-06b34c8ffdbd'}]}
> Thread-13::DEBUG::2014-06-21
> 16:17:14,187::task::595::TaskManager.Task::(_updateState)
> Task=`97b78287-45d2-4d5a-8336-460987df3840`::moving
> from state preparing
> -> state finished
> Thread-13::DEBUG::2014-06-21
> 16:17:14,187::resourceManager::940::ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources {}
> Thread-13::DEBUG::2014-06-21
> 16:17:14,187::resourceManager::977::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-13::DEBUG::2014-06-21
> 16:17:14,188::task::990::TaskManager.Task::(_decref)
> Task=`97b78287-45d2-4d5a-8336-460987df3840`::ref 0
> aborting False
> Thread-13::DEBUG::2014-06-21
> 16:17:14,195::BindingXMLRPC::251::vds::(wrapper)
> client [192.168.10.119]
> flowID [6e44c0a3]
> Thread-13::DEBUG::2014-06-21
> 16:17:14,195::task::595::TaskManager.Task::(_updateState)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::moving
> from state init ->
> state preparing
> Thread-13::INFO::2014-06-21
> 16:17:14,196::logUtils::44::dispatcher::(wrapper) Run
> and protect:
> createStoragePool(poolType=None,
> spUUID='806d2356-12cf-437c-8917-dd13ee823e36',
> poolName='testing',
> masterDom='dc661957-c0c1-44ba-a5b9-e6558904207b',
> domList=['dc661957-c0c1-44ba-a5b9-e6558904207b'],
> masterVersion=2,
> lockPolicy=None, lockRenewalIntervalSec=5,
> leaseTimeSec=60,
> ioOpTimeoutSec=10, leaseRetries=3, options=None)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,196::misc::756::SamplingMethod::(__call__)
> Trying to enter
> sampling method (storage.sdc.refreshStorage)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,197::misc::758::SamplingMethod::(__call__)
> Got in to sampling
> method
> Thread-13::DEBUG::2014-06-21
> 16:17:14,197::misc::756::SamplingMethod::(__call__)
> Trying to enter
> sampling method (storage.iscsi.rescan)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,198::misc::758::SamplingMethod::(__call__)
> Got in to sampling
> method
> Thread-13::DEBUG::2014-06-21
> 16:17:14,198::iscsi::407::Storage.ISCSI::(rescan)
> Performing SCSI scan,
> this will take up to 30 seconds
> Thread-13::DEBUG::2014-06-21
> 16:17:14,199::iscsiadm::92::Storage.Misc.excCmd::(_runCmd)
> '/usr/bin/sudo -n /sbin/iscsiadm -m session -R' (cwd None)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,228::misc::766::SamplingMethod::(__call__)
> Returning last result
> Thread-13::DEBUG::2014-06-21
> 16:17:14,229::multipath::110::Storage.Misc.excCmd::(rescan)
> '/usr/bin/sudo -n /sbin/multipath -r' (cwd None)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,294::multipath::110::Storage.Misc.excCmd::(rescan)
> SUCCESS:
> <err> = ''; <rc> = 0
> Thread-13::DEBUG::2014-06-21
> 16:17:14,295::lvm::497::OperationMutex::(_invalidateAllPvs)
> Operation
> 'lvm invalidate operation' got the operation mutex
> Thread-13::DEBUG::2014-06-21
> 16:17:14,295::lvm::499::OperationMutex::(_invalidateAllPvs)
> Operation
> 'lvm invalidate operation' released the operation mutex
> Thread-13::DEBUG::2014-06-21
> 16:17:14,296::lvm::508::OperationMutex::(_invalidateAllVgs)
> Operation
> 'lvm invalidate operation' got the operation mutex
> Thread-13::DEBUG::2014-06-21
> 16:17:14,296::lvm::510::OperationMutex::(_invalidateAllVgs)
> Operation
> 'lvm invalidate operation' released the operation mutex
> Thread-13::DEBUG::2014-06-21
> 16:17:14,297::lvm::528::OperationMutex::(_invalidateAllLvs)
> Operation
> 'lvm invalidate operation' got the operation mutex
> Thread-13::DEBUG::2014-06-21
> 16:17:14,297::lvm::530::OperationMutex::(_invalidateAllLvs)
> Operation
> 'lvm invalidate operation' released the operation mutex
> Thread-13::DEBUG::2014-06-21
> 16:17:14,298::misc::766::SamplingMethod::(__call__)
> Returning last result
> Thread-13::DEBUG::2014-06-21
> 16:17:14,318::fileSD::150::Storage.StorageDomain::(__init__)
> Reading
> domain in path
> /rhev/data-center/mnt/glusterSD/192.168.10.120:_vmimage/dc661957-c0c1-44ba-a5b9-e6558904207b
> Thread-13::DEBUG::2014-06-21
> 16:17:14,322::persistentDict::192::Storage.PersistentDict::(__init__)
> Created a persistent dict with FileMetadataRW backend
> Thread-13::DEBUG::2014-06-21
> 16:17:14,328::persistentDict::234::Storage.PersistentDict::(refresh)
> read lines (FileMetadataRW)=['CLASS=Data',
> 'DESCRIPTION=vmimage',
> 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3',
> 'LEASETIMESEC=60', 'LOCKPOLICY=',
> 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=',
> 'REMOTE_PATH=192.168.10.120:/vmimage', 'ROLE=Regular',
> 'SDUUID=dc661957-c0c1-44ba-a5b9-e6558904207b',
> 'TYPE=GLUSTERFS',
> 'VERSION=3',
> '_SHA_CKSUM=9fdc035c398d2cd8b5c31bf5eea2882c8782ed57']
> Thread-13::DEBUG::2014-06-21
> 16:17:14,334::fileSD::609::Storage.StorageDomain::(imageGarbageCollector)
> Removing
> remnants of deleted images []
> Thread-13::INFO::2014-06-21
> 16:17:14,335::sd::383::Storage.StorageDomain::(_registerResourceNamespaces)
> Resource namespace
> dc661957-c0c1-44ba-a5b9-e6558904207b_imageNS already
> registered
> Thread-13::INFO::2014-06-21
> 16:17:14,335::sd::391::Storage.StorageDomain::(_registerResourceNamespaces)
> Resource namespace
> dc661957-c0c1-44ba-a5b9-e6558904207b_volumeNS already
> registered
> Thread-13::INFO::2014-06-21
> 16:17:14,336::fileSD::350::Storage.StorageDomain::(validate)
> sdUUID=dc661957-c0c1-44ba-a5b9-e6558904207b
> Thread-13::DEBUG::2014-06-21
> 16:17:14,340::persistentDict::234::Storage.PersistentDict::(refresh)
> read lines (FileMetadataRW)=['CLASS=Data',
> 'DESCRIPTION=vmimage',
> 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3',
> 'LEASETIMESEC=60', 'LOCKPOLICY=',
> 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=',
> 'REMOTE_PATH=192.168.10.120:/vmimage', 'ROLE=Regular',
> 'SDUUID=dc661957-c0c1-44ba-a5b9-e6558904207b',
> 'TYPE=GLUSTERFS',
> 'VERSION=3',
> '_SHA_CKSUM=9fdc035c398d2cd8b5c31bf5eea2882c8782ed57']
> Thread-13::DEBUG::2014-06-21
> 16:17:14,341::resourceManager::198::ResourceManager.Request::(__init__)
> ResName=`Storage.806d2356-12cf-437c-8917-dd13ee823e36`ReqID=`de2ede47-22fa-43b8-9f3b-dc714a45b450`::Request
> was made in '/usr/share/vdsm/storage/hsm.py' line '980' at
> 'createStoragePool'
> Thread-13::DEBUG::2014-06-21
> 16:17:14,342::resourceManager::542::ResourceManager::(registerResource)
> Trying to register resource
> 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' for
> lock type 'exclusive'
> Thread-13::DEBUG::2014-06-21
> 16:17:14,342::resourceManager::601::ResourceManager::(registerResource)
> Resource
> 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' is
> free. Now
> locking as 'exclusive' (1 active user)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,343::resourceManager::238::ResourceManager.Request::(grant)
> ResName=`Storage.806d2356-12cf-437c-8917-dd13ee823e36`ReqID=`de2ede47-22fa-43b8-9f3b-dc714a45b450`::Granted
> request
> Thread-13::DEBUG::2014-06-21
> 16:17:14,343::task::827::TaskManager.Task::(resourceAcquired)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::_resourcesAcquired:
> Storage.806d2356-12cf-437c-8917-dd13ee823e36 (exclusive)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,344::task::990::TaskManager.Task::(_decref)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::ref 1
> aborting False
> Thread-13::DEBUG::2014-06-21
> 16:17:14,345::resourceManager::198::ResourceManager.Request::(__init__)
> ResName=`Storage.dc661957-c0c1-44ba-a5b9-e6558904207b`ReqID=`71bf6917-b501-4016-ad8e-8b84849da8cb`::Request
> was made in '/usr/share/vdsm/storage/hsm.py' line '982' at
> 'createStoragePool'
> Thread-13::DEBUG::2014-06-21
> 16:17:14,345::resourceManager::542::ResourceManager::(registerResource)
> Trying to register resource
> 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' for
> lock type 'exclusive'
> Thread-13::DEBUG::2014-06-21
> 16:17:14,346::resourceManager::601::ResourceManager::(registerResource)
> Resource
> 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' is
> free. Now
> locking as 'exclusive' (1 active user)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,346::resourceManager::238::ResourceManager.Request::(grant)
> ResName=`Storage.dc661957-c0c1-44ba-a5b9-e6558904207b`ReqID=`71bf6917-b501-4016-ad8e-8b84849da8cb`::Granted
> request
> Thread-13::DEBUG::2014-06-21
> 16:17:14,347::task::827::TaskManager.Task::(resourceAcquired)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::_resourcesAcquired:
> Storage.dc661957-c0c1-44ba-a5b9-e6558904207b (exclusive)
> Thread-13::DEBUG::2014-06-21
> 16:17:14,347::task::990::TaskManager.Task::(_decref)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::ref 1
> aborting False
> Thread-13::INFO::2014-06-21
> 16:17:14,347::sp::133::Storage.StoragePool::(setBackend)
> updating pool
> 806d2356-12cf-437c-8917-dd13ee823e36 backend from type
> NoneType instance
> 0x39e278bf00 to type StoragePoolDiskBackend instance
> 0x7f764c093cb0
> Thread-13::INFO::2014-06-21
> 16:17:14,348::sp::548::Storage.StoragePool::(create)
> spUUID=806d2356-12cf-437c-8917-dd13ee823e36
> poolName=testing
> master_sd=dc661957-c0c1-44ba-a5b9-e6558904207b
> domList=['dc661957-c0c1-44ba-a5b9-e6558904207b']
> masterVersion=2
> {'LEASETIMESEC': 60, 'IOOPTIMEOUTSEC': 10,
> 'LEASERETRIES': 3,
> 'LOCKRENEWALINTERVALSEC': 5}
> Thread-13::INFO::2014-06-21
> 16:17:14,348::fileSD::350::Storage.StorageDomain::(validate)
> sdUUID=dc661957-c0c1-44ba-a5b9-e6558904207b
> Thread-13::DEBUG::2014-06-21
> 16:17:14,352::persistentDict::234::Storage.PersistentDict::(refresh)
> read lines (FileMetadataRW)=['CLASS=Data',
> 'DESCRIPTION=vmimage',
> 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3',
> 'LEASETIMESEC=60', 'LOCKPOLICY=',
> 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=',
> 'REMOTE_PATH=192.168.10.120:/vmimage', 'ROLE=Regular',
> 'SDUUID=dc661957-c0c1-44ba-a5b9-e6558904207b',
> 'TYPE=GLUSTERFS',
> 'VERSION=3',
> '_SHA_CKSUM=9fdc035c398d2cd8b5c31bf5eea2882c8782ed57']
> Thread-13::DEBUG::2014-06-21
> 16:17:14,357::persistentDict::234::Storage.PersistentDict::(refresh)
> read lines (FileMetadataRW)=['CLASS=Data',
> 'DESCRIPTION=vmimage',
> 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3',
> 'LEASETIMESEC=60', 'LOCKPOLICY=',
> 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=',
> 'REMOTE_PATH=192.168.10.120:/vmimage', 'ROLE=Regular',
> 'SDUUID=dc661957-c0c1-44ba-a5b9-e6558904207b',
> 'TYPE=GLUSTERFS',
> 'VERSION=3',
> '_SHA_CKSUM=9fdc035c398d2cd8b5c31bf5eea2882c8782ed57']
> Thread-13::WARNING::2014-06-21
> 16:17:14,358::fileUtils::167::Storage.fileUtils::(createdir)
> Dir
> /rhev/data-center/806d2356-12cf-437c-8917-dd13ee823e36
> already exists
> Thread-13::DEBUG::2014-06-21
> 16:17:14,358::persistentDict::167::Storage.PersistentDict::(transaction)
> Starting transaction
> Thread-13::DEBUG::2014-06-21
> 16:17:14,359::persistentDict::175::Storage.PersistentDict::(transaction)
> Finished transaction
> Thread-13::INFO::2014-06-21
> 16:17:14,359::clusterlock::184::SANLock::(acquireHostId)
> Acquiring host
> id for domain dc661957-c0c1-44ba-a5b9-e6558904207b
> (id: 250)
> Thread-24::DEBUG::2014-06-21
> 16:17:14,394::task::595::TaskManager.Task::(_updateState)
> Task=`c4430b80-31d9-4a1d-bee8-fae01a438da6`::moving
> from state init ->
> state preparing
> Thread-24::INFO::2014-06-21
> 16:17:14,395::logUtils::44::dispatcher::(wrapper) Run
> and protect:
> repoStats(options=None)
> Thread-24::INFO::2014-06-21
> 16:17:14,395::logUtils::47::dispatcher::(wrapper) Run
> and protect:
> repoStats, Return response: {}
> Thread-24::DEBUG::2014-06-21
> 16:17:14,396::task::1185::TaskManager.Task::(prepare)
> Task=`c4430b80-31d9-4a1d-bee8-fae01a438da6`::finished: {}
> Thread-24::DEBUG::2014-06-21
> 16:17:14,396::task::595::TaskManager.Task::(_updateState)
> Task=`c4430b80-31d9-4a1d-bee8-fae01a438da6`::moving
> from state preparing
> -> state finished
> Thread-24::DEBUG::2014-06-21
> 16:17:14,396::resourceManager::940::ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources {}
> Thread-24::DEBUG::2014-06-21
> 16:17:14,396::resourceManager::977::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-24::DEBUG::2014-06-21
> 16:17:14,397::task::990::TaskManager.Task::(_decref)
> Task=`c4430b80-31d9-4a1d-bee8-fae01a438da6`::ref 0
> aborting False
> Thread-13::ERROR::2014-06-21
> 16:17:15,361::task::866::TaskManager.Task::(_setError)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::Unexpected error
> Traceback (most recent call last):
> File "/usr/share/vdsm/storage/task.py", line 873,
> in _run
> return fn(*args, **kargs)
> File "/usr/share/vdsm/logUtils.py", line 45, in wrapper
> res = f(*args, **kwargs)
> File "/usr/share/vdsm/storage/hsm.py", line 988, in
> createStoragePool
> leaseParams)
> File "/usr/share/vdsm/storage/sp.py", line 573, in
> create
> self._acquireTemporaryClusterLock(msdUUID,
> leaseParams)
> File "/usr/share/vdsm/storage/sp.py", line 515, in
> _acquireTemporaryClusterLock
> msd.acquireHostId(self.id <http://self.id>)
> File "/usr/share/vdsm/storage/sd.py", line 467, in
> acquireHostId
> self._clusterLock.acquireHostId(hostId, async)
> File "/usr/share/vdsm/storage/clusterlock.py", line
> 199, in acquireHostId
> raise se.AcquireHostIdFailure(self._sdUUID, e)
> AcquireHostIdFailure: Cannot acquire host id:
> ('dc661957-c0c1-44ba-a5b9-e6558904207b',
> SanlockException(90, 'Sanlock
> lockspace add failure', 'Message too long'))
> Thread-13::DEBUG::2014-06-21
> 16:17:15,363::task::885::TaskManager.Task::(_run)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::Task._run:
> d815e5e5-0202-4137-94be-21dc5e2b61c9 (None,
> '806d2356-12cf-437c-8917-dd13ee823e36', 'testing',
> 'dc661957-c0c1-44ba-a5b9-e6558904207b',
> ['dc661957-c0c1-44ba-a5b9-e6558904207b'], 2, None, 5,
> 60, 10, 3) {}
> failed - stopping task
> Thread-13::DEBUG::2014-06-21
> 16:17:15,364::task::1211::TaskManager.Task::(stop)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::stopping
> in state preparing
> (force False)
> Thread-13::DEBUG::2014-06-21
> 16:17:15,364::task::990::TaskManager.Task::(_decref)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::ref 1
> aborting True
> Thread-13::INFO::2014-06-21
> 16:17:15,365::task::1168::TaskManager.Task::(prepare)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::aborting:
> Task is aborted:
> 'Cannot acquire host id' - code 661
> Thread-13::DEBUG::2014-06-21
> 16:17:15,365::task::1173::TaskManager.Task::(prepare)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::Prepare:
> aborted: Cannot
> acquire host id
> Thread-13::DEBUG::2014-06-21
> 16:17:15,365::task::990::TaskManager.Task::(_decref)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::ref 0
> aborting True
> Thread-13::DEBUG::2014-06-21
> 16:17:15,366::task::925::TaskManager.Task::(_doAbort)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::Task._doAbort:
> force False
> Thread-13::DEBUG::2014-06-21
> 16:17:15,366::resourceManager::977::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-13::DEBUG::2014-06-21
> 16:17:15,366::task::595::TaskManager.Task::(_updateState)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::moving
> from state preparing
> -> state aborting
> Thread-13::DEBUG::2014-06-21
> 16:17:15,366::task::550::TaskManager.Task::(__state_aborting)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::_aborting: recover
> policy none
> Thread-13::DEBUG::2014-06-21
> 16:17:15,367::task::595::TaskManager.Task::(_updateState)
> Task=`d815e5e5-0202-4137-94be-21dc5e2b61c9`::moving
> from state aborting
> -> state failed
> Thread-13::DEBUG::2014-06-21
> 16:17:15,367::resourceManager::940::ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources
> {'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b': <
> ResourceRef
> 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b',
> isValid: 'True' obj:
> 'None'>,
> 'Storage.806d2356-12cf-437c-8917-dd13ee823e36': <
> ResourceRef
> 'Storage.806d2356-12cf-437c-8917-dd13ee823e36',
> isValid: 'True' obj:
> 'None'>}
> Thread-13::DEBUG::2014-06-21
> 16:17:15,367::resourceManager::977::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-13::DEBUG::2014-06-21
> 16:17:15,368::resourceManager::616::ResourceManager::(releaseResource)
> Trying to release resource
> 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b'
> Thread-13::DEBUG::2014-06-21
> 16:17:15,369::resourceManager::635::ResourceManager::(releaseResource)
> Released resource
> 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' (0
> active users)
> Thread-13::DEBUG::2014-06-21
> 16:17:15,369::resourceManager::641::ResourceManager::(releaseResource)
> Resource
> 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b' is
> free, finding
> out if anyone is waiting for it.
> Thread-13::DEBUG::2014-06-21
> 16:17:15,369::resourceManager::649::ResourceManager::(releaseResource)
> No one is waiting for resource
> 'Storage.dc661957-c0c1-44ba-a5b9-e6558904207b',
> Clearing records.
> Thread-13::DEBUG::2014-06-21
> 16:17:15,370::resourceManager::616::ResourceManager::(releaseResource)
> Trying to release resource
> 'Storage.806d2356-12cf-437c-8917-dd13ee823e36'
> Thread-13::DEBUG::2014-06-21
> 16:17:15,370::resourceManager::635::ResourceManager::(releaseResource)
> Released resource
> 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' (0
> active users)
> Thread-13::DEBUG::2014-06-21
> 16:17:15,370::resourceManager::641::ResourceManager::(releaseResource)
> Resource
> 'Storage.806d2356-12cf-437c-8917-dd13ee823e36' is
> free, finding
> out if anyone is waiting for it.
> Thread-13::DEBUG::2014-06-21
> 16:17:15,371::resourceManager::649::ResourceManager::(releaseResource)
> No one is waiting for resource
> 'Storage.806d2356-12cf-437c-8917-dd13ee823e36',
> Clearing records.
> Thread-13::ERROR::2014-06-21
> 16:17:15,371::dispatcher::65::Storage.Dispatcher.Protect::(run)
> {'status': {'message': "Cannot acquire host id:
> ('dc661957-c0c1-44ba-a5b9-e6558904207b',
> SanlockException(90, 'Sanlock
> lockspace add failure', 'Message too long'))", 'code':
> 661}}
>
>
> My oVirt version: 3.4.2-1.el6 (CentOS 6.5)
> The hypervisor hosts run GlusterFS
> 3.5.0-3.fc19.(Fedora 19)
> The two storage servers run GlusterFS 3.5.0-2.el6
> (Centos 6.5)
>
> So I am NOT using local storage of the hypervisor
> hosts for the
> GlusterFS bricks.
>
> What can I do to solve this error?
>
> By the way, the options on the GlusterFS volume are as
> follows:
>
> Volume Name: vmimage
> Type: Replicate
> Volume ID: 348e1d45-1b80-420b-91c2-93f0d764f227
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.10.120:/export/gluster01/brick
> Brick2: 192.168.10.149:/export/gluster01/brick
> Options Reconfigured:
> network.ping-timeout: 10
> cluster.quorum-count: 1
> cluster.quorum-type: auto
> server.allow-insecure: on
> storage.owner-uid: 36
> storage.owner-gid: 36
>
> OK, fixed it. For someone else's reference, I had to set the
> following
> options on the gluster volume:
>
> network.remote-dio: on
> performance.io-cache: off
> performance.read-ahead: off
> performance.quick-read: off
> cluster.eager-lock: enable
>
> Apparently that's done by the 'optimize for virt store'
> checkbox, but
> obviously not when the volume is created manually. Having this
> in the
> documentation on ovirt.org <http://ovirt.org> would have saved
> me a lot of time and
> frustration.
>
>
>
> its a wiki, how about adding this for the next guy?
>
> thanks,
> Itamar
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org <mailto:Users at ovirt.org>
> http://lists.ovirt.org/mailman/listinfo/users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20140624/6a45ef54/attachment-0001.html>
More information about the Users
mailing list