Posix FS storage
-----Original Message-----
From: Itamar Heim [mailto:iheim@redhat.com]
Sent: Monday, June 18, 2012 6:29 PM
To: зоррыч
Cc: 'Vijay Bellur'; 'Robert Middleswarth'; users(a)ovirt.org; 'Daniel
Paikov'
Subject: Re: [Users] Ovirt 3.1 and gluster (creation in ovirt)
On 06/18/2012 04:50 PM, зоррыч wrote:
Any ideas for solutions?
Is this a bug?
*From:*users-bounces@ovirt.org [mailto:users-bounces@ovirt.org] *On
Behalf Of *зоррыч
*Sent:* Sunday, June 17, 2012 12:04 AM
*To:* 'Vijay Bellur'; 'Robert Middleswarth'
*Cc:* users(a)ovirt.org; 'Daniel Paikov'
*Subject:* Re: [Users] Ovirt 3.1 and gluster (creation in ovirt)
I have updated GlusterFS and volume successfully created
Thank you!
But I was not able to mount a storage domain.
Vdsm.log:
Thread-21025::DEBUG::2012-06-16
15:43:21,495::BindingXMLRPC::160::vds::(wrapper) [10.1.20.2]
Thread-21025::DEBUG::2012-06-16
15:43:21,495::task::588::TaskManager.Task::(_updateState)
Task=`8d841c96-43e3-4d4b-b115-a36c4adf695a`::moving from state init ->
state preparing
Thread-21025::INFO::2012-06-16
15:43:21,503::logUtils::37::dispatcher::(wrapper) Run and protect:
validateStorageServerConnection(domType=6,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port':
'',
'connection': '10.1.20.7:/sd2', 'iqn': '',
'portal': '', 'user': '',
'vfs_type': 'glusterfs', 'password': '******',
'id':
'00000000-0000-0000-0000-000000000000'}], options=None)
Thread-21025::INFO::2012-06-16
15:43:21,503::logUtils::39::dispatcher::(wrapper) Run and protect:
validateStorageServerConnection, Return response: {'statuslist':
[{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-21025::DEBUG::2012-06-16
15:43:21,503::task::1172::TaskManager.Task::(prepare)
Task=`8d841c96-43e3-4d4b-b115-a36c4adf695a`::finished: {'statuslist':
[{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-21025::DEBUG::2012-06-16
15:43:21,503::task::588::TaskManager.Task::(_updateState)
Task=`8d841c96-43e3-4d4b-b115-a36c4adf695a`::moving from state
preparing
-> state finished
Thread-21025::DEBUG::2012-06-16
15:43:21,503::resourceManager::809::ResourceManager.Owner::(releaseAll
) Owner.releaseAll requests {} resources {}
Thread-21025::DEBUG::2012-06-16
15:43:21,504::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-21025::DEBUG::2012-06-16
15:43:21,504::task::978::TaskManager.Task::(_decref)
Task=`8d841c96-43e3-4d4b-b115-a36c4adf695a`::ref 0 aborting False
Thread-21026::DEBUG::2012-06-16
15:43:21,526::BindingXMLRPC::160::vds::(wrapper) [10.1.20.2]
Thread-21026::DEBUG::2012-06-16
15:43:21,526::task::588::TaskManager.Task::(_updateState)
Task=`2a6538e5-b961-478a-bce6-f5ded1a62bca`::moving from state init ->
state preparing
Thread-21026::INFO::2012-06-16
15:43:21,527::logUtils::37::dispatcher::(wrapper) Run and protect:
connectStorageServer(domType=6,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port':
'',
'connection': '10.1.20.7:/sd2', 'iqn': '',
'portal': '', 'user': '',
'vfs_type': 'glusterfs', 'password': '******',
'id':
'e7766e1d-f2c6-45ee-900e-00c6689649cd'}], options=None)
Thread-21026::DEBUG::2012-06-16
15:43:21,530::lvm::460::OperationMutex::(_invalidateAllPvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-21026::DEBUG::2012-06-16
15:43:21,531::lvm::462::OperationMutex::(_invalidateAllPvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-21026::DEBUG::2012-06-16
15:43:21,531::lvm::472::OperationMutex::(_invalidateAllVgs) Operation
'lvm invalidate operation' got the operation mutex
Thread-21026::DEBUG::2012-06-16
15:43:21,531::lvm::474::OperationMutex::(_invalidateAllVgs) Operation
'lvm invalidate operation' released the operation mutex
Thread-21026::DEBUG::2012-06-16
15:43:21,531::lvm::493::OperationMutex::(_invalidateAllLvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-21026::DEBUG::2012-06-16
15:43:21,531::lvm::495::OperationMutex::(_invalidateAllLvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-21026::INFO::2012-06-16
15:43:21,532::logUtils::39::dispatcher::(wrapper) Run and protect:
connectStorageServer, Return response: {'statuslist': [{'status': 0,
'id': 'e7766e1d-f2c6-45ee-900e-00c6689649cd'}]}
Thread-21026::DEBUG::2012-06-16
15:43:21,532::task::1172::TaskManager.Task::(prepare)
Task=`2a6538e5-b961-478a-bce6-f5ded1a62bca`::finished: {'statuslist':
[{'status': 0, 'id': 'e7766e1d-f2c6-45ee-900e-00c6689649cd'}]}
Thread-21026::DEBUG::2012-06-16
15:43:21,532::task::588::TaskManager.Task::(_updateState)
Task=`2a6538e5-b961-478a-bce6-f5ded1a62bca`::moving from state
preparing
-> state finished
Thread-21026::DEBUG::2012-06-16
15:43:21,532::resourceManager::809::ResourceManager.Owner::(releaseAll
) Owner.releaseAll requests {} resources {}
Thread-21026::DEBUG::2012-06-16
15:43:21,532::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-21026::DEBUG::2012-06-16
15:43:21,532::task::978::TaskManager.Task::(_decref)
Task=`2a6538e5-b961-478a-bce6-f5ded1a62bca`::ref 0 aborting False
Thread-21027::DEBUG::2012-06-16
15:43:21,610::BindingXMLRPC::160::vds::(wrapper) [10.1.20.2]
Thread-21027::DEBUG::2012-06-16
15:43:21,611::task::588::TaskManager.Task::(_updateState)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::moving from state init ->
state preparing
Thread-21027::INFO::2012-06-16
15:43:21,611::logUtils::37::dispatcher::(wrapper) Run and protect:
createStorageDomain(storageType=6,
sdUUID='711293b8-019c-4f41-8cab-df03dd843556', domainName='dfdf',
typeSpecificArg='10.1.20.7:/sd2', domClass=1, domVersion='0',
options=None)
Thread-21027::DEBUG::2012-06-16
15:43:21,611::misc::1053::SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
Thread-21027::DEBUG::2012-06-16
15:43:21,611::misc::1055::SamplingMethod::(__call__) Got in to
sampling method
Thread-21027::DEBUG::2012-06-16
15:43:21,612::misc::1053::SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
Thread-21027::DEBUG::2012-06-16
15:43:21,612::misc::1055::SamplingMethod::(__call__) Got in to
sampling method
Thread-21027::DEBUG::2012-06-16
15:43:21,612::__init__::1164::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /sbin/iscsiadm -m session -R' (cwd None)
Thread-21027::DEBUG::2012-06-16
15:43:21,639::__init__::1164::Storage.Misc.excCmd::(_log) FAILED:
<err> = 'iscsiadm: No session found.\n'; <rc> = 21
Thread-21027::DEBUG::2012-06-16
15:43:21,639::misc::1063::SamplingMethod::(__call__) Returning last
result
Thread-21027::DEBUG::2012-06-16
15:43:21,805::__init__::1164::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /sbin/multipath' (cwd None)
Thread-21027::DEBUG::2012-06-16
15:43:21,863::__init__::1164::Storage.Misc.excCmd::(_log) SUCCESS:
<err> = ''; <rc> = 0
Thread-21027::DEBUG::2012-06-16
15:43:21,864::lvm::460::OperationMutex::(_invalidateAllPvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-21027::DEBUG::2012-06-16
15:43:21,864::lvm::462::OperationMutex::(_invalidateAllPvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-21027::DEBUG::2012-06-16
15:43:21,864::lvm::472::OperationMutex::(_invalidateAllVgs) Operation
'lvm invalidate operation' got the operation mutex
Thread-21027::DEBUG::2012-06-16
15:43:21,865::lvm::474::OperationMutex::(_invalidateAllVgs) Operation
'lvm invalidate operation' released the operation mutex
Thread-21027::DEBUG::2012-06-16
15:43:21,865::lvm::493::OperationMutex::(_invalidateAllLvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-21027::DEBUG::2012-06-16
15:43:21,865::lvm::495::OperationMutex::(_invalidateAllLvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-21027::DEBUG::2012-06-16
15:43:21,865::misc::1063::SamplingMethod::(__call__) Returning last
result
Thread-21027::DEBUG::2012-06-16
15:43:21,866::lvm::352::OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
Thread-21027::DEBUG::2012-06-16
15:43:21,867::__init__::1164::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /sbin/lvm vgs --config " devices { preferred_names =
[\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0
disable_after_error_count=3 filter = [ \\"a%35000c50001770ea3%\\
<file:///\\%22a%2535000c50001770ea3%25\>", \\"r%.*%\\
<file:///\\%22r%25.*%25\>" ] } global { locking_type=1
prioritise_write_locks=1 wait_for_locks=1 } backup { retain_min = 50
retain_days = 0 } " --noheadings --units b --nosuffix --separator | -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_m
da_size,vg_mda_free 711293b8-019c-4f41-8cab-df03dd843556' (cwd None)
Thread-21027::DEBUG::2012-06-16
15:43:22,013::__init__::1164::Storage.Misc.excCmd::(_log) FAILED:
<err> = ' Volume group "711293b8-019c-4f41-8cab-df03dd843556" not
found\n'; <rc> = 5
Thread-21027::WARNING::2012-06-16
15:43:22,014::lvm::356::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 []
[' Volume group "711293b8-019c-4f41-8cab-df03dd843556" not found']
Thread-21027::DEBUG::2012-06-16
15:43:22,016::lvm::379::OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
Thread-21027::INFO::2012-06-16
15:43:22,020::nfsSD::64::Storage.StorageDomain::(create)
sdUUID=711293b8-019c-4f41-8cab-df03dd843556 domainName=dfdf
remotePath=10.1.20.7:/sd2 domClass=1
Thread-21027::DEBUG::2012-06-16
15:43:22,037::persistentDict::185::Storage.PersistentDict::(__init__)
Created a persistent dict with FileMetadataRW backend
Thread-21027::DEBUG::2012-06-16
15:43:22,038::persistentDict::226::Storage.PersistentDict::(refresh)
read lines (FileMetadataRW)=[]
Thread-21027::WARNING::2012-06-16
15:43:22,038::persistentDict::248::Storage.PersistentDict::(refresh)
data has no embedded checksum - trust it as it is
Thread-21027::DEBUG::2012-06-16
15:43:22,038::persistentDict::162::Storage.PersistentDict::(transactio
n)
Starting transaction
Thread-21027::DEBUG::2012-06-16
15:43:22,038::persistentDict::168::Storage.PersistentDict::(transactio
n)
Flushing changes
Thread-21027::DEBUG::2012-06-16
15:43:22,039::persistentDict::287::Storage.PersistentDict::(flush)
about to write lines (FileMetadataRW)=['CLASS=Data',
'DESCRIPTION=dfdf', 'IOOPTIMEOUTSEC=1', 'LEASERETRIES=3',
'LEASETIMESEC=5', 'LOCKPOLICY=', 'LOCKRENEWALINTERVALSEC=5',
'POOL_UUID=', 'REMOTE_PATH=10.1.20.7:/sd2', 'ROLE=Regular',
'SDUUID=711293b8-019c-4f41-8cab-df03dd843556',
'TYPE=SHAREDFS', 'VERSION=0',
'_SHA_CKSUM=2e3c4fc88aa713dedbb0d708375966e158327797']
Thread-21027::DEBUG::2012-06-16
15:43:22,041::persistentDict::170::Storage.PersistentDict::(transactio
n)
Finished transaction
Thread-21027::DEBUG::2012-06-16
15:43:22,042::fileSD::107::Storage.StorageDomain::(__init__) Reading
domain in path
/rhev/data-center/mnt/10.1.20.7:_sd2/711293b8-019c-4f41-8cab-df03dd843
556
Thread-21027::DEBUG::2012-06-16
15:43:22,042::persistentDict::185::Storage.PersistentDict::(__init__)
Created a persistent dict with FileMetadataRW backend
Thread-21027::ERROR::2012-06-16
15:43:22,043::task::853::TaskManager.Task::(_setError)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::Unexpected error
Traceback (most recent call last):
File "/usr/share/vdsm/storage/task.py", line 861, in _run
return fn(*args, **kargs)
File "/usr/share/vdsm/logUtils.py", line 38, in wrapper
res = f(*args, **kwargs)
File "/usr/share/vdsm/storage/hsm.py", line 2136, in
createStorageDomain
typeSpecificArg, storageType, domVersion)
File "/usr/share/vdsm/storage/nfsSD.py", line 90, in create
fsd = cls(os.path.join(mntPoint, sdUUID))
File "/usr/share/vdsm/storage/fileSD.py", line 113, in __init__
sdUUID = metadata[sd.DMDK_SDUUID]
File "/usr/share/vdsm/storage/persistentDict.py", line 85, in
__getitem__
return dec(self._dict[key])
File "/usr/share/vdsm/storage/persistentDict.py", line 193, in
__getitem__
with self._accessWrapper():
File "/usr/lib64/python2.6/contextlib.py", line 16, in __enter__
return self.gen.next()
File "/usr/share/vdsm/storage/persistentDict.py", line 147, in
_accessWrapper
self.refresh()
File "/usr/share/vdsm/storage/persistentDict.py", line 224, in refresh
lines = self._metaRW.readlines()
File "/usr/share/vdsm/storage/fileSD.py", line 82, in readlines
return misc.stripNewLines(self._oop.directReadLines(self._metafile))
File "/usr/share/vdsm/storage/processPool.py", line 63, in wrapper
return self.runExternally(func, *args, **kwds)
File "/usr/share/vdsm/storage/processPool.py", line 74, in
runExternally
return self._procPool.runExternally(*args, **kwargs)
File "/usr/share/vdsm/storage/processPool.py", line 170, in
runExternally
raise err
OSError: [Errno 22] Invalid argument:
'/rhev/data-center/mnt/10.1.20.7:_sd2/711293b8-019c-4f41-8cab-df03dd843556/dom_md/metadata'
Thread-21027::DEBUG::2012-06-16
15:43:22,044::task::872::TaskManager.Task::(_run)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::Task._run:
fdeab954-7958-4cf3-8dd3-09e06329bf92 (6,
'711293b8-019c-4f41-8cab-df03dd843556', 'dfdf', '10.1.20.7:/sd2',
1,
'0') {} failed - stopping task
Thread-21027::DEBUG::2012-06-16
15:43:22,044::task::1199::TaskManager.Task::(stop)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::stopping in state
preparing (force False)
Thread-21027::DEBUG::2012-06-16
15:43:22,044::task::978::TaskManager.Task::(_decref)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::ref 1 aborting True
Thread-21027::INFO::2012-06-16
15:43:22,045::task::1157::TaskManager.Task::(prepare)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::aborting: Task is aborted:
u"[Errno 22] Invalid argument:
'/rhev/data-center/mnt/10.1.20.7:_sd2/711293b8-019c-4f41-8cab-df03dd843556/dom_md/metadata'"
- code 100
Thread-21027::DEBUG::2012-06-16
15:43:22,045::task::1162::TaskManager.Task::(prepare)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::Prepare: aborted: [Errno
22] Invalid argument:
'/rhev/data-center/mnt/10.1.20.7:_sd2/711293b8-019c-4f41-8cab-df03dd843556/dom_md/metadata'
Thread-21027::DEBUG::2012-06-16
15:43:22,045::task::978::TaskManager.Task::(_decref)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::ref 0 aborting True
Thread-21027::DEBUG::2012-06-16
15:43:22,045::task::913::TaskManager.Task::(_doAbort)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::Task._doAbort: force
False
Thread-21027::DEBUG::2012-06-16
15:43:22,045::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-21027::DEBUG::2012-06-16
15:43:22,046::task::588::TaskManager.Task::(_updateState)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::moving from state
preparing
-> state aborting
Thread-21027::DEBUG::2012-06-16
15:43:22,046::task::537::TaskManager.Task::(__state_aborting)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::_aborting: recover policy
none
Thread-21027::DEBUG::2012-06-16
15:43:22,046::task::588::TaskManager.Task::(_updateState)
Task=`fdeab954-7958-4cf3-8dd3-09e06329bf92`::moving from state
aborting
-> state failed
Thread-21027::DEBUG::2012-06-16
15:43:22,046::resourceManager::809::ResourceManager.Owner::(releaseAll
) Owner.releaseAll requests {} resources {}
Thread-21027::DEBUG::2012-06-16
15:43:22,047::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-21027::ERROR::2012-06-16
15:43:22,047::dispatcher::69::Storage.Dispatcher.Protect::(run) [Errno
22] Invalid argument:
'/rhev/data-center/mnt/10.1.20.7:_sd2/711293b8-019c-4f41-8cab-df03dd843556/dom_md/metadata'
Traceback (most recent call last):
File "/usr/share/vdsm/storage/dispatcher.py", line 61, in run
result = ctask.prepare(self.func, *args, **kwargs)
File "/usr/share/vdsm/storage/task.py", line 1164, in prepare
raise self.error
OSError: [Errno 22] Invalid argument:
'/rhev/data-center/mnt/10.1.20.7:_sd2/711293b8-019c-4f41-8cab-df03dd843556/dom_md/metadata'
Thread-21035::DEBUG::2012-06-16
15:43:26,510::task::588::TaskManager.Task::(_updateState)
Task=`283d6276-7bdd-4c35-ad6f-6e8514c277f0`::moving from state init ->
state preparing
Thread-21035::INFO::2012-06-16
15:43:26,510::logUtils::37::dispatcher::(wrapper) Run and protect:
repoStats(options=None)
Thread-21035::INFO::2012-06-16
15:43:26,511::logUtils::39::dispatcher::(wrapper) Run and protect:
repoStats, Return response: {}
Thread-21035::DEBUG::2012-06-16
15:43:26,511::task::1172::TaskManager.Task::(prepare)
Task=`283d6276-7bdd-4c35-ad6f-6e8514c277f0`::finished: {}
Thread-21035::DEBUG::2012-06-16
15:43:26,511::task::588::TaskManager.Task::(_updateState)
Task=`283d6276-7bdd-4c35-ad6f-6e8514c277f0`::moving from state
preparing
-> state finished
Thread-21035::DEBUG::2012-06-16
15:43:26,511::resourceManager::809::ResourceManager.Owner::(releaseAll
) Owner.releaseAll requests {} resources {}
Thread-21035::DEBUG::2012-06-16
15:43:26,511::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-21035::DEBUG::2012-06-16
15:43:26,511::task::978::TaskManager.Task::(_decref)
Task=`283d6276-7bdd-4c35-ad6f-6e8514c277f0`::ref 0 aborting False
^C
[root@noc-3-synt mnt]# cat
/rhev/data-center/mnt/10.1.20.7:_sd2/711293b8-019c-4f41-8cab-df03dd843
556/dom_md/metadata
CLASS=Data
DESCRIPTION=dfdf
IOOPTIMEOUTSEC=1
LEASERETRIES=3
LEASETIMESEC=5
LOCKPOLICY=
LOCKRENEWALINTERVALSEC=5
POOL_UUID=
REMOTE_PATH=10.1.20.7:/sd2
ROLE=Regular
SDUUID=711293b8-019c-4f41-8cab-df03dd843556
TYPE=SHAREDFS
VERSION=0
_SHA_CKSUM=2e3c4fc88aa713dedbb0d708375966e158327797
*From:*Vijay Bellur [mailto:vbellur@redhat.com]
*Sent:* Saturday, June 16, 2012 10:19 PM
*To:* Robert Middleswarth
*Cc:* зоррыч; users(a)ovirt.org <mailto:users@ovirt.org>; Daniel Paikov
*Subject:* Re: [Users] Ovirt 3.1 and gluster (creation in ovirt)
On 06/16/2012 11:08 AM, Robert Middleswarth wrote:
I am seeing the same thing. I also notice that glusterfs seems to die
every-time I try. I am wonder if this could be a glusterfs / f17 issue.
Are you running GlusterFS 3.2.x in Fedora 17? For this volume creation
to complete successfully from oVirt, you will need GlusterFS 3.3.0.
You can download Fedora RPMs for 3.3.0 from:
http://download.gluster.com/pub/gluster/glusterfs/3.3/LATEST/Fedora/
-Vijay
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users