[Users] Issue adding a gluster storage domain.
Daniel Rowe
daniel.fathom13 at gmail.com
Mon Oct 29 06:16:34 UTC 2012
Hi
I can't seem to get a gluster storage domain added I am using Fedora
17 on both the nodes management machine. I have the gluster volumes
showing in ovirt and I can manually mount the gluster volume both
locally on the node and on the management machine.
If I manually mount the volume on the node with /usr/bin/mount -t
glusterfs -o vers=3 localhost:/bsdvol1
/rhev/data-center/mnt/localhost:_bsdvol1 as root then add the domain
in the web interface it works.
Although I have the domain done with the manual mounting, I am
wondering what is going on.
Thread-1934::INFO::2012-10-29
13:27:09,793::logUtils::37::dispatcher::(wrapper) Run and protect:
validateStorageServerConnection(domType=6,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '',
'connection': 'localhost:/bsdvol1', 'iqn': '', 'portal': '', 'user':
'', 'vfs_type': 'glusterfs', 'password': '******', 'id':
'00000000-0000-0000-0000-000000000000'}], options=None)
Thread-1934::INFO::2012-10-29
13:27:09,793::logUtils::39::dispatcher::(wrapper) Run and protect:
validateStorageServerConnection, Return response: {'statuslist':
[{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-1934::DEBUG::2012-10-29
13:27:09,793::task::1172::TaskManager.Task::(prepare)
Task=`4f9131eb-45cd-4e82-bc35-e29ed14f0818`::finished: {'statuslist':
[{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-1934::DEBUG::2012-10-29
13:27:09,793::task::588::TaskManager.Task::(_updateState)
Task=`4f9131eb-45cd-4e82-bc35-e29ed14f0818`::moving from state
preparing -> state finished
Thread-1934::DEBUG::2012-10-29
13:27:09,794::resourceManager::809::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-1934::DEBUG::2012-10-29
13:27:09,794::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-1934::DEBUG::2012-10-29
13:27:09,794::task::978::TaskManager.Task::(_decref)
Task=`4f9131eb-45cd-4e82-bc35-e29ed14f0818`::ref 0 aborting False
Thread-1935::DEBUG::2012-10-29
13:27:09,842::BindingXMLRPC::156::vds::(wrapper) [192.168.1.10]
Thread-1935::DEBUG::2012-10-29
13:27:09,842::task::588::TaskManager.Task::(_updateState)
Task=`ef920d52-3082-495f-bfbc-2eb508c3de18`::moving from state init ->
state preparing
Thread-1935::INFO::2012-10-29
13:27:09,843::logUtils::37::dispatcher::(wrapper) Run and protect:
connectStorageServer(domType=6,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '',
'connection': 'localhost:/bsdvol1', 'mnt_options': 'vers=3', 'portal':
'', 'user': '', 'iqn': '', 'vfs_type': 'glusterfs', 'password':
'******', 'id': 'bbe1bbc5-62b8-4115-b409-6ddea910a688'}],
options=None)
Thread-1935::DEBUG::2012-10-29
13:27:09,851::__init__::1249::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /usr/bin/mount -t glusterfs -o vers=3
localhost:/bsdvol1 /rhev/data-center/mnt/localhost:_bsdvol1' (cwd
None)
Thread-1935::ERROR::2012-10-29
13:27:09,930::hsm::1932::Storage.HSM::(connectStorageServer) Could not
connect to storageServer
Traceback (most recent call last):
File "/usr/share/vdsm/storage/hsm.py", line 1929, in connectStorageServer
conObj.connect()
File "/usr/share/vdsm/storage/storageServer.py", line 179, in connect
self._mount.mount(self.options, self._vfsType)
File "/usr/share/vdsm/storage/mount.py", line 190, in mount
return self._runcmd(cmd, timeout)
File "/usr/share/vdsm/storage/mount.py", line 206, in _runcmd
raise MountError(rc, ";".join((out, err)))
MountError: (1, 'unknown option vers (ignored)\nMount failed. Please
check the log file for more details.\n;ERROR: failed to create logfile
"/var/log/glusterfs/rhev-data-center-mnt-localhost:_bsdvol1.log"
(Permission denied)\nERROR: failed to open logfile
/var/log/glusterfs/rhev-data-center-mnt-localhost:_bsdvol1.log\n')
Thread-1935::DEBUG::2012-10-29
13:27:09,931::lvm::457::OperationMutex::(_invalidateAllPvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-1935::DEBUG::2012-10-29
13:27:09,932::lvm::459::OperationMutex::(_invalidateAllPvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-1935::DEBUG::2012-10-29
13:27:09,932::lvm::469::OperationMutex::(_invalidateAllVgs) Operation
'lvm invalidate operation' got the operation mutex
Thread-1935::DEBUG::2012-10-29
13:27:09,932::lvm::471::OperationMutex::(_invalidateAllVgs) Operation
'lvm invalidate operation' released the operation mutex
Thread-1935::DEBUG::2012-10-29
13:27:09,933::lvm::490::OperationMutex::(_invalidateAllLvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-1935::DEBUG::2012-10-29
13:27:09,933::lvm::492::OperationMutex::(_invalidateAllLvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-1935::INFO::2012-10-29
13:27:09,934::logUtils::39::dispatcher::(wrapper) Run and protect:
connectStorageServer, Return response: {'statuslist': [{'status': 477,
'id': 'bbe1bbc5-62b8-4115-b409-6ddea910a688'}]}
Thread-1935::DEBUG::2012-10-29
13:27:09,934::task::1172::TaskManager.Task::(prepare)
Task=`ef920d52-3082-495f-bfbc-2eb508c3de18`::finished: {'statuslist':
[{'status': 477, 'id': 'bbe1bbc5-62b8-4115-b409-6ddea910a688'}]}
Thread-1935::DEBUG::2012-10-29
13:27:09,934::task::588::TaskManager.Task::(_updateState)
Task=`ef920d52-3082-495f-bfbc-2eb508c3de18`::moving from state
preparing -> state finished
Thread-1935::DEBUG::2012-10-29
13:27:09,934::resourceManager::809::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-1935::DEBUG::2012-10-29
13:27:09,935::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-1935::DEBUG::2012-10-29
13:27:09,935::task::978::TaskManager.Task::(_decref)
Task=`ef920d52-3082-495f-bfbc-2eb508c3de18`::ref 0 aborting False
Thread-1937::DEBUG::2012-10-29
13:27:10,073::BindingXMLRPC::156::vds::(wrapper) [192.168.1.10]
Thread-1937::DEBUG::2012-10-29
13:27:10,073::task::588::TaskManager.Task::(_updateState)
Task=`322f7d1b-0e8d-458b-baf2-5c7f8c5f4303`::moving from state init ->
state preparing
Thread-1937::INFO::2012-10-29
13:27:10,073::logUtils::37::dispatcher::(wrapper) Run and protect:
createStorageDomain(storageType=6,
sdUUID='d8591d07-1611-46cc-9122-fc7badea0f4c',
domainName='DCS_Bis_Sys_gluster',
typeSpecificArg='localhost:/bsdvol1', domClass=1, domVersion='0',
options=None)
Thread-1937::DEBUG::2012-10-29
13:27:10,074::misc::1053::SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
Thread-1937::DEBUG::2012-10-29
13:27:10,074::misc::1055::SamplingMethod::(__call__) Got in to
sampling method
Thread-1937::DEBUG::2012-10-29
13:27:10,074::misc::1053::SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
Thread-1937::DEBUG::2012-10-29
13:27:10,074::misc::1055::SamplingMethod::(__call__) Got in to
sampling method
Thread-1937::DEBUG::2012-10-29
13:27:10,075::__init__::1249::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /sbin/iscsiadm -m session -R' (cwd None)
Thread-1937::DEBUG::2012-10-29
13:27:10,092::__init__::1249::Storage.Misc.excCmd::(_log) FAILED:
<err> = 'iscsiadm: No session found.\n'; <rc> = 21
Thread-1937::DEBUG::2012-10-29
13:27:10,092::misc::1063::SamplingMethod::(__call__) Returning last
result
Thread-1937::DEBUG::2012-10-29
13:27:15,382::__init__::1249::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /sbin/multipath' (cwd None)
Thread-1937::DEBUG::2012-10-29
13:27:15,414::__init__::1249::Storage.Misc.excCmd::(_log) SUCCESS:
<err> = ''; <rc> = 0
Thread-1937::DEBUG::2012-10-29
13:27:15,415::lvm::457::OperationMutex::(_invalidateAllPvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-1937::DEBUG::2012-10-29
13:27:15,415::lvm::459::OperationMutex::(_invalidateAllPvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-1937::DEBUG::2012-10-29
13:27:15,415::lvm::469::OperationMutex::(_invalidateAllVgs) Operation
'lvm invalidate operation' got the operation mutex
Thread-1937::DEBUG::2012-10-29
13:27:15,416::lvm::471::OperationMutex::(_invalidateAllVgs) Operation
'lvm invalidate operation' released the operation mutex
Thread-1937::DEBUG::2012-10-29
13:27:15,416::lvm::490::OperationMutex::(_invalidateAllLvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-1937::DEBUG::2012-10-29
13:27:15,416::lvm::492::OperationMutex::(_invalidateAllLvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-1937::DEBUG::2012-10-29
13:27:15,416::misc::1063::SamplingMethod::(__call__) Returning last
result
Thread-1937::DEBUG::2012-10-29
13:27:15,417::lvm::349::OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
Thread-1937::DEBUG::2012-10-29
13:27:15,419::__init__::1249::Storage.Misc.excCmd::(_log)
'/usr/bin/sudo -n /sbin/lvm vgs --config " devices { preferred_names =
[\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0
disable_after_error_count=3 filter = [
\\"a%36848f690e0b7860017e62e4257089a6e|36848f690e0b78600181b77200a8142b8%\\",
\\"r%.*%\\" ] } global { locking_type=1 prioritise_write_locks=1
wait_for_locks=1 } backup { retain_min = 50 retain_days = 0 } "
--noheadings --units b --nosuffix --separator | -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free
d8591d07-1611-46cc-9122-fc7badea0f4c' (cwd None)
Thread-1937::DEBUG::2012-10-29
13:27:15,536::__init__::1249::Storage.Misc.excCmd::(_log) FAILED:
<err> = ' Volume group "d8591d07-1611-46cc-9122-fc7badea0f4c" not
found\n'; <rc> = 5
Thread-1937::WARNING::2012-10-29
13:27:15,538::lvm::353::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 []
[' Volume group "d8591d07-1611-46cc-9122-fc7badea0f4c" not found']
Thread-1937::DEBUG::2012-10-29
13:27:15,538::lvm::376::OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
Thread-1937::INFO::2012-10-29
13:27:15,542::nfsSD::64::Storage.StorageDomain::(create)
sdUUID=d8591d07-1611-46cc-9122-fc7badea0f4c
domainName=DCS_Bis_Sys_gluster remotePath=localhost:/bsdvol1
domClass=1
Thread-1937::ERROR::2012-10-29
13:27:15,550::task::853::TaskManager.Task::(_setError)
Task=`322f7d1b-0e8d-458b-baf2-5c7f8c5f4303`::Unexpected error
Traceback (most recent call last):
File "/usr/share/vdsm/storage/task.py", line 861, in _run
return fn(*args, **kargs)
File "/usr/share/vdsm/logUtils.py", line 38, in wrapper
res = f(*args, **kwargs)
File "/usr/share/vdsm/storage/hsm.py", line 2136, in createStorageDomain
typeSpecificArg, storageType, domVersion)
File "/usr/share/vdsm/storage/nfsSD.py", line 75, in create
cls._preCreateValidation(sdUUID, mntPoint, remotePath, version)
File "/usr/share/vdsm/storage/nfsSD.py", line 44, in _preCreateValidation
raise se.StorageDomainFSNotMounted(domPath)
StorageDomainFSNotMounted: Storage domain remote path not mounted:
('/rhev/data-center/mnt/localhost:_bsdvol1',)
[root at bsdvmhvnode01 mnt]# cat
/var/log/glusterfs/rhev-data-center-mnt-localhost:_bsdvol1.log
[2012-10-29 13:27:42.311022] I [glusterfsd.c:1666:main]
0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version
3.3.1
[2012-10-29 13:27:42.328054] I
[io-cache.c:1549:check_cache_ users at ovirt.org. size_ok]
0-bsdvol1-quick-read: Max cache size is 50636935168
[2012-10-29 13:27:42.328157] I [io-cache.c:1549:check_cache_size_ok]
0-bsdvol1-io-cache: Max cache size is 50636935168
[2012-10-29 13:27:42.329037] I [client.c:2142:notify]
0-bsdvol1-client-0: parent translators are ready, attempting connect
on transport
Given volfile:
+------------------------------------------------------------------------------+
1: volume bsdvol1-client-0
2: type protocol/client
3: option remote-host 192.168.1.11
4: option remote-subvolume /data/bsdvol1
5: option transport-type tcp
6: option username ed9b899c-26c1-4d8b-a2f8-ffd2d0b10fed
7: option password 37a8f9ed-e27c-41cb-a049-da6a65042920
8: end-volume
9:
10: volume bsdvol1-write-behind
11: type performance/write-behind
12: subvolumes bsdvol1-client-0
13: end-volume
14:
15: volume bsdvol1-read-ahead
16: type performance/read-ahead
17: subvolumes bsdvol1-write-behind
18: end-volume
19:
20: volume bsdvol1-io-cache
21: type performance/io-cache
22: subvolumes bsdvol1-read-ahead
23: end-volume
24:
25: volume bsdvol1-quick-read
26: type performance/quick-read
27: subvolumes bsdvol1-io-cache
28: end-volume
29:
30: volume bsdvol1-md-cache
31: type performance/md-cache
32: subvolumes bsdvol1-quick-read
33: end-volume
34:
35: volume bsdvol1
36: type debug/io-stats
37: option latency-measurement off
38: option count-fop-hits off
39: subvolumes bsdvol1-md-cache
40: end-volume
+------------------------------------------------------------------------------+
[2012-10-29 13:27:42.334438] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
0-bsdvol1-client-0: changing port to 24009 (from 0)
[2012-10-29 13:27:46.331768] I
[client-handshake.c:1636:select_server_supported_programs]
0-bsdvol1-client-0: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2012-10-29 13:27:46.332216] I
[client-handshake.c:1433:client_setvolume_cbk] 0-bsdvol1-client-0:
Connected to 192.168.1.11:24009, attached to remote volume
'/data/bsdvol1'.
[2012-10-29 13:27:46.332249] I
[client-handshake.c:1445:client_setvolume_cbk] 0-bsdvol1-client-0:
Server and Client lk-version numbers are not same, reopening the fds
[2012-10-29 13:27:46.337041] I [fuse-bridge.c:4191:fuse_graph_setup]
0-fuse: switched to graph 0
[2012-10-29 13:27:46.337157] I
[client-handshake.c:453:client_set_lk_version_cbk] 0-bsdvol1-client-0:
Server lk version = 1
[2012-10-29 13:27:46.337466] I [fuse-bridge.c:3376:fuse_init]
0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13
kernel 7.20
More information about the Users
mailing list