[Users] gluster storage

Robert Middleswarth robert at middleswarth.net
Wed Jul 18 14:31:15 UTC 2012


This may be a dumb questions but did you chown the storage to 36.36 
before attempting to mount?

aka

chown -R 36.36 /mnt/point

Thanks
Robert

On 07/18/2012 10:28 AM, ?????? wrote:
>
> I'm trying to mount manually created gluster storage.
>
> But getting the error:
>
> Vdsm.log:
>
> Thread-1530::DEBUG::2012-07-18 
> 09:59:41,604::BindingXMLRPC::160::vds::(wrapper) [10.1.20.2]
>
> Thread-1530::DEBUG::2012-07-18 
> 09:59:41,605::task::568::TaskManager.Task::(_updateState) 
> Task=`90d34839-2ea2-4be5-b9b5-b3f4c330dc16`::moving from state init -> 
> state preparing
>
> Thread-1530::INFO::2012-07-18 
> 09:59:41,605::logUtils::37::dispatcher::(wrapper) Run and protect: 
> validateStorageServerConnection(domType=6, 
> spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '', 
> 'connection': '127.0.0.1:/sd3', 'iqn': '', 'portal': '', 'user': '', 
> 'vfs_type': 'glusterfs', 'password': '******', 'id': 
> '00000000-0000-0000-0000-000000000000'}], options=None)
>
> Thread-1530::INFO::2012-07-18 
> 09:59:41,605::logUtils::39::dispatcher::(wrapper) Run and protect: 
> validateStorageServerConnection, Return response: {'statuslist': 
> [{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
>
> Thread-1530::DEBUG::2012-07-18 
> 09:59:41,605::task::1151::TaskManager.Task::(prepare) 
> Task=`90d34839-2ea2-4be5-b9b5-b3f4c330dc16`::finished: {'statuslist': 
> [{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
>
> Thread-1530::DEBUG::2012-07-18 
> 09:59:41,605::task::568::TaskManager.Task::(_updateState) 
> Task=`90d34839-2ea2-4be5-b9b5-b3f4c330dc16`::moving from state 
> preparing -> state finished
>
> Thread-1530::DEBUG::2012-07-18 
> 09:59:41,606::resourceManager::809::ResourceManager.Owner::(releaseAll) Owner.releaseAll 
> requests {} resources {}
>
> Thread-1530::DEBUG::2012-07-18 
> 09:59:41,606::resourceManager::844::ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {}
>
> Thread-1530::DEBUG::2012-07-18 
> 09:59:41,606::task::957::TaskManager.Task::(_decref) 
> Task=`90d34839-2ea2-4be5-b9b5-b3f4c330dc16`::ref 0 aborting False
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:41,634::BindingXMLRPC::160::vds::(wrapper) [10.1.20.2]
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:41,634::task::568::TaskManager.Task::(_updateState) 
> Task=`19e8db57-2942-495a-a58c-398b79646a65`::moving from state init -> 
> state preparing
>
> Thread-1531::INFO::2012-07-18 
> 09:59:41,635::logUtils::37::dispatcher::(wrapper) Run and protect: 
> connectStorageServer(domType=6, 
> spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '', 
> 'connection': '127.0.0.1:/sd3', 'iqn': '', 'portal': '', 'user': '', 
> 'vfs_type': 'glusterfs', 'password': '******', 'id': 
> 'dd989346-1bd1-4772-8cc3-a8913680af01'}], options=None)
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:41,637::__init__::1164::Storage.Misc.excCmd::(_log) 
> '/usr/bin/sudo -n /bin/mount -t glusterfs 127.0.0.1:/sd3 
> /rhev/data-center/mnt/127.0.0.1:_sd3' (cwd None)
>
> Thread-1533::DEBUG::2012-07-18 
> 09:59:43,910::task::568::TaskManager.Task::(_updateState) 
> Task=`d350c85f-51f9-4ee2-be80-1dfbc072e789`::moving from state init -> 
> state preparing
>
> Thread-1533::INFO::2012-07-18 
> 09:59:43,911::logUtils::37::dispatcher::(wrapper) Run and protect: 
> repoStats(options=None)
>
> Thread-1533::INFO::2012-07-18 
> 09:59:43,911::logUtils::39::dispatcher::(wrapper) Run and protect: 
> repoStats, Return response: {}
>
> Thread-1533::DEBUG::2012-07-18 
> 09:59:43,911::task::1151::TaskManager.Task::(prepare) 
> Task=`d350c85f-51f9-4ee2-be80-1dfbc072e789`::finished: {}
>
> Thread-1533::DEBUG::2012-07-18 
> 09:59:43,911::task::568::TaskManager.Task::(_updateState) 
> Task=`d350c85f-51f9-4ee2-be80-1dfbc072e789`::moving from state 
> preparing -> state finished
>
> Thread-1533::DEBUG::2012-07-18 
> 09:59:43,911::resourceManager::809::ResourceManager.Owner::(releaseAll) Owner.releaseAll 
> requests {} resources {}
>
> Thread-1533::DEBUG::2012-07-18 
> 09:59:43,911::resourceManager::844::ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {}
>
> Thread-1533::DEBUG::2012-07-18 
> 09:59:43,912::task::957::TaskManager.Task::(_decref) 
> Task=`d350c85f-51f9-4ee2-be80-1dfbc072e789`::ref 0 aborting False
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,719::lvm::477::OperationMutex::(_invalidateAllPvs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,720::lvm::479::OperationMutex::(_invalidateAllPvs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,720::lvm::488::OperationMutex::(_invalidateAllVgs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,720::lvm::490::OperationMutex::(_invalidateAllVgs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,721::lvm::508::OperationMutex::(_invalidateAllLvs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,721::lvm::510::OperationMutex::(_invalidateAllLvs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1531::INFO::2012-07-18 
> 09:59:45,721::logUtils::39::dispatcher::(wrapper) Run and protect: 
> connectStorageServer, Return response: {'statuslist': [{'status': 0, 
> 'id': 'dd989346-1bd1-4772-8cc3-a8913680af01'}]}
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,721::task::1151::TaskManager.Task::(prepare) 
> Task=`19e8db57-2942-495a-a58c-398b79646a65`::finished: {'statuslist': 
> [{'status': 0, 'id': 'dd989346-1bd1-4772-8cc3-a8913680af01'}]}
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,722::task::568::TaskManager.Task::(_updateState) 
> Task=`19e8db57-2942-495a-a58c-398b79646a65`::moving from state 
> preparing -> state finished
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,722::resourceManager::809::ResourceManager.Owner::(releaseAll) Owner.releaseAll 
> requests {} resources {}
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,722::resourceManager::844::ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {}
>
> Thread-1531::DEBUG::2012-07-18 
> 09:59:45,722::task::957::TaskManager.Task::(_decref) 
> Task=`19e8db57-2942-495a-a58c-398b79646a65`::ref 0 aborting False
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,802::BindingXMLRPC::160::vds::(wrapper) [10.1.20.2]
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,803::task::568::TaskManager.Task::(_updateState) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::moving from state init -> 
> state preparing
>
> Thread-1535::INFO::2012-07-18 
> 09:59:45,803::logUtils::37::dispatcher::(wrapper) Run and protect: 
> createStorageDomain(storageType=6, 
> sdUUID='2118b60e-d905-49c8-8379-3de2ef7e9711', domainName='xcv', 
> typeSpecificArg='127.0.0.1:/sd3', domClass=1, domVersion='0', 
> options=None)
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,803::misc::1054::SamplingMethod::(__call__) Trying to enter 
> sampling method (storage.sdc.refreshStorage)
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,803::misc::1056::SamplingMethod::(__call__) Got in to 
> sampling method
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,804::misc::1054::SamplingMethod::(__call__) Trying to enter 
> sampling method (storage.iscsi.rescan)
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,804::misc::1056::SamplingMethod::(__call__) Got in to 
> sampling method
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,804::__init__::1164::Storage.Misc.excCmd::(_log) 
> '/usr/bin/sudo -n /sbin/iscsiadm -m session -R' (cwd None)
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,819::__init__::1164::Storage.Misc.excCmd::(_log) FAILED: 
> <err> = 'iscsiadm: No session found.\n'; <rc> = 21
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,819::misc::1064::SamplingMethod::(__call__) Returning last result
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:45,984::__init__::1164::Storage.Misc.excCmd::(_log) 
> '/usr/bin/sudo -n /sbin/multipath' (cwd None)
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,021::__init__::1164::Storage.Misc.excCmd::(_log) SUCCESS: 
> <err> = ''; <rc> = 0
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,021::lvm::477::OperationMutex::(_invalidateAllPvs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,021::lvm::479::OperationMutex::(_invalidateAllPvs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,022::lvm::488::OperationMutex::(_invalidateAllVgs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,022::lvm::490::OperationMutex::(_invalidateAllVgs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,022::lvm::508::OperationMutex::(_invalidateAllLvs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,022::lvm::510::OperationMutex::(_invalidateAllLvs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,023::misc::1064::SamplingMethod::(__call__) Returning last result
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,023::lvm::368::OperationMutex::(_reloadvgs) Operation 'lvm 
> reload operation' got the operation mutex
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,024::__init__::1164::Storage.Misc.excCmd::(_log) 
> '/usr/bin/sudo -n /sbin/lvm vgs --config " devices { preferred_names = 
> [\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0 
> disable_after_error_count=3 filter = [ \\"a%35000c500048c2263%\\", 
> \\"r%.*%\\" ] }  global { locking_type=1  prioritise_write_locks=1  
> wait_for_locks=1 }  backup {  retain_min = 50  retain_days = 0 } " 
> --noheadings --units b --nosuffix --separator | -o 
> uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free 
> 2118b60e-d905-49c8-8379-3de2ef7e9711' (cwd None)
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,171::__init__::1164::Storage.Misc.excCmd::(_log) FAILED: 
> <err> = '  Volume group "2118b60e-d905-49c8-8379-3de2ef7e9711" not 
> found\n'; <rc> = 5
>
> Thread-1535::WARNING::2012-07-18 
> 09:59:46,172::lvm::373::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] 
> ['  Volume group "2118b60e-d905-49c8-8379-3de2ef7e9711" not found']
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,172::lvm::397::OperationMutex::(_reloadvgs) Operation 'lvm 
> reload operation' released the operation mutex
>
> Thread-1535::INFO::2012-07-18 
> 09:59:46,177::nfsSD::64::Storage.StorageDomain::(create) 
> sdUUID=2118b60e-d905-49c8-8379-3de2ef7e9711 domainName=xcv 
> remotePath=127.0.0.1:/sd3 domClass=1
>
> Thread-1535::ERROR::2012-07-18 
> 09:59:46,180::task::833::TaskManager.Task::(_setError) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::Unexpected error
>
> Traceback (most recent call last):
>
>   File "/usr/share/vdsm/storage/task.py", line 840, in _run
>
>     return fn(*args, **kargs)
>
>   File "/usr/share/vdsm/logUtils.py", line 38, in wrapper
>
>     res = f(*args, **kwargs)
>
>   File "/usr/share/vdsm/storage/hsm.py", line 2143, in createStorageDomain
>
>     typeSpecificArg, storageType, domVersion)
>
>   File "/usr/share/vdsm/storage/nfsSD.py", line 75, in create
>
> cls._preCreateValidation(sdUUID, mntPoint, remotePath, version)
>
>   File "/usr/share/vdsm/storage/nfsSD.py", line 46, in 
> _preCreateValidation
>
> fileSD.validateDirAccess(domPath)
>
>   File "/usr/share/vdsm/storage/fileSD.py", line 51, in validateDirAccess
>
> getProcPool().fileUtils.validateAccess(dirPath)
>
>   File "/usr/share/vdsm/storage/remoteFileHandler.py", line 270, in 
> callCrabRPCFunction
>
>     *args, **kwargs)
>
>   File "/usr/share/vdsm/storage/remoteFileHandler.py", line 186, in 
> callCrabRPCFunction
>
>     res, err = pickle.loads(rawResponse)
>
>   File "/usr/lib64/python2.6/pickle.py", line 1374, in loads
>
>     return Unpickler(file).load()
>
>   File "/usr/lib64/python2.6/pickle.py", line 858, in load
>
>     dispatch[key](self)
>
>   File "/usr/lib64/python2.6/pickle.py", line 1090, in load_global
>
>     klass = self.find_class(module, name)
>
>   File "/usr/lib64/python2.6/pickle.py", line 1124, in find_class
>
>     __import__(module)
>
> ImportError: No module named storage_exception
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,181::task::852::TaskManager.Task::(_run) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::Task._run: 
> 6d187cad-6851-48c0-9fc8-f19698b385ca (6, 
> '2118b60e-d905-49c8-8379-3de2ef7e9711', 'xcv', '127.0.0.1:/sd3', 1, 
> '0') {} failed - stopping task
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,181::task::1177::TaskManager.Task::(stop) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::stopping in state 
> preparing (force False)
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,181::task::957::TaskManager.Task::(_decref) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::ref 1 aborting True
>
> Thread-1535::INFO::2012-07-18 
> 09:59:46,181::task::1134::TaskManager.Task::(prepare) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::aborting: Task is 
> aborted: u'No module named storage_exception' - code 100
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,182::task::1139::TaskManager.Task::(prepare) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::Prepare: aborted: No 
> module named storage_exception
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,182::task::957::TaskManager.Task::(_decref) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::ref 0 aborting True
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,182::task::892::TaskManager.Task::(_doAbort) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::Task._doAbort: force False
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,182::resourceManager::844::ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {}
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,183::task::568::TaskManager.Task::(_updateState) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::moving from state 
> preparing -> state aborting
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,183::task::523::TaskManager.Task::(__state_aborting) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::_aborting: recover policy 
> none
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,183::task::568::TaskManager.Task::(_updateState) 
> Task=`6d187cad-6851-48c0-9fc8-f19698b385ca`::moving from state 
> aborting -> state failed
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,183::resourceManager::809::ResourceManager.Owner::(releaseAll) Owner.releaseAll 
> requests {} resources {}
>
> Thread-1535::DEBUG::2012-07-18 
> 09:59:46,183::resourceManager::844::ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {}
>
> Thread-1535::ERROR::2012-07-18 
> 09:59:46,184::dispatcher::69::Storage.Dispatcher.Protect::(run) No 
> module named storage_exception
>
> Traceback (most recent call last):
>
>   File "/usr/share/vdsm/storage/dispatcher.py", line 61, in run
>
>     result = ctask.prepare(self.func, *args, **kwargs)
>
>   File "/usr/share/vdsm/storage/task.py", line 1142, in prepare
>
>     raise self.error
>
> ImportError: No module named storage_exception
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,376::BindingXMLRPC::160::vds::(wrapper) [10.1.20.2]
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,377::task::568::TaskManager.Task::(_updateState) 
> Task=`37c8be8a-d5a9-4fca-b502-754795857aea`::moving from state init -> 
> state preparing
>
> Thread-1539::INFO::2012-07-18 
> 09:59:46,377::logUtils::37::dispatcher::(wrapper) Run and protect: 
> disconnectStorageServer(domType=6, 
> spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '', 
> 'connection': '127.0.0.1:/sd3', 'iqn': '', 'portal': '', 'user': '', 
> 'vfs_type': 'glusterfs', 'password': '******', 'id': 
> '00000000-0000-0000-0000-000000000000'}], options=None)
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,377::__init__::1164::Storage.Misc.excCmd::(_log) 
> '/usr/bin/sudo -n /bin/umount -f -l 
> /rhev/data-center/mnt/127.0.0.1:_sd3' (cwd None)
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,398::misc::1054::SamplingMethod::(__call__) Trying to enter 
> sampling method (storage.sdc.refreshStorage)
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,399::misc::1056::SamplingMethod::(__call__) Got in to 
> sampling method
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,399::misc::1054::SamplingMethod::(__call__) Trying to enter 
> sampling method (storage.iscsi.rescan)
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,399::misc::1056::SamplingMethod::(__call__) Got in to 
> sampling method
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,399::__init__::1164::Storage.Misc.excCmd::(_log) 
> '/usr/bin/sudo -n /sbin/iscsiadm -m session -R' (cwd None)
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,414::__init__::1164::Storage.Misc.excCmd::(_log) FAILED: 
> <err> = 'iscsiadm: No session found.\n'; <rc> = 21
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,415::misc::1064::SamplingMethod::(__call__) Returning last result
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,579::__init__::1164::Storage.Misc.excCmd::(_log) 
> '/usr/bin/sudo -n /sbin/multipath' (cwd None)
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,616::__init__::1164::Storage.Misc.excCmd::(_log) SUCCESS: 
> <err> = ''; <rc> = 0
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,616::lvm::477::OperationMutex::(_invalidateAllPvs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,616::lvm::479::OperationMutex::(_invalidateAllPvs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,617::lvm::488::OperationMutex::(_invalidateAllVgs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,617::lvm::490::OperationMutex::(_invalidateAllVgs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,617::lvm::508::OperationMutex::(_invalidateAllLvs) Operation 
> 'lvm invalidate operation' got the operation mutex
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,617::lvm::510::OperationMutex::(_invalidateAllLvs) Operation 
> 'lvm invalidate operation' released the operation mutex
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,618::misc::1064::SamplingMethod::(__call__) Returning last result
>
> Thread-1539::INFO::2012-07-18 
> 09:59:46,618::logUtils::39::dispatcher::(wrapper) Run and protect: 
> disconnectStorageServer, Return response: {'statuslist': [{'status': 
> 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,618::task::1151::TaskManager.Task::(prepare) 
> Task=`37c8be8a-d5a9-4fca-b502-754795857aea`::finished: {'statuslist': 
> [{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,618::task::568::TaskManager.Task::(_updateState) 
> Task=`37c8be8a-d5a9-4fca-b502-754795857aea`::moving from state 
> preparing -> state finished
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,618::resourceManager::809::ResourceManager.Owner::(releaseAll) Owner.releaseAll 
> requests {} resources {}
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,619::resourceManager::844::ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {}
>
> Thread-1539::DEBUG::2012-07-18 
> 09:59:46,619::task::957::TaskManager.Task::(_decref) 
> Task=`37c8be8a-d5a9-4fca-b502-754795857aea`::ref 0 aborting False
>
> Thread-1543::DEBUG::2012-07-18 
> 09:59:54,010::task::568::TaskManager.Task::(_updateState) 
> Task=`a1230d49-dae8-4a2d-ae39-543aa1e530bb`::moving from state init -> 
> state preparing
>
> Thread-1543::INFO::2012-07-18 
> 09:59:54,011::logUtils::37::dispatcher::(wrapper) Run and protect: 
> repoStats(options=None)
>
> Thread-1543::INFO::2012-07-18 
> 09:59:54,011::logUtils::39::dispatcher::(wrapper) Run and protect: 
> repoStats, Return response: {}
>
> Thread-1543::DEBUG::2012-07-18 
> 09:59:54,011::task::1151::TaskManager.Task::(prepare) 
> Task=`a1230d49-dae8-4a2d-ae39-543aa1e530bb`::finished: {}
>
> Thread-1543::DEBUG::2012-07-18 
> 09:59:54,011::task::568::TaskManager.Task::(_updateState) 
> Task=`a1230d49-dae8-4a2d-ae39-543aa1e530bb`::moving from state 
> preparing -> state finished
>
> Thread-1543::DEBUG::2012-07-18 
> 09:59:54,012::resourceManager::809::ResourceManager.Owner::(releaseAll) Owner.releaseAll 
> requests {} resources {}
>
> Thread-1543::DEBUG::2012-07-18 
> 09:59:54,012::resourceManager::844::ResourceManager.Owner::(cancelAll) 
> Owner.cancelAll requests {}
>
> Thread-1543::DEBUG::2012-07-18 
> 09:59:54,012::task::957::TaskManager.Task::(_decref) 
> Task=`a1230d49-dae8-4a2d-ae39-543aa1e530bb`::ref 0 aborting False
>
> ^C
>
> Command:
>
> [root at noc-2-m77 tmp]# rpm -qa | grep vdsm
>
> vdsm-xmlrpc-4.10.0-0.184.7.el6.noarch
>
> vdsm-python-4.10.0-0.184.7.el6.x86_64
>
> vdsm-4.10.0-0.184.7.el6.x86_64
>
> vdsm-cli-4.10.0-0.184.7.el6.noarch
>
> vdsm-gluster-4.10.0-0.184.7.el6.noarch
>
> [root at noc-2-m77 tmp]# rpm -qa | grep lvm
>
> lvm2-2.02.95-10.el6.x86_64
>
> lvm2-libs-2.02.95-10.el6.x86_64
>
> [root at noc-2-m77 tmp]# rpm -qa | grep gluster
>
> glusterfs-fuse-3.3.0-1.el6.x86_64
>
> glusterfs-3.3.0-1.el6.x86_64
>
> glusterfs-server-3.3.0-1.el6.x86_64
>
> glusterfs-rdma-3.3.0-1.el6.x86_64
>
> vdsm-gluster-4.10.0-0.184.7.el6.noarch
>
> glusterfs-geo-replication-3.3.0-1.el6.x86_64
>
> [root at noc-2-m77 tmp]#
>
> [root at noc-2-m77 tmp]# gluster volume info all
>
> Volume Name: sd3
>
> Type: Stripe
>
> Volume ID: bb966ba1-9b35-436a-bbc7-0503be5a1114
>
> Status: Started
>
> Number of Bricks: 1 x 2 = 2
>
> Transport-type: tcp
>
> Bricks:
>
> Brick1: 10.1.20.10:/mht
>
> Brick2: 10.2.20.2:/mht
>
> [root at noc-2-m77 tmp]#
>
> [root at noc-2-m77 tmp]# /usr/bin/sudo -n /bin/mount -t glusterfs 
> 127.0.0.1:/sd3 /tmp/test2
>
> [root at noc-2-m77 tmp]# cd /tmp/test2
>
> [root at noc-2-m77 test2]# ls
>
> [root at noc-2-m77 test2]# mkdir 12
>
> [root at noc-2-m77 test2]# ls
>
> 12
>
> [root at noc-2-m77 test2]# rm 21
>
> rm: cannot remove `21': No such file or directory
>
> [root at noc-2-m77 test2]# rm 12
>
> rm: cannot remove `12': Is a directory
>
> [root at noc-2-m77 test2]# ls
>
> 12
>
> [root at noc-2-m77 test2]# rm -r 12
>
> rm: remove directory `12'? y
>
> [root at noc-2-m77 test2]# ls
>
> [root at noc-2-m77 test2]#
>
> [root at noc-2-m77 test2]# df -h
>
> Filesystem Size  Used Avail Use% Mounted on
>
> /dev/mapper/vg_noc2m77-lv_root
>
> 50G  3.9G   44G   9% /
>
> tmpfs 7.9G     0  7.9G   0% /dev/shm
>
> /dev/sda1 497M  243M  230M  52% /boot
>
> /dev/mapper/vg_noc2m77-lv_home
>
> 491G  7.4G  459G   2% /mht
>
> 127.0.0.1:/sd3 979G   15G  915G   2% /tmp/test2
>
> [root at noc-2-m77 test2]#
>
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users


-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20120718/2ae35078/attachment-0001.html>


More information about the Users mailing list