[Users] allinone setup can't add storage

Cristian Falcas cristi.falcas at gmail.com
Thu Nov 8 06:37:56 UTC 2012


This is what I got:

Error in GUI: Cannot add Storage. Internal error, Storage Connection
doesn't exist.

Permissions:
ls -la /media/ceva2/Ovirt/Storage/
total 8
drwxrwxr-x. 2 vdsm kvm 4096 Nov  6 09:04 .
drwxr-xr-x. 5 vdsm kvm 4096 Nov  2 10:55 ..

df | grep /media/ceva2
/dev/mapper/1ATA_WDC_WD2500BB-00GUA0_WD-WCAL73625324p2 144237688 124962468
11948376  92% /media/ceva2


This is with the latest nightly on ovirt and vdsm.

This is the error from vdsm:


Thread-366::DEBUG::2012-11-08
08:31:41,244::BindingXMLRPC::161::vds::(wrapper) [79.112.94.67]
Thread-366::DEBUG::2012-11-08
08:31:41,244::task::568::TaskManager.Task::(_updateState)
Task=`2b324a0a-96ef-4ecd-8ad3-b8588ed93fd5`::moving from state init ->
state preparing
Thread-366::INFO::2012-11-08
08:31:41,245::logUtils::37::dispatcher::(wrapper) Run and protect:
validateStorageServerConnection(domType=4,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'connection':
'/media/ceva2/Ovirt/Storage', 'iqn': '', 'portal': '', 'user': '',
'password': '******', 'id': '00000000-0000-0000-0000-000000000000', 'port':
''}], options=None)
Thread-366::INFO::2012-11-08
08:31:41,245::logUtils::39::dispatcher::(wrapper) Run and protect:
validateStorageServerConnection, Return response: {'statuslist':
[{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-366::DEBUG::2012-11-08
08:31:41,245::task::1151::TaskManager.Task::(prepare)
Task=`2b324a0a-96ef-4ecd-8ad3-b8588ed93fd5`::finished: {'statuslist':
[{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-366::DEBUG::2012-11-08
08:31:41,245::task::568::TaskManager.Task::(_updateState)
Task=`2b324a0a-96ef-4ecd-8ad3-b8588ed93fd5`::moving from state preparing ->
state finished
Thread-366::DEBUG::2012-11-08
08:31:41,245::resourceManager::809::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-366::DEBUG::2012-11-08
08:31:41,245::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-366::DEBUG::2012-11-08
08:31:41,246::task::957::TaskManager.Task::(_decref)
Task=`2b324a0a-96ef-4ecd-8ad3-b8588ed93fd5`::ref 0 aborting False
Thread-367::DEBUG::2012-11-08
08:31:41,304::BindingXMLRPC::161::vds::(wrapper) [79.112.94.67]
Thread-367::DEBUG::2012-11-08
08:31:41,305::task::568::TaskManager.Task::(_updateState)
Task=`2844a5a4-7148-4d44-858c-fc75abab1a5f`::moving from state init ->
state preparing
Thread-367::INFO::2012-11-08
08:31:41,305::logUtils::37::dispatcher::(wrapper) Run and protect:
connectStorageServer(domType=4,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'connection':
'/media/ceva2/Ovirt/Storage', 'iqn': '', 'portal': '', 'user': '',
'password': '******', 'id': '00000000-0000-0000-0000-000000000000', 'port':
''}], options=None)
Thread-367::ERROR::2012-11-08
08:31:41,430::hsm::2057::Storage.HSM::(connectStorageServer) Could not
connect to storageServer
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/hsm.py", line 2054, in connectStorageServer
    conObj.connect()
  File "/usr/share/vdsm/storage/storageServer.py", line 462, in connect
    if not self.checkTarget():
  File "/usr/share/vdsm/storage/storageServer.py", line 449, in checkTarget
    fileSD.validateDirAccess(self._path))
  File "/usr/share/vdsm/storage/fileSD.py", line 51, in validateDirAccess
    getProcPool().fileUtils.validateAccess(dirPath)
  File "/usr/share/vdsm/storage/remoteFileHandler.py", line 277, in
callCrabRPCFunction
    *args, **kwargs)
  File "/usr/share/vdsm/storage/remoteFileHandler.py", line 180, in
callCrabRPCFunction
    rawLength = self._recvAll(LENGTH_STRUCT_LENGTH, timeout)
  File "/usr/share/vdsm/storage/remoteFileHandler.py", line 149, in _recvAll
    timeLeft):
  File "/usr/lib64/python2.7/contextlib.py", line 84, in helper
    return GeneratorContextManager(func(*args, **kwds))
  File "/usr/share/vdsm/storage/remoteFileHandler.py", line 136, in _poll
    raise Timeout()
Timeout
Thread-367::INFO::2012-11-08
08:31:41,432::logUtils::39::dispatcher::(wrapper) Run and protect:
connectStorageServer, Return response: {'statuslist': [{'status': 100,
'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-367::DEBUG::2012-11-08
08:31:41,433::task::1151::TaskManager.Task::(prepare)
Task=`2844a5a4-7148-4d44-858c-fc75abab1a5f`::finished: {'statuslist':
[{'status': 100, 'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-367::DEBUG::2012-11-08
08:31:41,433::task::568::TaskManager.Task::(_updateState)
Task=`2844a5a4-7148-4d44-858c-fc75abab1a5f`::moving from state preparing ->
state finished
Thread-367::DEBUG::2012-11-08
08:31:41,434::resourceManager::809::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-367::DEBUG::2012-11-08
08:31:41,434::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-367::DEBUG::2012-11-08
08:31:41,435::task::957::TaskManager.Task::(_decref)
Task=`2844a5a4-7148-4d44-858c-fc75abab1a5f`::ref 0 aborting False
Thread-370::DEBUG::2012-11-08
08:31:41,784::BindingXMLRPC::161::vds::(wrapper) [79.112.94.67]
Thread-370::DEBUG::2012-11-08
08:31:41,784::task::568::TaskManager.Task::(_updateState)
Task=`aa8526b6-0772-4909-bad3-b2cd7ad589cc`::moving from state init ->
state preparing
Thread-370::INFO::2012-11-08
08:31:41,784::logUtils::37::dispatcher::(wrapper) Run and protect:
disconnectStorageServer(domType=4,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'connection':
'/media/ceva2/Ovirt/Storage', 'iqn': '', 'portal': '', 'user': '',
'password': '******', 'id': '00000000-0000-0000-0000-000000000000', 'port':
''}], options=None)
Thread-370::DEBUG::2012-11-08
08:31:41,785::misc::1026::SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
Thread-370::DEBUG::2012-11-08
08:31:41,785::misc::1028::SamplingMethod::(__call__) Got in to sampling
method
Thread-370::DEBUG::2012-11-08
08:31:41,785::misc::1026::SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
Thread-370::DEBUG::2012-11-08
08:31:41,785::misc::1028::SamplingMethod::(__call__) Got in to sampling
method
Thread-370::DEBUG::2012-11-08
08:31:41,785::misc::84::Storage.Misc.excCmd::(<lambda>) '/bin/sudo -n
/sbin/iscsiadm -m session -R' (cwd None)
Thread-370::DEBUG::2012-11-08
08:31:41,810::misc::84::Storage.Misc.excCmd::(<lambda>) FAILED: <err> =
'iscsiadm: No session found.\n'; <rc> = 21
Thread-370::DEBUG::2012-11-08
08:31:41,811::misc::1036::SamplingMethod::(__call__) Returning last result
MainProcess|Thread-370::DEBUG::2012-11-08
08:31:41,814::misc::84::Storage.Misc.excCmd::(<lambda>) '/bin/dd
of=/sys/class/scsi_host/host0/scan' (cwd None)
MainProcess|Thread-370::DEBUG::2012-11-08
08:31:41,817::misc::84::Storage.Misc.excCmd::(<lambda>) '/bin/dd
of=/sys/class/scsi_host/host1/scan' (cwd None)
MainProcess|Thread-370::DEBUG::2012-11-08
08:31:41,820::misc::84::Storage.Misc.excCmd::(<lambda>) '/bin/dd
of=/sys/class/scsi_host/host2/scan' (cwd None)
MainProcess|Thread-370::DEBUG::2012-11-08
08:31:41,822::misc::84::Storage.Misc.excCmd::(<lambda>) '/bin/dd
of=/sys/class/scsi_host/host3/scan' (cwd None)
MainProcess|Thread-370::DEBUG::2012-11-08
08:31:42,827::misc::84::Storage.Misc.excCmd::(<lambda>) '/bin/dd
of=/sys/class/scsi_host/host4/scan' (cwd None)
MainProcess|Thread-370::DEBUG::2012-11-08
08:31:42,835::misc::84::Storage.Misc.excCmd::(<lambda>) '/bin/dd
of=/sys/class/scsi_host/host5/scan' (cwd None)
MainProcess|Thread-370::DEBUG::2012-11-08
08:31:42,842::iscsi::388::Storage.ISCSI::(forceIScsiScan) Performing SCSI
scan, this will take up to 30 seconds
Thread-370::DEBUG::2012-11-08
08:31:44,846::misc::84::Storage.Misc.excCmd::(<lambda>) '/bin/sudo -n
/sbin/multipath' (cwd None)
Thread-370::DEBUG::2012-11-08
08:31:44,902::misc::84::Storage.Misc.excCmd::(<lambda>) SUCCESS: <err> =
''; <rc> = 0
Thread-370::DEBUG::2012-11-08
08:31:44,903::lvm::477::OperationMutex::(_invalidateAllPvs) Operation 'lvm
invalidate operation' got the operation mutex
Thread-370::DEBUG::2012-11-08
08:31:44,904::lvm::479::OperationMutex::(_invalidateAllPvs) Operation 'lvm
invalidate operation' released the operation mutex
Thread-370::DEBUG::2012-11-08
08:31:44,905::lvm::488::OperationMutex::(_invalidateAllVgs) Operation 'lvm
invalidate operation' got the operation mutex
Thread-370::DEBUG::2012-11-08
08:31:44,905::lvm::490::OperationMutex::(_invalidateAllVgs) Operation 'lvm
invalidate operation' released the operation mutex
Thread-370::DEBUG::2012-11-08
08:31:44,906::lvm::508::OperationMutex::(_invalidateAllLvs) Operation 'lvm
invalidate operation' got the operation mutex
Thread-370::DEBUG::2012-11-08
08:31:44,906::lvm::510::OperationMutex::(_invalidateAllLvs) Operation 'lvm
invalidate operation' released the operation mutex
Thread-370::DEBUG::2012-11-08
08:31:44,907::misc::1036::SamplingMethod::(__call__) Returning last result
Thread-370::INFO::2012-11-08
08:31:44,907::logUtils::39::dispatcher::(wrapper) Run and protect:
disconnectStorageServer, Return response: {'statuslist': [{'status': 0,
'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-370::DEBUG::2012-11-08
08:31:44,908::task::1151::TaskManager.Task::(prepare)
Task=`aa8526b6-0772-4909-bad3-b2cd7ad589cc`::finished: {'statuslist':
[{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]}
Thread-370::DEBUG::2012-11-08
08:31:44,908::task::568::TaskManager.Task::(_updateState)
Task=`aa8526b6-0772-4909-bad3-b2cd7ad589cc`::moving from state preparing ->
state finished
Thread-370::DEBUG::2012-11-08
08:31:44,909::resourceManager::809::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-370::DEBUG::2012-11-08
08:31:44,909::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-370::DEBUG::2012-11-08
08:31:44,910::task::957::TaskManager.Task::(_decref)
Task=`aa8526b6-0772-4909-bad3-b2cd7ad589cc`::ref 0 aborting False



On Wed, Nov 7, 2012 at 10:00 PM, Itamar Heim <iheim at redhat.com> wrote:

> On 11/07/2012 02:59 PM, Cristian Falcas wrote:
>
>> Bummer.
>>
>> And the sdk from nightly is not working because it's too new, from what
>> I sow on the mailing list?
>>
>> I have errors even when I'm trying to add the storage manually, from the
>> engine.Should I give up on nightly and use the beta ones, or there is an
>> other solution for this?
>>
>
> shouldn't fail from admin. which vdsm version, logs?
>
>
>>
>> On Wed, Nov 7, 2012 at 3:47 PM, Moran Goldboim <mgoldboi at redhat.com
>> <mailto:mgoldboi at redhat.com>> wrote:
>>
>>     you are using ovirt-sdk from stable repo and engine from nightly
>>     repo, those doesn't work together.
>>
>>
>>
>>     On 11/07/2012 01:40 PM, Cristian Falcas wrote:
>>
>>         Hi all,
>>
>>         Can someone help me with this error:
>>
>>         AIO: Adding Local Datacenter and cluster...
>>         [ ERROR ]
>>         Error: could not create ovirtsdk API object
>>
>>
>>         trace from the log file
>>
>>         2012-11-07 13:34:44::DEBUG::all_in_one___**100::220::root::
>>         Initiating the API object
>>         2012-11-07 13:34:44::ERROR::all_in_one___**100::231::root::
>>
>>         Traceback (most recent call last):
>>            File
>>         "/usr/share/ovirt-engine/__**scripts/plugins/all_in_one___**
>> 100.py",
>>         line 228, in initAPI
>>              ca_file=basedefs.FILE_CA_CRT__**_SRC,
>>
>>         TypeError: __init__() got an unexpected keyword argument 'ca_file'
>>
>>         2012-11-07 13:34:44::DEBUG::setup___**sequences::62::root::
>>
>>         Traceback (most recent call last):
>>            File "/usr/share/ovirt-engine/__**scripts/setup_sequences.py",
>>
>>         line 60, in run
>>              function()
>>            File
>>         "/usr/share/ovirt-engine/__**scripts/plugins/all_in_one___**
>> 100.py",
>>         line 232, in initAPI
>>              raise Exception(ERROR_CREATE_API___**OBJECT)
>>
>>         Exception: Error: could not create ovirtsdk API object
>>
>>
>>         Versions installed:
>>
>>         ovirt-engine-3.1.0-3.20121106.**__git6891171.fc17.noarch
>>         ovirt-engine-backend-3.1.0-3._**_20121106.git6891171.fc17.__**
>> noarch
>>         ovirt-engine-cli-3.1.0.6-1.__**fc17.noarch
>>         ovirt-engine-config-3.1.0-3.__**20121106.git6891171.fc17.__**
>> noarch
>>         ovirt-engine-dbscripts-3.1.0-_**_3.20121106.git6891171.fc17.__**
>> noarch
>>         ovirt-engine-genericapi-3.1.0-**__3.20121106.git6891171.fc17._**
>> _noarch
>>         ovirt-engine-notification-__**service-3.1.0-3.20121106.__**
>> git6891171.fc17.noarch
>>         ovirt-engine-restapi-3.1.0-3._**_20121106.git6891171.fc17.__**
>> noarch
>>         ovirt-engine-sdk-3.1.0.4-1.__**fc17.noarch
>>         ovirt-engine-setup-3.1.0-3.__**20121106.git6891171.fc17.__**
>> noarch
>>         ovirt-engine-setup-plugin-__**allinone-3.1.0-3.20121106.__**
>> git6891171.fc17.noarch
>>         ovirt-engine-tools-common-3.1.**__0-3.20121106.git6891171.**
>> fc17.__noarch
>>         ovirt-engine-userportal-3.1.0-**__3.20121106.git6891171.fc17._**
>> _noarch
>>         ovirt-engine-webadmin-portal-_**_3.1.0-3.20121106.git6891171._**
>> _fc17.noarch
>>
>>
>>         ______________________________**___________________
>>         Users mailing list
>>         Users at ovirt.org <mailto:Users at ovirt.org>
>>         http://lists.ovirt.org/__**mailman/listinfo/users<http://lists.ovirt.org/__mailman/listinfo/users>
>>         <http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>> >
>>
>>
>>
>>
>>
>>
>> ______________________________**_________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20121108/8ff5ccf2/attachment-0001.html>


More information about the Users mailing list