[ovirt-users] Error creating a storage domain

Bloemen, Jurriën Jurrien.Bloemen at dmc.amcnetworks.com
Tue Jul 21 07:57:10 UTC 2015


Hi all,

I have to add that this is the latest 3.5 version. Adding the other glusterfs storage was done by an older version of 3.5. I don't know if something has changed in between version?

Thanks,

Jurriën

ovirt-engine-websocket-proxy-3.5.3.1-1.el7.centos.noarch
ovirt-release35-004-1.noarch
ovirt-engine-backend-3.5.3.1-1.el7.centos.noarch
ovirt-host-deploy-1.3.1-1.el7.noarch
ovirt-engine-userportal-3.5.3.1-1.el7.centos.noarch
ovirt-engine-jboss-as-7.1.1-1.el7.x86_64
ovirt-engine-lib-3.5.3.1-1.el7.centos.noarch
ovirt-engine-setup-3.5.3.1-1.el7.centos.noarch
ovirt-engine-webadmin-portal-3.5.3.1-1.el7.centos.noarch
ovirt-engine-dbscripts-3.5.3.1-1.el7.centos.noarch
ovirt-engine-3.5.3.1-1.el7.centos.noarch
ovirt-engine-extensions-api-impl-3.5.3.1-1.el7.centos.noarch
ovirt-image-uploader-3.5.1-1.el7.centos.noarch
ovirt-engine-cli-3.5.0.5-1.el7.centos.noarch
ovirt-host-deploy-java-1.3.1-1.el7.noarch
ovirt-iso-uploader-3.5.2-1.el7.centos.noarch
ovirt-engine-setup-base-3.5.3.1-1.el7.centos.noarch
ovirt-engine-setup-plugin-ovirt-engine-3.5.3.1-1.el7.centos.noarch
ovirt-engine-setup-plugin-websocket-proxy-3.5.3.1-1.el7.centos.noarch
ovirt-engine-restapi-3.5.3.1-1.el7.centos.noarch
ovirt-engine-tools-3.5.3.1-1.el7.centos.noarch
ovirt-engine-sdk-python-3.5.2.1-1.el7.centos.noarch
ovirt-engine-setup-plugin-ovirt-engine-common-3.5.3.1-1.el7.centos.noarch



On 20-07-15 18:30, Bloemen, Jurriën wrote:

Hi all,


I have some trouble with adding a gluster storage domain:


Error while executing action AddGlusterFsStorageDomain: Error creating a storage domain


I have mounted the gluster volume by hand and that works fine. Also checked the rights and ownership and they are 755 and vdsm:kvm.

I also tried to reinstall the oVirtH node.


May worth of telling is that I already have one gluster mount running perfectly but add a new one does not work.


Can somebody help me?


This is the engine.log of the oVirt manager:


2015-07-20 18:23:30,766 INFO  [org.ovirt.engine.core.bll.storage.AddStorageServerConnectionCommand] (ajp--127.0.0.1-8702-3) [61712435] Lock Acquired to object EngineLock [exclusiveLocks= key: superstore001-stor.cs.example.com:/ovirtprd01 value: STORAGE_CONNECTION

, sharedLocks= ]

2015-07-20 18:23:30,791 INFO  [org.ovirt.engine.core.bll.storage.AddStorageServerConnectionCommand] (ajp--127.0.0.1-8702-3) [61712435] Running command: AddStorageServerConnectionCommand internal: false. Entities affected :  ID: aaa00000-0000-0000-0000-123456789aaa Type: SystemAction group CREATE_STORAGE_DOMAIN with role type ADMIN

2015-07-20 18:23:30,811 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVDSCommand] (ajp--127.0.0.1-8702-3) [61712435] START, ConnectStorageServerVDSCommand(HostName = twin189, HostId = ad392b6d-12b8-4f4d-98a1-49e87443eddc, storagePoolId = 00000000-0000-0000-0000-000000000000, storageType = GLUSTERFS, connectionList = [{ id: null, connection: superstore001-stor.cs.example.com:/ovirtprd01, iqn: null, vfsType: glusterfs, mountOptions: null, nfsVersion: null, nfsRetrans: null, nfsTimeo: null };]), log id: 44c1a9f2

2015-07-20 18:23:31,074 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVDSCommand] (ajp--127.0.0.1-8702-3) [61712435] FINISH, ConnectStorageServerVDSCommand, return: {00000000-0000-0000-0000-000000000000=0}, log id: 44c1a9f2

2015-07-20 18:23:31,085 INFO  [org.ovirt.engine.core.bll.storage.AddStorageServerConnectionCommand] (ajp--127.0.0.1-8702-3) [61712435] Lock freed to object EngineLock [exclusiveLocks= key: superstore001-stor.cs.example.com:/ovirtprd01 value: STORAGE_CONNECTION

, sharedLocks= ]

2015-07-20 18:23:31,138 WARN  [org.ovirt.engine.core.dal.job.ExecutionMessageDirector] (ajp--127.0.0.1-8702-3) [8d11fec] The message key AddGlusterFsStorageDomain is missing from bundles/ExecutionMessages

2015-07-20 18:23:31,156 INFO  [org.ovirt.engine.core.bll.storage.AddGlusterFsStorageDomainCommand] (ajp--127.0.0.1-8702-3) [8d11fec] Running command: AddGlusterFsStorageDomainCommand internal: false. Entities affected :  ID: aaa00000-0000-0000-0000-123456789aaa Type: SystemAction group CREATE_STORAGE_DOMAIN with role type ADMIN

2015-07-20 18:23:31,183 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVDSCommand] (ajp--127.0.0.1-8702-3) [8d11fec] START, ConnectStorageServerVDSCommand(HostName = twin189, HostId = ad392b6d-12b8-4f4d-98a1-49e87443eddc, storagePoolId = 00000000-0000-0000-0000-000000000000, storageType = GLUSTERFS, connectionList = [{ id: ece6ad42-be90-4980-98dd-d7ae12cc6709, connection: superstore001-stor.cs.example.com:/ovirtprd01, iqn: null, vfsType: glusterfs, mountOptions: null, nfsVersion: null, nfsRetrans: null, nfsTimeo: null };]), log id: 3bb1d8b0

2015-07-20 18:23:31,207 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVDSCommand] (ajp--127.0.0.1-8702-3) [8d11fec] FINISH, ConnectStorageServerVDSCommand, return: {ece6ad42-be90-4980-98dd-d7ae12cc6709=0}, log id: 3bb1d8b0

2015-07-20 18:23:31,223 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStorageDomainVDSCommand] (ajp--127.0.0.1-8702-3) [8d11fec] START, CreateStorageDomainVDSCommand(HostName = twin189, HostId = ad392b6d-12b8-4f4d-98a1-49e87443eddc, storageDomain=StorageDomainStatic[ovirtprd01, d6df7930-342a-493a-b70b-fb1c52b0828c], args=superstore001-stor.cs.example.com:/ovirtprd01), log id: 42df6467

2015-07-20 18:23:31,356 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStorageDomainVDSCommand] (ajp--127.0.0.1-8702-3) [8d11fec] Failed in CreateStorageDomainVDS method

2015-07-20 18:23:31,358 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStorageDomainVDSCommand] (ajp--127.0.0.1-8702-3) [8d11fec] Command org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStorageDomainVDSCommand return value

 StatusOnlyReturnForXmlRpc [mStatus=StatusForXmlRpc [mCode=351, mMessage=Error creating a storage domain: (u'storageType=7, sdUUID=d6df7930-342a-493a-b70b-fb1c52b0828c, domainName=ovirtprd01, domClass=1, typeSpecificArg=superstore001-stor.cs.example.com:/ovirtprd01 domVersion=3',)]]

2015-07-20 18:23:31,364 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStorageDomainVDSCommand] (ajp--127.0.0.1-8702-3) [8d11fec] HostName = twin189

2015-07-20 18:23:31,366 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStorageDomainVDSCommand] (ajp--127.0.0.1-8702-3) [8d11fec] Command CreateStorageDomainVDSCommand(HostName = twin189, HostId = ad392b6d-12b8-4f4d-98a1-49e87443eddc, storageDomain=StorageDomainStatic[ovirtprd01, d6df7930-342a-493a-b70b-fb1c52b0828c], args=superstore001-stor.cs.example.com:/ovirtprd01) execution failed. Exception: VDSErrorException: VDSGenericException: VDSErrorException: Failed to CreateStorageDomainVDS, error = Error creating a storage domain: (u'storageType=7, sdUUID=d6df7930-342a-493a-b70b-fb1c52b0828c, domainName=ovirtprd01, domClass=1, typeSpecificArg=superstore001-stor.cs.example.com:/ovirtprd01 domVersion=3',), code = 351

2015-07-20 18:23:31,375 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStorageDomainVDSCommand] (ajp--127.0.0.1-8702-3) [8d11fec] FINISH, CreateStorageDomainVDSCommand, log id: 42df6467

2015-07-20 18:23:31,377 ERROR [org.ovirt.engine.core.bll.storage.AddGlusterFsStorageDomainCommand] (ajp--127.0.0.1-8702-3) [8d11fec] Command org.ovirt.engine.core.bll.storage.AddGlusterFsStorageDomainCommand throw Vdc Bll exception. With error message VdcBLLException: org.ovirt.engine.core.vdsbroker.vdsbroker.VDSErrorException: VDSGenericException: VDSErrorException: Failed to CreateStorageDomainVDS, error = Error creating a storage domain: (u'storageType=7, sdUUID=d6df7930-342a-493a-b70b-fb1c52b0828c, domainName=ovirtprd01, domClass=1, typeSpecificArg=superstore001-stor.cs.example.com:/ovirtprd01 domVersion=3',), code = 351 (Failed with error StorageDomainCreationError and code 351)

2015-07-20 18:23:31,386 INFO  [org.ovirt.engine.core.bll.storage.AddGlusterFsStorageDomainCommand] (ajp--127.0.0.1-8702-3) [8d11fec] Command [id=b1e0d0ec-526b-42e4-997f-ff55614e0797]: Compensating NEW_ENTITY_ID of org.ovirt.engine.core.common.businessentities.StorageDomainDynamic; snapshot: d6df7930-342a-493a-b70b-fb1c52b0828c.

2015-07-20 18:23:31,390 INFO  [org.ovirt.engine.core.bll.storage.AddGlusterFsStorageDomainCommand] (ajp--127.0.0.1-8702-3) [8d11fec] Command [id=b1e0d0ec-526b-42e4-997f-ff55614e0797]: Compensating NEW_ENTITY_ID of org.ovirt.engine.core.common.businessentities.StorageDomainStatic; snapshot: d6df7930-342a-493a-b70b-fb1c52b0828c.

2015-07-20 18:23:31,397 ERROR [org.ovirt.engine.core.bll.storage.AddGlusterFsStorageDomainCommand] (ajp--127.0.0.1-8702-3) [8d11fec] Transaction rolled-back for command: org.ovirt.engine.core.bll.storage.AddGlusterFsStorageDomainCommand.

2015-07-20 18:23:31,404 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-3) [8d11fec] Correlation ID: 8d11fec, Job ID: 9c59789a-ab38-4171-96c5-3d8da688e2bb, Call Stack: null, Custom Event ID: -1, Message: Failed to add Storage Domain ovirtprd01. (User: admin at internal)

2015-07-20 18:23:31,451 INFO  [org.ovirt.engine.core.bll.storage.RemoveStorageServerConnectionCommand] (ajp--127.0.0.1-8702-3) [406a3269] Lock Acquired to object EngineLock [exclusiveLocks= key: ece6ad42-be90-4980-98dd-d7ae12cc6709 value: STORAGE_CONNECTION

key: superstore001-stor.cs.example.com:/ovirtprd01 value: STORAGE_CONNECTION

, sharedLocks= ]

2015-07-20 18:23:31,607 INFO  [org.ovirt.engine.core.bll.storage.RemoveStorageServerConnectionCommand] (ajp--127.0.0.1-8702-3) [406a3269] Running command: RemoveStorageServerConnectionCommand internal: false. Entities affected :  ID: aaa00000-0000-0000-0000-123456789aaa Type: SystemAction group CREATE_STORAGE_DOMAIN with role type ADMIN

2015-07-20 18:23:31,615 INFO  [org.ovirt.engine.core.bll.storage.RemoveStorageServerConnectionCommand] (ajp--127.0.0.1-8702-3) [406a3269] Removing connection ece6ad42-be90-4980-98dd-d7ae12cc6709 from database

2015-07-20 18:23:31,631 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.DisconnectStorageServerVDSCommand] (ajp--127.0.0.1-8702-3) [406a3269] START, DisconnectStorageServerVDSCommand(HostName = twin189, HostId = ad392b6d-12b8-4f4d-98a1-49e87443eddc, storagePoolId = 00000000-0000-0000-0000-000000000000, storageType = GLUSTERFS, connectionList = [{ id: ece6ad42-be90-4980-98dd-d7ae12cc6709, connection: superstore001-stor.cs.example.com:/ovirtprd01, iqn: null, vfsType: glusterfs, mountOptions: null, nfsVersion: null, nfsRetrans: null, nfsTimeo: null };]), log id: 372ee09d

2015-07-20 18:23:31,723 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.DisconnectStorageServerVDSCommand] (ajp--127.0.0.1-8702-3) [406a3269] FINISH, DisconnectStorageServerVDSCommand, return: {ece6ad42-be90-4980-98dd-d7ae12cc6709=0}, log id: 372ee09d

2015-07-20 18:23:31,728 INFO  [org.ovirt.engine.core.bll.storage.RemoveStorageServerConnectionCommand] (ajp--127.0.0.1-8702-3) [406a3269] Lock freed to object EngineLock [exclusiveLocks= key: ece6ad42-be90-4980-98dd-d7ae12cc6709 value: STORAGE_CONNECTION

key: superstore001-stor.cs.example.com:/ovirtprd01 value: STORAGE_CONNECTION

, sharedLocks= ]


And this is de vdsm.log of the hypervisor:


JsonRpc (StompReactor)::DEBUG::2015-07-20 16:23:25,456::stompReactor::98::Broker.StompAdapter::(handle_frame) Handling message <StompFrame command='SEND'>

JsonRpcServer::DEBUG::2015-07-20 16:23:25,458::__init__::506::jsonrpc.JsonRpcServer::(serve_requests) Waiting for request

Thread-199::DEBUG::2015-07-20 16:23:25,460::stompReactor::163::yajsonrpc.StompServer::(send) Sending response

JsonRpc (StompReactor)::DEBUG::2015-07-20 16:23:28,483::stompReactor::98::Broker.StompAdapter::(handle_frame) Handling message <StompFrame command='SEND'>

JsonRpcServer::DEBUG::2015-07-20 16:23:28,484::__init__::506::jsonrpc.JsonRpcServer::(serve_requests) Waiting for request

Thread-200::DEBUG::2015-07-20 16:23:28,485::stompReactor::163::yajsonrpc.StompServer::(send) Sending response

JsonRpc (StompReactor)::DEBUG::2015-07-20 16:23:30,124::stompReactor::98::Broker.StompAdapter::(handle_frame) Handling message <StompFrame command='SEND'>

JsonRpcServer::DEBUG::2015-07-20 16:23:30,125::__init__::506::jsonrpc.JsonRpcServer::(serve_requests) Waiting for request

Thread-201::DEBUG::2015-07-20 16:23:30,125::__init__::469::jsonrpc.JsonRpcServer::(_serveRequest) Calling 'StoragePool.connectStorageServer' in bridge with {u'connectionParams': [{u'id': u'00000000-0000-0000-0000-000000000000', u'connection': u'superstore001-stor.cs.example.com:/ovirtprd01', u'iqn': u'', u'user': u'', u'tpgt': u'1', u'vfs_type': u'glusterfs', u'password': u'', u'port': u''}], u'storagepoolID': u'00000000-0000-0000-0000-000000000000', u'domainType': 7}

Thread-201::DEBUG::2015-07-20 16:23:30,127::task::595::Storage.TaskManager.Task::(_updateState) Task=`417462a8-3817-4e44-ac89-20ea4e3c8709`::moving from state init -> state preparing

Thread-201::INFO::2015-07-20 16:23:30,127::logUtils::44::dispatcher::(wrapper) Run and protect: connectStorageServer(domType=7, spUUID=u'00000000-0000-0000-0000-000000000000', conList=[{u'port': u'', u'connection': u'superstore001-stor.cs.example.com:/ovirtprd01', u'iqn': u'', u'user': u'', u'tpgt': u'1', u'vfs_type': u'glusterfs', u'password': '******', u'id': u'00000000-0000-0000-0000-000000000000'}], options=None)

Thread-201::DEBUG::2015-07-20 16:23:30,137::fileUtils::142::Storage.fileUtils::(createdir) Creating directory: /rhev/data-center/mnt/glusterSD/superstore001-stor.cs.example.com:_ovirtprd01

Thread-201::DEBUG::2015-07-20 16:23:30,138::mount::227::Storage.Misc.excCmd::(_runcmd) /usr/bin/sudo -n /usr/bin/mount -t glusterfs superstore001-stor.cs.example.com:/ovirtprd01 /rhev/data-center/mnt/glusterSD/superstore001-stor.cs.example.com:_ovirtprd01 (cwd None)

Thread-201::DEBUG::2015-07-20 16:23:30,371::hsm::2375::Storage.HSM::(__prefetchDomains) glusterDomPath: glusterSD/*

Thread-201::DEBUG::2015-07-20 16:23:30,377::hsm::2387::Storage.HSM::(__prefetchDomains) Found SD uuids: ()

Thread-201::DEBUG::2015-07-20 16:23:30,377::hsm::2443::Storage.HSM::(connectStorageServer) knownSDs: {}

Thread-201::INFO::2015-07-20 16:23:30,377::logUtils::47::dispatcher::(wrapper) Run and protect: connectStorageServer, Return response: {'statuslist': [{'status': 0, 'id': u'00000000-0000-0000-0000-000000000000'}]}

Thread-201::DEBUG::2015-07-20 16:23:30,377::task::1191::Storage.TaskManager.Task::(prepare) Task=`417462a8-3817-4e44-ac89-20ea4e3c8709`::finished: {'statuslist': [{'status': 0, 'id': u'00000000-0000-0000-0000-000000000000'}]}

Thread-201::DEBUG::2015-07-20 16:23:30,377::task::595::Storage.TaskManager.Task::(_updateState) Task=`417462a8-3817-4e44-ac89-20ea4e3c8709`::moving from state preparing -> state finished

Thread-201::DEBUG::2015-07-20 16:23:30,377::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {}

Thread-201::DEBUG::2015-07-20 16:23:30,378::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {}

Thread-201::DEBUG::2015-07-20 16:23:30,378::task::993::Storage.TaskManager.Task::(_decref) Task=`417462a8-3817-4e44-ac89-20ea4e3c8709`::ref 0 aborting False

Thread-201::DEBUG::2015-07-20 16:23:30,378::__init__::500::jsonrpc.JsonRpcServer::(_serveRequest) Return 'StoragePool.connectStorageServer' in bridge with [{'status': 0, 'id': u'00000000-0000-0000-0000-000000000000'}]

Thread-201::DEBUG::2015-07-20 16:23:30,378::stompReactor::163::yajsonrpc.StompServer::(send) Sending response

JsonRpc (StompReactor)::DEBUG::2015-07-20 16:23:30,497::stompReactor::98::Broker.StompAdapter::(handle_frame) Handling message <StompFrame command='SEND'>

JsonRpcServer::DEBUG::2015-07-20 16:23:30,497::__init__::506::jsonrpc.JsonRpcServer::(serve_requests) Waiting for request

Thread-203::DEBUG::2015-07-20 16:23:30,497::__init__::469::jsonrpc.JsonRpcServer::(_serveRequest) Calling 'StoragePool.connectStorageServer' in bridge with {u'connectionParams': [{u'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709', u'connection': u'superstore001-stor.cs.example.com:/ovirtprd01', u'iqn': u'', u'user': u'', u'tpgt': u'1', u'vfs_type': u'glusterfs', u'password': u'', u'port': u''}], u'storagepoolID': u'00000000-0000-0000-0000-000000000000', u'domainType': 7}

Thread-203::DEBUG::2015-07-20 16:23:30,498::task::595::Storage.TaskManager.Task::(_updateState) Task=`627a1260-b231-493f-b0f1-14c0a8501f49`::moving from state init -> state preparing

Thread-203::INFO::2015-07-20 16:23:30,499::logUtils::44::dispatcher::(wrapper) Run and protect: connectStorageServer(domType=7, spUUID=u'00000000-0000-0000-0000-000000000000', conList=[{u'port': u'', u'connection': u'superstore001-stor.cs.example.com:/ovirtprd01', u'iqn': u'', u'user': u'', u'tpgt': u'1', u'vfs_type': u'glusterfs', u'password': '******', u'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709'}], options=None)

Thread-203::DEBUG::2015-07-20 16:23:30,505::hsm::2375::Storage.HSM::(__prefetchDomains) glusterDomPath: glusterSD/*

Thread-203::DEBUG::2015-07-20 16:23:30,511::hsm::2387::Storage.HSM::(__prefetchDomains) Found SD uuids: ()

Thread-203::DEBUG::2015-07-20 16:23:30,511::hsm::2443::Storage.HSM::(connectStorageServer) knownSDs: {}

Thread-203::INFO::2015-07-20 16:23:30,511::logUtils::47::dispatcher::(wrapper) Run and protect: connectStorageServer, Return response: {'statuslist': [{'status': 0, 'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709'}]}

Thread-203::DEBUG::2015-07-20 16:23:30,511::task::1191::Storage.TaskManager.Task::(prepare) Task=`627a1260-b231-493f-b0f1-14c0a8501f49`::finished: {'statuslist': [{'status': 0, 'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709'}]}

Thread-203::DEBUG::2015-07-20 16:23:30,511::task::595::Storage.TaskManager.Task::(_updateState) Task=`627a1260-b231-493f-b0f1-14c0a8501f49`::moving from state preparing -> state finished

Thread-203::DEBUG::2015-07-20 16:23:30,511::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {}

Thread-203::DEBUG::2015-07-20 16:23:30,511::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {}

Thread-203::DEBUG::2015-07-20 16:23:30,511::task::993::Storage.TaskManager.Task::(_decref) Task=`627a1260-b231-493f-b0f1-14c0a8501f49`::ref 0 aborting False

Thread-203::DEBUG::2015-07-20 16:23:30,511::__init__::500::jsonrpc.JsonRpcServer::(_serveRequest) Return 'StoragePool.connectStorageServer' in bridge with [{'status': 0, 'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709'}]

Thread-203::DEBUG::2015-07-20 16:23:30,512::stompReactor::163::yajsonrpc.StompServer::(send) Sending response

JsonRpc (StompReactor)::DEBUG::2015-07-20 16:23:30,532::stompReactor::98::Broker.StompAdapter::(handle_frame) Handling message <StompFrame command='SEND'>

JsonRpcServer::DEBUG::2015-07-20 16:23:30,533::__init__::506::jsonrpc.JsonRpcServer::(serve_requests) Waiting for request

Thread-205::DEBUG::2015-07-20 16:23:30,533::__init__::469::jsonrpc.JsonRpcServer::(_serveRequest) Calling 'StorageDomain.create' in bridge with {u'name': u'ovirtprd01', u'domainType': 7, u'domainClass': 1, u'typeArgs': u'superstore001-stor.cs.example.com:/ovirtprd01', u'version': u'3', u'storagedomainID': u'd6df7930-342a-493a-b70b-fb1c52b0828c'}

Thread-205::DEBUG::2015-07-20 16:23:30,534::task::595::Storage.TaskManager.Task::(_updateState) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::moving from state init -> state preparing

Thread-205::INFO::2015-07-20 16:23:30,534::logUtils::44::dispatcher::(wrapper) Run and protect: createStorageDomain(storageType=7, sdUUID=u'd6df7930-342a-493a-b70b-fb1c52b0828c', domainName=u'ovirtprd01', typeSpecificArg=u'superstore001-stor.cs.example.com:/ovirtprd01', domClass=1, domVersion=u'3', options=None)

Thread-205::DEBUG::2015-07-20 16:23:30,534::misc::741::Storage.SamplingMethod::(__call__) Trying to enter sampling method (storage.sdc.refreshStorage)

Thread-205::DEBUG::2015-07-20 16:23:30,534::misc::743::Storage.SamplingMethod::(__call__) Got in to sampling method

Thread-205::DEBUG::2015-07-20 16:23:30,534::misc::741::Storage.SamplingMethod::(__call__) Trying to enter sampling method (storage.iscsi.rescan)

Thread-205::DEBUG::2015-07-20 16:23:30,535::misc::743::Storage.SamplingMethod::(__call__) Got in to sampling method

Thread-205::DEBUG::2015-07-20 16:23:30,535::iscsi::403::Storage.ISCSI::(rescan) Performing SCSI scan, this will take up to 30 seconds

Thread-205::DEBUG::2015-07-20 16:23:30,535::iscsiadm::92::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n /sbin/iscsiadm -m session -R (cwd None)

Thread-205::DEBUG::2015-07-20 16:23:30,545::misc::751::Storage.SamplingMethod::(__call__) Returning last result

Thread-205::DEBUG::2015-07-20 16:23:30,545::misc::741::Storage.SamplingMethod::(__call__) Trying to enter sampling method (storage.hba.rescan)

Thread-205::DEBUG::2015-07-20 16:23:30,545::misc::743::Storage.SamplingMethod::(__call__) Got in to sampling method

Thread-205::DEBUG::2015-07-20 16:23:30,545::hba::53::Storage.HBA::(rescan) Starting scan

Thread-205::DEBUG::2015-07-20 16:23:30,545::utils::739::Storage.HBA::(execCmd) /usr/bin/sudo -n /usr/libexec/vdsm/fc-scan (cwd None)

Thread-205::DEBUG::2015-07-20 16:23:30,565::hba::66::Storage.HBA::(rescan) Scan finished

Thread-205::DEBUG::2015-07-20 16:23:30,565::misc::751::Storage.SamplingMethod::(__call__) Returning last result

Thread-205::DEBUG::2015-07-20 16:23:30,565::multipath::128::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n /sbin/multipath (cwd None)

Thread-205::DEBUG::2015-07-20 16:23:30,606::multipath::128::Storage.Misc.excCmd::(rescan) SUCCESS: <err> = ''; <rc> = 0

Thread-205::DEBUG::2015-07-20 16:23:30,607::lvm::489::Storage.OperationMutex::(_invalidateAllPvs) Operation 'lvm invalidate operation' got the operation mutex

Thread-205::DEBUG::2015-07-20 16:23:30,607::lvm::491::Storage.OperationMutex::(_invalidateAllPvs) Operation 'lvm invalidate operation' released the operation mutex

Thread-205::DEBUG::2015-07-20 16:23:30,607::lvm::500::Storage.OperationMutex::(_invalidateAllVgs) Operation 'lvm invalidate operation' got the operation mutex

Thread-205::DEBUG::2015-07-20 16:23:30,607::lvm::502::Storage.OperationMutex::(_invalidateAllVgs) Operation 'lvm invalidate operation' released the operation mutex

Thread-205::DEBUG::2015-07-20 16:23:30,607::lvm::520::Storage.OperationMutex::(_invalidateAllLvs) Operation 'lvm invalidate operation' got the operation mutex

Thread-205::DEBUG::2015-07-20 16:23:30,607::lvm::522::Storage.OperationMutex::(_invalidateAllLvs) Operation 'lvm invalidate operation' released the operation mutex

Thread-205::DEBUG::2015-07-20 16:23:30,608::misc::751::Storage.SamplingMethod::(__call__) Returning last result

Thread-205::ERROR::2015-07-20 16:23:30,608::sdc::137::Storage.StorageDomainCache::(_findDomain) looking for unfetched domain d6df7930-342a-493a-b70b-fb1c52b0828c

Thread-205::ERROR::2015-07-20 16:23:30,608::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDomain) looking for domain d6df7930-342a-493a-b70b-fb1c52b0828c

Thread-205::DEBUG::2015-07-20 16:23:30,608::lvm::365::Storage.OperationMutex::(_reloadvgs) Operation 'lvm reload operation' got the operation mutex

Thread-205::DEBUG::2015-07-20 16:23:30,609::lvm::288::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n /sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 obtain_device_list_from_udev=0 filter = [ '\''a|/dev/mapper/3600304801a8505001cd001f904750aa2|'\'', '\''r|.*|'\'' ] }  global {  locking_type=1  prioritise_write_locks=1  wait_for_locks=1  use_lvmetad=0 }  backup {  retain_min = 50  retain_days = 0 } ' --noheadings --units b --nosuffix --separator '|' --ignoreskippedcluster -o uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name d6df7930-342a-493a-b70b-fb1c52b0828c (cwd None)

Thread-205::DEBUG::2015-07-20 16:23:30,630::lvm::288::Storage.Misc.excCmd::(cmd) FAILED: <err> = '  Volume group "d6df7930-342a-493a-b70b-fb1c52b0828c" not found\n  Skipping volume group d6df7930-342a-493a-b70b-fb1c52b0828c\n'; <rc> = 5

Thread-205::WARNING::2015-07-20 16:23:30,631::lvm::370::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] ['  Volume group "d6df7930-342a-493a-b70b-fb1c52b0828c" not found', '  Skipping volume group d6df7930-342a-493a-b70b-fb1c52b0828c']

Thread-205::DEBUG::2015-07-20 16:23:30,631::lvm::407::Storage.OperationMutex::(_reloadvgs) Operation 'lvm reload operation' released the operation mutex

Thread-205::ERROR::2015-07-20 16:23:30,644::sdc::143::Storage.StorageDomainCache::(_findDomain) domain d6df7930-342a-493a-b70b-fb1c52b0828c not found

Traceback (most recent call last):

  File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain

  File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain

StorageDomainDoesNotExist: Storage domain does not exist: (u'd6df7930-342a-493a-b70b-fb1c52b0828c',)

Thread-205::INFO::2015-07-20 16:23:30,644::nfsSD::69::Storage.StorageDomain::(create) sdUUID=d6df7930-342a-493a-b70b-fb1c52b0828c domainName=ovirtprd01 remotePath=superstore001-stor.cs.example.com:/ovirtprd01 domClass=1

Thread-205::ERROR::2015-07-20 16:23:30,659::task::866::Storage.TaskManager.Task::(_setError) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::Unexpected error

Traceback (most recent call last):

  File "/usr/share/vdsm/storage/task.py", line 873, in _run

  File "/usr/share/vdsm/logUtils.py", line 45, in wrapper

  File "/usr/share/vdsm/storage/hsm.py", line 2670, in createStorageDomain

  File "/usr/share/vdsm/storage/nfsSD.py", line 80, in create

  File "/usr/share/vdsm/storage/nfsSD.py", line 49, in _preCreateValidation

  File "/usr/share/vdsm/storage/fileSD.py", line 88, in validateFileSystemFeatures

  File "/usr/share/vdsm/storage/outOfProcess.py", line 320, in directTouch

  File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 507, in touch

  File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 391, in _sendCommand

OSError: [Errno 2] No such file or directory

Thread-205::DEBUG::2015-07-20 16:23:30,659::task::885::Storage.TaskManager.Task::(_run) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::Task._run: 050e9378-ba78-4e6f-b986-0dda7bb09aa7 (7, u'd6df7930-342a-493a-b70b-fb1c52b0828c', u'ovirtprd01', u'superstore001-stor.cs.example.com:/ovirtprd01', 1, u'3') {} failed - stopping task

Thread-205::DEBUG::2015-07-20 16:23:30,659::task::1217::Storage.TaskManager.Task::(stop) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::stopping in state preparing (force False)

Thread-205::DEBUG::2015-07-20 16:23:30,659::task::993::Storage.TaskManager.Task::(_decref) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::ref 1 aborting True

Thread-205::INFO::2015-07-20 16:23:30,659::task::1171::Storage.TaskManager.Task::(prepare) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::aborting: Task is aborted: u'[Errno 2] No such file or directory' - code 100

Thread-205::DEBUG::2015-07-20 16:23:30,659::task::1176::Storage.TaskManager.Task::(prepare) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::Prepare: aborted: [Errno 2] No such file or directory

Thread-205::DEBUG::2015-07-20 16:23:30,660::task::993::Storage.TaskManager.Task::(_decref) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::ref 0 aborting True

Thread-205::DEBUG::2015-07-20 16:23:30,660::task::928::Storage.TaskManager.Task::(_doAbort) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::Task._doAbort: force False

Thread-205::DEBUG::2015-07-20 16:23:30,660::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {}

Thread-205::DEBUG::2015-07-20 16:23:30,660::task::595::Storage.TaskManager.Task::(_updateState) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::moving from state preparing -> state aborting

Thread-205::DEBUG::2015-07-20 16:23:30,660::task::550::Storage.TaskManager.Task::(__state_aborting) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::_aborting: recover policy none

Thread-205::DEBUG::2015-07-20 16:23:30,660::task::595::Storage.TaskManager.Task::(_updateState) Task=`050e9378-ba78-4e6f-b986-0dda7bb09aa7`::moving from state aborting -> state failed

Thread-205::DEBUG::2015-07-20 16:23:30,660::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {}

Thread-205::DEBUG::2015-07-20 16:23:30,660::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {}

Thread-205::ERROR::2015-07-20 16:23:30,660::dispatcher::79::Storage.Dispatcher::(wrapper) [Errno 2] No such file or directory

Traceback (most recent call last):

  File "/usr/share/vdsm/storage/dispatcher.py", line 71, in wrapper

  File "/usr/share/vdsm/storage/task.py", line 103, in wrapper

  File "/usr/share/vdsm/storage/task.py", line 1179, in prepare

OSError: [Errno 2] No such file or directory

Thread-205::DEBUG::2015-07-20 16:23:30,660::stompReactor::163::yajsonrpc.StompServer::(send) Sending response

JsonRpc (StompReactor)::DEBUG::2015-07-20 16:23:30,945::stompReactor::98::Broker.StompAdapter::(handle_frame) Handling message <StompFrame command='SEND'>

JsonRpcServer::DEBUG::2015-07-20 16:23:30,945::__init__::506::jsonrpc.JsonRpcServer::(serve_requests) Waiting for request

Thread-208::DEBUG::2015-07-20 16:23:30,945::__init__::469::jsonrpc.JsonRpcServer::(_serveRequest) Calling 'StoragePool.disconnectStorageServer' in bridge with {u'connectionParams': [{u'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709', u'connection': u'superstore001-stor.cs.example.com:/ovirtprd01', u'iqn': u'', u'user': u'', u'tpgt': u'1', u'vfs_type': u'glusterfs', u'password': u'', u'port': u''}], u'storagepoolID': u'00000000-0000-0000-0000-000000000000', u'domainType': 7}

Thread-208::DEBUG::2015-07-20 16:23:30,946::task::595::Storage.TaskManager.Task::(_updateState) Task=`9f38f64c-24e9-4f6b-9487-39384266bc71`::moving from state init -> state preparing

Thread-208::INFO::2015-07-20 16:23:30,947::logUtils::44::dispatcher::(wrapper) Run and protect: disconnectStorageServer(domType=7, spUUID=u'00000000-0000-0000-0000-000000000000', conList=[{u'port': u'', u'connection': u'superstore001-stor.cs.example.com:/ovirtprd01', u'iqn': u'', u'user': u'', u'tpgt': u'1', u'vfs_type': u'glusterfs', u'password': '******', u'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709'}], options=None)

Thread-208::DEBUG::2015-07-20 16:23:30,947::mount::227::Storage.Misc.excCmd::(_runcmd) /usr/bin/sudo -n /usr/bin/umount -f -l /rhev/data-center/mnt/glusterSD/superstore001-stor.cs.example.com:_ovirtprd01 (cwd None)

Thread-208::DEBUG::2015-07-20 16:23:30,958::misc::741::Storage.SamplingMethod::(__call__) Trying to enter sampling method (storage.sdc.refreshStorage)

Thread-208::DEBUG::2015-07-20 16:23:30,958::misc::743::Storage.SamplingMethod::(__call__) Got in to sampling method

Thread-208::DEBUG::2015-07-20 16:23:30,958::misc::741::Storage.SamplingMethod::(__call__) Trying to enter sampling method (storage.iscsi.rescan)

Thread-208::DEBUG::2015-07-20 16:23:30,958::misc::743::Storage.SamplingMethod::(__call__) Got in to sampling method

Thread-208::DEBUG::2015-07-20 16:23:30,958::iscsi::403::Storage.ISCSI::(rescan) Performing SCSI scan, this will take up to 30 seconds

Thread-208::DEBUG::2015-07-20 16:23:30,958::iscsiadm::92::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n /sbin/iscsiadm -m session -R (cwd None)

Thread-208::DEBUG::2015-07-20 16:23:30,971::misc::751::Storage.SamplingMethod::(__call__) Returning last result

Thread-208::DEBUG::2015-07-20 16:23:30,971::misc::741::Storage.SamplingMethod::(__call__) Trying to enter sampling method (storage.hba.rescan)

Thread-208::DEBUG::2015-07-20 16:23:30,971::misc::743::Storage.SamplingMethod::(__call__) Got in to sampling method

Thread-208::DEBUG::2015-07-20 16:23:30,971::hba::53::Storage.HBA::(rescan) Starting scan

Thread-208::DEBUG::2015-07-20 16:23:30,971::utils::739::Storage.HBA::(execCmd) /usr/bin/sudo -n /usr/libexec/vdsm/fc-scan (cwd None)

Thread-208::DEBUG::2015-07-20 16:23:30,989::hba::66::Storage.HBA::(rescan) Scan finished

Thread-208::DEBUG::2015-07-20 16:23:30,989::misc::751::Storage.SamplingMethod::(__call__) Returning last result

Thread-208::DEBUG::2015-07-20 16:23:30,989::multipath::128::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n /sbin/multipath (cwd None)

Thread-208::DEBUG::2015-07-20 16:23:31,025::multipath::128::Storage.Misc.excCmd::(rescan) SUCCESS: <err> = ''; <rc> = 0

Thread-208::DEBUG::2015-07-20 16:23:31,026::lvm::489::Storage.OperationMutex::(_invalidateAllPvs) Operation 'lvm invalidate operation' got the operation mutex

Thread-208::DEBUG::2015-07-20 16:23:31,026::lvm::491::Storage.OperationMutex::(_invalidateAllPvs) Operation 'lvm invalidate operation' released the operation mutex

Thread-208::DEBUG::2015-07-20 16:23:31,026::lvm::500::Storage.OperationMutex::(_invalidateAllVgs) Operation 'lvm invalidate operation' got the operation mutex

Thread-208::DEBUG::2015-07-20 16:23:31,026::lvm::502::Storage.OperationMutex::(_invalidateAllVgs) Operation 'lvm invalidate operation' released the operation mutex

Thread-208::DEBUG::2015-07-20 16:23:31,026::lvm::520::Storage.OperationMutex::(_invalidateAllLvs) Operation 'lvm invalidate operation' got the operation mutex

Thread-208::DEBUG::2015-07-20 16:23:31,026::lvm::522::Storage.OperationMutex::(_invalidateAllLvs) Operation 'lvm invalidate operation' released the operation mutex

Thread-208::DEBUG::2015-07-20 16:23:31,026::misc::751::Storage.SamplingMethod::(__call__) Returning last result

Thread-208::INFO::2015-07-20 16:23:31,026::logUtils::47::dispatcher::(wrapper) Run and protect: disconnectStorageServer, Return response: {'statuslist': [{'status': 0, 'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709'}]}

Thread-208::DEBUG::2015-07-20 16:23:31,026::task::1191::Storage.TaskManager.Task::(prepare) Task=`9f38f64c-24e9-4f6b-9487-39384266bc71`::finished: {'statuslist': [{'status': 0, 'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709'}]}

Thread-208::DEBUG::2015-07-20 16:23:31,027::task::595::Storage.TaskManager.Task::(_updateState) Task=`9f38f64c-24e9-4f6b-9487-39384266bc71`::moving from state preparing -> state finished

Thread-208::DEBUG::2015-07-20 16:23:31,027::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {}

Thread-208::DEBUG::2015-07-20 16:23:31,027::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {}

Thread-208::DEBUG::2015-07-20 16:23:31,027::task::993::Storage.TaskManager.Task::(_decref) Task=`9f38f64c-24e9-4f6b-9487-39384266bc71`::ref 0 aborting False

Thread-208::DEBUG::2015-07-20 16:23:31,027::__init__::500::jsonrpc.JsonRpcServer::(_serveRequest) Return 'StoragePool.disconnectStorageServer' in bridge with [{'status': 0, 'id': u'ece6ad42-be90-4980-98dd-d7ae12cc6709'}]

Thread-208::DEBUG::2015-07-20 16:23:31,027::stompReactor::163::yajsonrpc.StompServer::(send) Sending response

JsonRpc (StompReactor)::DEBUG::2015-07-20 16:23:31,505::stompReactor::98::Broker.StompAdapter::(handle_frame) Handling message <StompFrame command='SEND'>

JsonRpcServer::DEBUG::2015-07-20 16:23:31,506::__init__::506::jsonrpc.JsonRpcServer::(serve_requests) Waiting for request

Thread-209::DEBUG::2015-07-20 16:23:31,507::stompReactor::163::yajsonrpc.StompServer::(send) Sending response

JsonRpc (StompReactor)::DEBUG::2015-07-20 16:23:34,537::stompReactor::98::Broker.StompAdapter::(handle_frame) Handling message <StompFrame command='SEND'>

JsonRpcServer::DEBUG::2015-07-20 16:23:34,537::__init__::506::jsonrpc.JsonRpcServer::(serve_requests) Waiting for request

Thread-210::DEBUG::2015-07-20 16:23:34,538::stompReactor::163::yajsonrpc.StompServer::(send) Sending response


Thanks in advance,


Jurriën


This message (including any attachments) may contain information that is privileged or confidential. If you are not the intended recipient, please notify the sender and delete this email immediately from your systems and destroy all copies of it. You may not, directly or indirectly, use, disclose, distribute, print or copy this email or any part of it if you are not the intended recipient


_______________________________________________
Users mailing list
Users at ovirt.org<mailto:Users at ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users


-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150721/1e39b671/attachment-0001.html>


More information about the Users mailing list