I’ve just reinstall the host and have the same issue, here is the ERROR messages from the vdsm logs :
Thread-43::ERROR::2016-09-06 16:02:54,399::hsm::2551::
Storage.HSM::( disconnectStorageServer) Could not disconnect from storageServer Thread-43::ERROR::2016-09-06 16:02:54,453::hsm::2551::
Storage.HSM::( disconnectStorageServer) Could not disconnect from storageServer Thread-43::ERROR::2016-09-06 16:02:54,475::hsm::2551::
Storage.HSM::( disconnectStorageServer) Could not disconnect from storageServer Thread-51::ERROR::2016-09-06 16:05:38,319::hsm::2453::
Storage.HSM::( connectStorageServer) Could not connect to storageServer Thread-52::ERROR::2016-09-06 16:05:38,636::sdc::137::
Storage.StorageDomainCache::(_ findDomain) looking for unfetched domain cc9ab4b2-9880-427b-8f3b- 61f03e520cbc Thread-52::ERROR::2016-09-06 16:05:38,637::sdc::154::
Storage.StorageDomainCache::(_ findUnfetchedDomain) looking for domain cc9ab4b2-9880-427b-8f3b- 61f03e520cbc Thread-52::ERROR::2016-09-06 16:05:38,756::sdc::143::
Storage.StorageDomainCache::(_ findDomain) domain cc9ab4b2-9880-427b-8f3b- 61f03e520cbc not found Thread-52::ERROR::2016-09-06 16:05:38,769::task::866::
Storage.TaskManager.Task::(_ setError) Task=`1cfc5c30-fc82-44d6-a296- 223fa9426417`::Unexpected error Thread-52::ERROR::2016-09-06 16:05:38,788::dispatcher::76::
Storage.Dispatcher::(wrapper) {'status': {'message': "Cannot find master domain: u'spUUID=00000002-0002-0002- 0002-000000000193, msdUUID=cc9ab4b2-9880-427b- 8f3b-61f03e520cbc'", 'code': 304}}
This SD is present on the host1 (CentOS 6.8) :
[root@unc-srv-hyp1 ~]$ ll /rhev/data-center/
00000002-0002-0002-0002-
000000000193/ mnt/
Not in the host2 (CentOS 7.2) :
[root@unc-srv-hyp2 ~]# ll /rhev/data-center/
total 4
drwxr-xr-x. 5 vdsm kvm 4096 6 sept. 16:03 mnt
The vdsm.log more complete :
Thread-88::DEBUG::2016-09-06 16:22:03,057::iscsi::424::
Storage.ISCSI::(rescan) Performing SCSI scan, this will take up to 30 seconds Thread-88::DEBUG::2016-09-06 16:22:03,057::iscsiadm::97::
Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n /sbin/iscsiadm -m session -R (cwd None) Thread-88::DEBUG::2016-09-06 16:22:03,078::misc::751::
Storage.SamplingMethod::(__ call__) Returning last result Thread-88::DEBUG::2016-09-06 16:22:03,079::misc::741::
Storage.SamplingMethod::(__ call__) Trying to enter sampling method (storage.hba.rescan) Thread-88::DEBUG::2016-09-06 16:22:03,079::misc::743::
Storage.SamplingMethod::(__ call__) Got in to sampling method Thread-88::DEBUG::2016-09-06 16:22:03,080::hba::53::
Storage.HBA::(rescan) Starting scan Thread-88::DEBUG::2016-09-06 16:22:03,080::utils::755::
Storage.HBA::(execCmd) /usr/bin/sudo -n /usr/libexec/vdsm/fc-scan (cwd None) Thread-88::DEBUG::2016-09-06 16:22:03,134::hba::66::
Storage.HBA::(rescan) Scan finished Thread-88::DEBUG::2016-09-06 16:22:03,134::misc::751::
Storage.SamplingMethod::(__ call__) Returning last result Thread-88::DEBUG::2016-09-06 16:22:03,135::multipath::131::
Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n /sbin/multipath (cwd None) Thread-88::DEBUG::2016-09-06 16:22:03,201::multipath::131::
Storage.Misc.excCmd::(rescan) SUCCESS: <err> = ''; <rc> = 0 Thread-88::DEBUG::2016-09-06 16:22:03,202::utils::755::
root::(execCmd) /sbin/udevadm settle --timeout=5 (cwd None) Thread-88::DEBUG::2016-09-06 16:22:03,227::utils::775::
root::(execCmd) SUCCESS: <err> = ''; <rc> = 0 Thread-88::DEBUG::2016-09-06 16:22:03,228::lvm::498::
Storage.OperationMutex::(_ invalidateAllPvs) Operation 'lvm invalidate operation' got the operation mutex Thread-88::DEBUG::2016-09-06 16:22:03,228::lvm::500::
Storage.OperationMutex::(_ invalidateAllPvs) Operation 'lvm invalidate operation' released the operation mutex Thread-88::DEBUG::2016-09-06 16:22:03,229::lvm::509::
Storage.OperationMutex::(_ invalidateAllVgs) Operation 'lvm invalidate operation' got the operation mutex Thread-88::DEBUG::2016-09-06 16:22:03,229::lvm::511::
Storage.OperationMutex::(_ invalidateAllVgs) Operation 'lvm invalidate operation' released the operation mutex Thread-88::DEBUG::2016-09-06 16:22:03,229::lvm::529::
Storage.OperationMutex::(_ invalidateAllLvs) Operation 'lvm invalidate operation' got the operation mutex Thread-88::DEBUG::2016-09-06 16:22:03,230::lvm::531::
Storage.OperationMutex::(_ invalidateAllLvs) Operation 'lvm invalidate operation' released the operation mutex Thread-88::DEBUG::2016-09-06 16:22:03,230::misc::751::
Storage.SamplingMethod::(__ call__) Returning last result Thread-88::ERROR::2016-09-06 16:22:03,230::sdc::137::
Storage.StorageDomainCache::(_ findDomain) looking for unfetched domain cc9ab4b2-9880-427b-8f3b- 61f03e520cbc Thread-88::ERROR::2016-09-06 16:22:03,231::sdc::154::
Storage.StorageDomainCache::(_ findUnfetchedDomain) looking for domain cc9ab4b2-9880-427b-8f3b- 61f03e520cbc Thread-88::DEBUG::2016-09-06 16:22:03,231::lvm::371::
Storage.OperationMutex::(_ reloadvgs) Operation 'lvm reload operation' got the operation mutex Thread-88::DEBUG::2016-09-06 16:22:03,233::lvm::291::
Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n /sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 obtain_device_list_from_udev=0 filter = [ '\''r|.*|'\'' ] } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days = 0 } ' --noheadings --units b --nosuffix --separator '|' --ignoreskippedcluster -o uuid,name,attr,size,free, extent_size,extent_count,free_ count,tags,vg_mda_size,vg_mda_ free,lv_count,pv_count,pv_name cc9ab4b2-9880-427b-8f3b- 61f03e520cbc (cwd None) Thread-88::DEBUG::2016-09-06 16:22:03,275::lvm::291::
Storage.Misc.excCmd::(cmd) FAILED: <err> = ' WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it!\n Volume group "cc9ab4b2-9880-427b-8f3b- 61f03e520cbc" not found\n Cannot process volume group cc9ab4b2-9880-427b-8f3b- 61f03e520cbc\n'; <rc> = 5 Thread-88::WARNING::2016-09-06 16:22:03,277::lvm::376::
Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] [' WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it!', ' Volume group "cc9ab4b2-9880-427b-8f3b- 61f03e520cbc" not found', ' Cannot process volume group cc9ab4b2-9880-427b-8f3b- 61f03e520cbc'] Thread-88::DEBUG::2016-09-06 16:22:03,277::lvm::416::
Storage.OperationMutex::(_ reloadvgs) Operation 'lvm reload operation' released the operation mutex Thread-90::DEBUG::2016-09-06 16:22:03,293::__init__::318::
IOProcessClient::(_run) Starting IOProcess... Thread-91::DEBUG::2016-09-06 16:22:03,314::__init__::318::
IOProcessClient::(_run) Starting IOProcess... Thread-88::ERROR::2016-09-06 16:22:03,334::sdc::143::
Storage.StorageDomainCache::(_ findDomain) domain cc9ab4b2-9880-427b-8f3b- 61f03e520cbc not found Traceback (most recent call last):
File "/usr/share/vdsm/storage/sdc.
py", line 141, in _findDomain dom = findMethod(sdUUID)
File "/usr/share/vdsm/storage/sdc.
py", line 171, in _findUnfetchedDomain raise se.StorageDomainDoesNotExist(
sdUUID) StorageDomainDoesNotExist: Storage domain does not exist: (u'cc9ab4b2-9880-427b-8f3b-
61f03e520cbc',) Thread-88::DEBUG::2016-09-06 16:22:03,335::resourceManager:
:616::Storage.ResourceManager: :(releaseResource) Trying to release resource 'Storage.00000002-0002-0002- 0002-000000000193' Thread-88::DEBUG::2016-09-06 16:22:03,336::resourceManager:
:635::Storage.ResourceManager: :(releaseResource) Released resource 'Storage.00000002-0002-0002- 0002-000000000193' (0 active users) Thread-88::DEBUG::2016-09-06 16:22:03,336::resourceManager:
:641::Storage.ResourceManager: :(releaseResource) Resource 'Storage.00000002-0002-0002- 0002-000000000193' is free, finding out if anyone is waiting for it. Thread-88::DEBUG::2016-09-06 16:22:03,337::resourceManager:
:649::Storage.ResourceManager: :(releaseResource) No one is waiting for resource 'Storage.00000002-0002-0002- 0002-000000000193', Clearing records. Thread-88::DEBUG::2016-09-06 16:22:03,337::resourceManager:
:616::Storage.ResourceManager: :(releaseResource) Trying to release resource 'Storage.HsmDomainMonitorLock' Thread-88::DEBUG::2016-09-06 16:22:03,338::resourceManager:
:635::Storage.ResourceManager: :(releaseResource) Released resource 'Storage.HsmDomainMonitorLock' (0 active users) Thread-88::DEBUG::2016-09-06 16:22:03,338::resourceManager:
:641::Storage.ResourceManager: :(releaseResource) Resource 'Storage.HsmDomainMonitorLock' is free, finding out if anyone is waiting for it. Thread-88::DEBUG::2016-09-06 16:22:03,338::resourceManager:
:649::Storage.ResourceManager: :(releaseResource) No one is waiting for resource 'Storage.HsmDomainMonitorLock' , Clearing records. Thread-88::ERROR::2016-09-06 16:22:03,338::task::866::
Storage.TaskManager.Task::(_ setError) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::Unexpected error Traceback (most recent call last):
File "/usr/share/vdsm/storage/task.
py", line 873, in _run return fn(*args, **kargs)
File "/usr/share/vdsm/logUtils.py", line 45, in wrapper
res = f(*args, **kwargs)
File "/usr/share/vdsm/storage/hsm.
py", line 1039, in connectStoragePool spUUID, hostID, msdUUID, masterVersion, domainsMap)
File "/usr/share/vdsm/storage/hsm.
py", line 1104, in _connectStoragePool res = pool.connect(hostID, msdUUID, masterVersion)
File "/usr/share/vdsm/storage/sp.
py", line 637, in connect self.__rebuild(msdUUID=
msdUUID, masterVersion=masterVersion) File "/usr/share/vdsm/storage/sp.
py", line 1179, in __rebuild self.setMasterDomain(msdUUID, masterVersion)
File "/usr/share/vdsm/storage/sp.
py", line 1390, in setMasterDomain raise se.StoragePoolMasterNotFound(
self.spUUID, msdUUID) StoragePoolMasterNotFound: Cannot find master domain: u'spUUID=00000002-0002-0002-
0002-000000000193, msdUUID=cc9ab4b2-9880-427b- 8f3b-61f03e520cbc' Thread-88::DEBUG::2016-09-06 16:22:03,339::task::885::
Storage.TaskManager.Task::(_ run) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::Task._run: 4898233c-f6c6-45fa-a2ee- 42e63e189063 (u'00000002-0002-0002-0002- 000000000193', 2, u'cc9ab4b2-9880-427b-8f3b- 61f03e520cbc', 7, {u'015799ac-ec66-4a9d-a8d6- 6e9ec980972d': u'active', u'2fcd37ce-cb88-4026-88df- d4d472b41ecf': u'active', u'76a1fed9-2e60-4b3e-9f00- efca8acd133d': u'active', u'7e40772a-fe94-4fb2-94c4- 6198bed04a6a': u'active', u'cb4c84c1-489a-433f-999a- f1aeec9d62cf': u'active', u'ea05d014-f8f0-4f1d-906e- 2e93c8907d7d': u'active', u'cc9ab4b2-9880-427b-8f3b- 61f03e520cbc': u'active'}) {} failed - stopping task Thread-88::DEBUG::2016-09-06 16:22:03,339::task::1217::
Storage.TaskManager.Task::( stop) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::stopping in state preparing (force False) Thread-88::DEBUG::2016-09-06 16:22:03,340::task::993::
Storage.TaskManager.Task::(_ decref) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::ref 1 aborting True Thread-88::INFO::2016-09-06 16:22:03,340::task::1171::
Storage.TaskManager.Task::( prepare) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::aborting: Task is aborted: 'Cannot find master domain' - code 304 Thread-88::DEBUG::2016-09-06 16:22:03,340::task::1176::
Storage.TaskManager.Task::( prepare) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::Prepare: aborted: Cannot find master domain Thread-88::DEBUG::2016-09-06 16:22:03,341::task::993::
Storage.TaskManager.Task::(_ decref) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::ref 0 aborting True Thread-88::DEBUG::2016-09-06 16:22:03,341::task::928::
Storage.TaskManager.Task::(_ doAbort) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::Task._doAbort: force False Thread-88::DEBUG::2016-09-06 16:22:03,341::resourceManager:
:977::Storage.ResourceManager. Owner::(cancelAll) Owner.cancelAll requests {} Thread-88::DEBUG::2016-09-06 16:22:03,341::task::595::
Storage.TaskManager.Task::(_ updateState) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::moving from state preparing -> state aborting Thread-88::DEBUG::2016-09-06 16:22:03,342::task::550::
Storage.TaskManager.Task::(__ state_aborting) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::_aborting: recover policy none Thread-88::DEBUG::2016-09-06 16:22:03,342::task::595::
Storage.TaskManager.Task::(_ updateState) Task=`4898233c-f6c6-45fa-a2ee- 42e63e189063`::moving from state aborting -> state failed Thread-88::DEBUG::2016-09-06 16:22:03,342::resourceManager:
:940::Storage.ResourceManager. Owner::(releaseAll) Owner.releaseAll requests {} resources {} Thread-88::DEBUG::2016-09-06 16:22:03,342::resourceManager:
:977::Storage.ResourceManager. Owner::(cancelAll) Owner.cancelAll requests {} Thread-88::ERROR::2016-09-06 16:22:03,343::dispatcher::76::
Storage.Dispatcher::(wrapper) {'status': {'message': "Cannot find master domain: u'spUUID=00000002-0002-0002- 0002-000000000193, msdUUID=cc9ab4b2-9880-427b- 8f3b-61f03e520cbc'", 'code': 304}}
I ‘ve restart the service lvm2-lvmetad.service and enable it as adviced in the log without success
The engine.log :
2016-09-06 16:20:02,971 INFO [org.ovirt.engine.core.bll.
InitVdsOnUpCommand] (DefaultQuartzScheduler_ Worker-26) [6f596728] Running command: InitVdsOnUpCommand internal: true. Entities affected : ID: 00000002-0002-0002-0002- 000000000193 Type: StoragePool 2016-09-06 16:20:02,982 INFO [org.ovirt.engine.core.bll.
storage. ConnectHostToStoragePoolServer sCommand] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] Running command: ConnectHostToStoragePoolServer sCommand internal: true. Entities affected : ID: 00000002-0002-0002-0002- 000000000193 Type: StoragePool 2016-09-06 16:20:03,035 INFO [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStorageServerVDSCommand ] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] START, ConnectStorageServerVDSCommand (HostName = unc-srv-hyp2, HostId = 5bf45c09-41d4-4125-a4bd- 81af2a100db8, storagePoolId = 00000002-0002-0002-0002- 000000000193, storageType = NFS, connectionList = [{ id: 2ceca65a-90a0-4daf-82a3- 366de490a71e, connection: unc-srv-oman.cfu.local:/var/ lib/exports/iso, iqn: null, vfsType: null, mountOptions: null, nfsVersion: null, nfsRetrans: null, nfsTimeo: null };{ id: 76a3f706-e7d4-49cd-9e0a- da6061b6b2d6, connection: unc-srv-hyp1.cfu.local:/ exports/import_domain, iqn: null, vfsType: null, mountOptions: null, nfsVersion: null, nfsRetrans: null, nfsTimeo: null };{ id: ea46418b-68b3-4f9b-9316- b1d57f17ecbc, connection: unc-srv-oman.cfu.local:/data/ Master, iqn: null, vfsType: null, mountOptions: null, nfsVersion: null, nfsRetrans: null, nfsTimeo: null };]), log id: 315222a9 2016-09-06 16:20:03,080 INFO [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStorageServerVDSCommand ] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] FINISH, ConnectStorageServerVDSCommand , return: {ea46418b-68b3-4f9b-9316- b1d57f17ecbc=0, 76a3f706-e7d4-49cd-9e0a- da6061b6b2d6=0, 2ceca65a-90a0-4daf-82a3- 366de490a71e=0}, log id: 315222a9 2016-09-06 16:20:03,115 INFO [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStorageServerVDSCommand ] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] START, ConnectStorageServerVDSCommand (HostName = unc-srv-hyp2, HostId = 5bf45c09-41d4-4125-a4bd- 81af2a100db8, storagePoolId = 00000002-0002-0002-0002- 000000000193, storageType = ISCSI, connectionList = [{ id: 3efad56e-a86d-4682-b799- 51f42713cda6, connection: 192.168.4.1, iqn: iqn.1984-05.com.dell: powervault.md3600i. 690b11c0005592a90000000051c3ef c7, vfsType: null, mountOptions: null, nfsVersion: null, nfsRetrans: null, nfsTimeo: null };]), log id: f9c9ff3 2016-09-06 16:22:03,952 INFO [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStorageServerVDSCommand ] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] FINISH, ConnectStorageServerVDSCommand , return: {3efad56e-a86d-4682-b799- 51f42713cda6=465}, log id: f9c9ff3 2016-09-06 16:22:03,976 INFO [org.ovirt.engine.core.bll.
storage.ISCSIStorageHelper] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] The lun with id xu5AAG-1FHh-Qxbx-ZXIe-55J0- vscw-kb6fdp was reported as problematic ! 2016-09-06 16:22:03,992 ERROR [org.ovirt.engine.core.dal.
dbbroker.auditloghandling. AuditLogDirector] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: The error message for connection 192.168.4.1 iqn.1984-05.com.dell: powervault.md3600i. 690b11c0005592a90000000051c3ef c7 (LUN mpathc) returned by VDSM was: Failed to setup iSCSI subsystem 2016-09-06 16:22:03,997 ERROR [org.ovirt.engine.core.bll.
storage.ISCSIStorageHelper] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] The connection with details 192.168.4.1 iqn.1984-05.com.dell: powervault.md3600i. 690b11c0005592a90000000051c3ef c7 (LUN mpathc) failed because of error code 465 and error message is: failed to setup iscsi subsystem
2016-09-06 16:22:03,998 INFO [org.ovirt.engine.core.bll.
storage. ConnectHostToStoragePoolServer sCommand] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] Host unc-srv-hyp2 storage connection was failed 2016-09-06 16:22:04,004 WARN [org.ovirt.engine.core.dal.
dbbroker.auditloghandling. AuditLogDirector] (DefaultQuartzScheduler_ Worker-26) [7d7c3300] Correlation ID: 7d7c3300, Call Stack: null, Custom Event ID: -1, Message: Failed to connect Host unc-srv-hyp2 to Storage Servers 2016-09-06 16:22:04,040 INFO [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStoragePoolVDSCommand] (org.ovirt.thread.pool-8- thread-20) START, ConnectStoragePoolVDSCommand( HostName = unc-srv-hyp2, HostId = 5bf45c09-41d4-4125-a4bd- 81af2a100db8, vdsId = 5bf45c09-41d4-4125-a4bd- 81af2a100db8, storagePoolId = 00000002-0002-0002-0002- 000000000193, masterVersion = 7), log id: 321780b7 2016-09-06 16:22:04,346 INFO [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStoragePoolVDSCommand] (org.ovirt.thread.pool-8- thread-20) Command org.ovirt.engine.core. vdsbroker.vdsbroker. ConnectStoragePoolVDSCommand return value StatusOnlyReturnForXmlRpc [mStatus=StatusForXmlRpc [mCode=304, mMessage=Cannot find master domain: u'spUUID=00000002-0002-0002-
0002-000000000193, msdUUID=cc9ab4b2-9880-427b- 8f3b-61f03e520cbc']] 2016-09-06 16:22:04,347 INFO [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStoragePoolVDSCommand] (org.ovirt.thread.pool-8- thread-20) HostName = unc-srv-hyp2 2016-09-06 16:22:04,348 ERROR [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStoragePoolVDSCommand] (org.ovirt.thread.pool-8- thread-20) Command ConnectStoragePoolVDSCommand( HostName = unc-srv-hyp2, HostId = 5bf45c09-41d4-4125-a4bd- 81af2a100db8, vdsId = 5bf45c09-41d4-4125-a4bd- 81af2a100db8, storagePoolId = 00000002-0002-0002-0002- 000000000193, masterVersion = 7) execution failed. Exception: IRSNoMasterDomainException: IRSGenericException: IRSErrorException: IRSNoMasterDomainException: Cannot find master domain: u'spUUID=00000002-0002-0002- 0002-000000000193, msdUUID=cc9ab4b2-9880-427b- 8f3b-61f03e520cbc' 2016-09-06 16:22:04,350 INFO [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStoragePoolVDSCommand] (org.ovirt.thread.pool-8- thread-20) FINISH, ConnectStoragePoolVDSCommand, log id: 321780b7 2016-09-06 16:22:04,351 ERROR [org.ovirt.engine.core.bll.
InitVdsOnUpCommand] (org.ovirt.thread.pool-8- thread-20) Could not connect host unc-srv-hyp2 to pool UNICEF with the message: null 2016-09-06 16:22:04,399 INFO [org.ovirt.engine.core.bll.
SetNonOperationalVdsCommand] (DefaultQuartzScheduler_ Worker-26) [60403257] Running command: SetNonOperationalVdsCommand internal: true. Entities affected : ID: 5bf45c09-41d4-4125-a4bd- 81af2a100db8 Type: VDS 2016-09-06 16:22:04,426 INFO [org.ovirt.engine.core.
vdsbroker. SetVdsStatusVDSCommand] (DefaultQuartzScheduler_ Worker-26) [60403257] START, SetVdsStatusVDSCommand( HostName = unc-srv-hyp2, HostId = 5bf45c09-41d4-4125-a4bd- 81af2a100db8, status=NonOperational, nonOperationalReason=STORAGE_ DOMAIN_UNREACHABLE, stopSpmFailureLogged=false), log id: 7f47d8b 2016-09-06 16:22:04,435 INFO [org.ovirt.engine.core.
vdsbroker. SetVdsStatusVDSCommand] (DefaultQuartzScheduler_ Worker-26) [60403257] FINISH, SetVdsStatusVDSCommand, log id: 7f47d8b 2016-09-06 16:22:04,448 WARN [org.ovirt.engine.core.dal.
dbbroker.auditloghandling. AuditLogDirector] (DefaultQuartzScheduler_ Worker-26) [60403257] Correlation ID: 60403257, Job ID: 4203a573-20f2-45e5-b251- e4e09f1ed951, Call Stack: null, Custom Event ID: -1, Message: Host unc-srv-hyp2 cannot access the Storage Domain(s) <UNKNOWN> attached to the Data Center UNICEF. Setting Host state to Non-Operational. 2016-09-06 16:22:04,484 WARN [org.ovirt.engine.core.dal.
dbbroker.auditloghandling. AuditLogDirector] (DefaultQuartzScheduler_ Worker-26) [60403257] Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: Power Management test failed for Host unc-srv-hyp2.There is no other host in the data center that can be used to test the power management settings. 2016-09-06 16:22:04,488 WARN [org.ovirt.engine.core.dal.
dbbroker.auditloghandling. AuditLogDirector] (DefaultQuartzScheduler_ Worker-26) [60403257] Correlation ID: 6f596728, Call Stack: null, Custom Event ID: -1, Message: Failed to connect Host unc-srv-hyp2 to Storage Pool UNICEF 2016-09-06 16:22:04,528 INFO [org.ovirt.engine.core.bll.
HandleVdsVersionCommand] (DefaultQuartzScheduler_ Worker-26) [709cb459] Running command: HandleVdsVersionCommand internal: true. Entities affected : ID: 5bf45c09-41d4-4125-a4bd- 81af2a100db8 Type: VDS 2016-09-06 16:22:04,530 INFO [org.ovirt.engine.core.
vdsbroker. VdsUpdateRunTimeInfo] (DefaultQuartzScheduler_ Worker-26) [709cb459] Host 5bf45c09-41d4-4125-a4bd- 81af2a100db8 : unc-srv-hyp2 is already in NonOperational status for reason STORAGE_DOMAIN_UNREACHABLE. SetNonOperationalVds command is skipped.
De : Simone Tiraboschi [mailto:stirabos@redhat.com]
Alain VONDRA
Chargé d'exploitation des Systèmes d'Information
Direction Administrative et Financière
+33 1 44 39 77 76
UNICEF France
3 rue Duguay Trouin 75006 PARIS
www.unicef.fr
Envoyé : lundi 5 septembre 2016 17:21
À : VONDRA Alain <AVONDRA@unicef.fr>
Cc : Yedidyah Bar David <didi@redhat.com>; users <users@ovirt.org>
Objet : Re: [ovirt-users] HELP Upgrade hypervisors from CentOS 6.8 to CentOS 7
On Mon, Sep 5, 2016 at 5:09 PM, VONDRA Alain <AVONDRA@unicef.fr> wrote:
To resume the situation and be the most clear I can.
I have :
- 2 hypervisors (physical) with CentOS 6.8 : HYP1 and HYP2
- 1 oVirt manager 3.5 (physical and not hosted-engine) with CentOS 7.2
And to be able to upgrade to oVirt 3.6 or more I need to upgrade the 2 hypervisors to CentOS 7.2.
So I began with HYP2, I’ve removed it from the Cluster, and installed it with CentOS 7.2, but when I wanted to join it to a new cluster as shown in the doc, HYP2 stayed always unresponsive.
We need vdsm logs to understand the issue here.
Could you please reproduce it?
So I’ve restored the HYP2 to Centos 6.8, reinstalled it on the previous cluster and everything works well, but I really want to upgrade to CentOS 7.2 to be up-to-date, because I will have soonly more than two versions late.
Alain VONDRA
Chargé d'exploitation des Systèmes d'Information
Direction Administrative et Financière
+33 1 44 39 77 76
UNICEF France
3 rue Duguay Trouin 75006 PARIS
www.unicef.fr
De : Yedidyah Bar David [mailto:didi@redhat.com]
Envoyé : lundi 5 septembre 2016 15:53
À : VONDRA Alain <AVONDRA@unicef.fr>
Cc : Simone Tiraboschi <stirabos@redhat.com>; users <users@ovirt.org>
Objet : Re: [ovirt-users] HELP Upgrade hypervisors from CentOS 6.8 to CentOS 7
On Mon, Sep 5, 2016 at 4:34 PM, VONDRA Alain <AVONDRA@unicef.fr> wrote:
As you say, “This should”, but is there anybody there who tried this operation ?
When I wrote [1], it worked for me.
If something does not work for you, I think it's best to open a bug
and attach all relevant logs, including hosted-engine-setup and vdsm
logs from the host and all engine logs from the engine vm.
I can't tell from your log snippet why your new host failed to attach
to the storage domain. If it's reproducible, please check/post vdsm logs.
[1] https://www.ovirt.org/documentation/how-to/hosted- engine-host-OS-upgrade/
Is there any other option to upgrade the hypervisors ?
Even if there is another option, the fact that you can't add a host
is probably problematic in itself, no? What if you actually need to
add a host?
Did you try adding a el6 host and it did work?
Best,
Alain VONDRA
Chargé d'exploitation des Systèmes d'Information
Direction Administrative et Financière
+33 1 44 39 77 76
UNICEF France
3 rue Duguay Trouin 75006 PARIS
www.unicef.fr
De : Yedidyah Bar David [mailto:didi@redhat.com]
Envoyé : lundi 5 septembre 2016 12:22
À : VONDRA Alain <AVONDRA@unicef.fr>
Cc : Simone Tiraboschi <stirabos@redhat.com>; users <users@ovirt.org>
Objet : Re: [ovirt-users] HELP Upgrade hypervisors from CentOS 6.8 to CentOS 7
On Mon, Sep 5, 2016 at 1:11 PM, VONDRA Alain <AVONDRA@unicef.fr> wrote:
Unlikely, I didn’t save them, I had to rollback the host quickly before the Week-end.
All that I can tell, that is all seemed to work well during the installation of the host, all the networks were connected to the SAN, but the host didn’t want to go UP, still staying unresponsive with the message below from the oVirt engine.
Can you assure me that the installation of a physical hypervisor with CentOS 7, is possible if I put it on a different cluster than the other host using CentOS 6.8 ?
Yes, this should work.
Thanks
Alain VONDRA
Chargé d'exploitation des Systèmes d'Information
Direction Administrative et Financière
+33 1 44 39 77 76
UNICEF France
3 rue Duguay Trouin 75006 PARIS
www.unicef.fr
De : Yedidyah Bar David [mailto:didi@redhat.com]
Envoyé : dimanche 4 septembre 2016 08:50
À : VONDRA Alain <AVONDRA@unicef.fr>
Cc : Simone Tiraboschi <stirabos@redhat.com>; users <users@ovirt.org>
Objet : Re: [ovirt-users] HELP Upgrade hypervisors from CentOS 6.8 to CentOS 7
On Fri, Sep 2, 2016 at 7:31 PM, VONDRA Alain <AVONDRA@unicef.fr> wrote:
Hi,
I’ve followed this doc, and hoped to find a solution with it, so I did’nt use the hosted-engine –deploy command, I’ve added a new host and at the end of the installation, the host stays unresponsive because unable to be attached to the same Storage volume .
Engine log :
2016-09-02 16:57:01,780 ERROR [org.ovirt.engine.core.bll.
InitVdsOnUpCommand] (org.ovirt.thread.pool-8- thread-28) Could not connect host unc-srv-hyp2 to pool UNICEF with the message: null 2016-09-02 17:00:01,634 ERROR [org.ovirt.engine.core.
vdsbroker.vdsbroker. ConnectStoragePoolVDSCommand] (org.ovirt.thread.pool-8- thread-3) Command ConnectStoragePoolVDSCommand( HostName = unc-srv-hyp2, HostId = ee1c57ce-1c77-47b1-b466- 7bf99382dd77, vdsId = ee1c57ce-1c77-47b1-b466- 7bf99382dd77, storagePoolId = 00000002-0002-0002-0002- 000000000193, masterVersion = 7) execution failed. Exception: IRSNoMasterDomainException: IRSGenericException: IRSErrorException: IRSNoMasterDomainException: Cannot find master domain: u'spUUID=00000002-0002-0002- 0002-000000000193, msdUUID=cc9ab4b2-9880-427b- 8f3b-61f03e520cbc'
Please check/post vdsm logs from the host. Thanks.
Alain VONDRA
Chargé d'exploitation des Systèmes d'Information
Direction Administrative et Financière
+33 1 44 39 77 76
UNICEF France
3 rue Duguay Trouin 75006 PARIS
www.unicef.fr
De : Simone Tiraboschi [mailto:stirabos@redhat.com]
Envoyé : vendredi 2 septembre 2016 18:20
À : VONDRA Alain <AVONDRA@unicef.fr>
Cc : Nir Soffer <nsoffer@redhat.com>; users <users@ovirt.org>
Objet : Re: [ovirt-users] HELP Upgrade hypervisors from CentOS 6.8 to CentOS 7
On Fri, Sep 2, 2016 at 5:21 PM, VONDRA Alain <AVONDRA@unicef.fr> wrote:
Hi,
I'd like to upgrade my oVirt environment from 3.5 to 3.6 and maybe 4, actually the manager oVirt is un 3.5 version, installed on a CentOS 7.2 and the two hypervisors are installed on CentOS 6.8.
I need anyway to upgrade the hosts to be able to move to 3.6.
I've tried to upgrade the first host, but I had of course issues telling that it can't be possible to mix different OS in the same cluster, I've also tried to create another cluster to put this host without success.
What is the best way to upgrade cleanly and safely ?
Thank you in advance for your advices
Follow this:
Simply ignore the steps that refers to the engine VM if you are with the engine on a physical system.
Alain VONDRA
Chargé d'exploitation des Systèmes d'Information
Direction Administrative et Financière
+33 1 44 39 77 76
UNICEF France
3 rue Duguay Trouin 75006
PARIS
www.unicef.fr
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
--Didi
--Didi
--Didi