[ovirt-users] ovirt 3.6 and self hosted engine: clarification on datacenter input

Simone Tiraboschi stirabos at redhat.com
Wed Oct 28 15:10:17 UTC 2015


On Wed, Oct 28, 2015 at 3:12 PM, Gianluca Cecchi <gianluca.cecchi at gmail.com>
wrote:

> On Tue, Oct 27, 2015 at 5:06 PM, Simone Tiraboschi <stirabos at redhat.com>
> wrote:
>
>>
>>
>>> I don't understand the meaning of the sentence above:
>>>
>>>           Local storage datacenter name is an internal name
>>>           and currently will not be shown in engine's admin UI.
>>>
>>>
>> It's just an internal label. I think we can just remove that question
>> always using the default value and nothing will change.
>>
>
> Probably better.
>
>
>>
>>
>>>
>>> How is the chosen "she_datacenter" name related with the "Default"
>>> datacenter where the hypervisor is put in? Do I have to manually create it
>>> (I don't see this se_datacenter in webadmin portal)?
>>>
>>> Also, I know there is open bug
>>>
>>> https://bugzilla.redhat.com/show_bug.cgi?id=1269768
>>>
>>> But it seems I'm not able to to import the storage domain...
>>> In events, when I import, I have the sequence
>>>
>>> Storage Domain she_sdomain was added by admin at internal
>>> VDSM ovc71.localdomain.local command failed: Cannot acquire host id:
>>> (u'9f1ec45d-0c32-4bfc-8b67-372d6f204fd1', SanlockException(22, 'Sanlock
>>> lockspace add failure', 'Invalid argument'))
>>> Failed to attach Storage Domains to Data Center Default. (User:
>>> admin at internal)
>>> Failed to attach Storage Domain she_sdomain to Data Center Default.
>>> (User: admin at internal)
>>>
>>> What should be the flow to compensate the bug? Do I have actually to
>>> attache it to "Default" datacenter or what? Is it expected to be fixed
>>> before 3.6?
>>>
>>
>> Postponing to 3.6.1 not being identified as a blocker.
>>
>
> But is this a regression from 3.5.x or did this problem exist also in all
> the 3.5 versions where sh engine was in place?
>

It's not a regression cause the hosted-engine storage domain wasn't neither
visible in 3.5.
Once again, also if you see it in the engine you cannot use it for anything
apart from the engine VM itself, you still have to add another storage
domain for regular VMs.


>
>
>>
>> You can try to add the first additional storage domain for other VMs.
>> The datacenter should came up and at that point you try importing the
>> hosted-engine storage domain.
>> You cannot add other VMs to that storage domain neither you'll can when
>> the auto-import will work.
>>
>>
>>>
>>>
> So I was indeed able to add a separate data NFS domain and able to attach
> it to the default DC that came then up as an active one.
> Then tried to import/attach also the sh engine domain; it went in locked
> state but then the sh engine VM itself went down (no qemu process on
> hypervisor).
> In /var/log/libvirt/qemu/HostedEngine.log of hypervisor I can see
>
> 2015-10-28 13:59:02.233+0000: shutting down
>
> expected?
>
> what now to have the sh engine come up again and see what happened?
> Any logs on hypevisor to check?
>
> In /var/log/sanlock.log
> 2015-10-28 14:57:14+0100 854 [829]: s4 lockspace
> 3662a51f-39de-4533-97fe-d49bf98e2d43:1:/rhev/data-center/mnt/ovc71.localdomain.local:_NFS__DOMAIN/3662a51f-39de-4533-97fe-d49bf98e2d43/dom_md/ids:0
> 2015-10-28 14:57:34+0100 874 [829]: s4:r3 resource
> 3662a51f-39de-4533-97fe-d49bf98e2d43:SDM:/rhev/data-center/mnt/ovc71.localdomain.local:_NFS__DOMAIN/3662a51f-39de-4533-97fe-d49bf98e2d43/dom_md/leases:1048576
> for 4,17,1698
> 2015-10-28 14:57:35+0100 875 [825]: s4 host 1 1 854
> 1bfba2b1-2353-4d4e-9000-f97585b54df1.ovc71.loca
> 2015-10-28 14:57:35+0100 875 [825]: s4 host 250 1 0
> 1bfba2b1-2353-4d4e-9000-f97585b54df1.ovc71.loca
> 2015-10-28 14:59:00+0100 960 [830]: s1:r4 resource
> 9f1ec45d-0c32-4bfc-8b67-372d6f204fd1:SDM:/rhev/data-center/mnt/ovc71.localdomain.local:_SHE__DOMAIN/9f1ec45d-0c32-4bfc-8b67-372d6f204fd1/dom_md/leases:1048576
> for 4,17,1698
> 2015-10-28 14:59:02+0100 962 [825]: s1 kill 3341 sig 9 count 1
> 2015-10-28 14:59:02+0100 962 [825]: dead 3341 ci 2 count 1
> 2015-10-28 14:59:08+0100 968 [830]: s5 lockspace
> 9f1ec45d-0c32-4bfc-8b67-372d6f204fd1:1:/rhev/data-center/mnt/ovc71.localdomain.local:_SHE__DOMAIN/9f1ec45d-0c32-4bfc-8b67-372d6f204fd1/dom_md/ids:0
> 2015-10-28 14:59:30+0100 990 [825]: s5 host 1 4 968
> 1bfba2b1-2353-4d4e-9000-f97585b54df1.ovc71.loca
> 2015-10-28 14:59:30+0100 990 [825]: s5 host 250 1 0
> aa89bb89-20a1-414b-8ee3-0430fdc330f8.ovc71.loca
>
>
>
> /var/log/vdsm/vdsm.log
> Thread-1247::DEBUG::2015-10-28
> 14:59:00,043::task::993::Storage.TaskManager.Task::(_decref) Task=`56
> dd2372-f454-4188-8bf3-ab543d677c14`::ref 0 aborting False
> Thread-1247::ERROR::2015-10-28 14:59:00,096::API::1847::vds::(_getHaInfo)
> failed to retrieve Hosted Engine HA info
> Traceback (most recent call last):
>   File "/usr/share/vdsm/API.py", line 1827, in _getHaInfo
>     stats = instance.get_all_stats()
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
> line 103, in get_all_stats
>     self._configure_broker_conn(broker)
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
> line 180, in _configure_broker_conn
>     dom_type=dom_type)
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
> line 176, in set_storage_domain
>     .format(sd_type, options, e))
> RequestError: Failed to set storage domain FilesystemBackend, options
> {'dom_type': 'nfs3', 'sd_uuid': '9f1ec45d-0c32-4bfc-8b67-372d6f204fd1'}:
> Request failed: <class
> 'ovirt_hosted_engine_ha.lib.storage_backends.BackendFailureException'>
> Thread-1247::INFO::2015-10-28
> 14:59:00,112::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Reques
> t handler for 127.0.0.1:42165 stopped
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,137::__init__::503::jsonrpc.JsonRpcServer::(_serveRequest) Calling
> 'StoragePool.connectStorageServer' in bridge with {u'connectionParams':
> [{u'id': u'189c29a5-6830-453c-aca3-7d82f2382dd8', u'connection':
> u'ovc71.localdomain.local:/SHE_DOMAIN', u'iqn': u'', u'user': u'',
> u'protocol_version': u'3', u'tpgt': u'1', u'password': '********', u'port':
> u''}], u'storagepoolID': u'00000000-0000-0000-0000-000000000000',
> u'domainType': 1}
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,138::task::595::Storage.TaskManager.Task::(_updateState)
> Task=`9ca908a0-45e2-41d5-802c-dc0bd2414a69`::moving from state init ->
> state preparing
> Thread-1248::INFO::2015-10-28
> 14:59:00,139::logUtils::48::dispatcher::(wrapper) Run and protect:
> connectStorageServer(domType=1,
> spUUID=u'00000000-0000-0000-0000-000000000000', conList=[{u'id':
> u'189c29a5-6830-453c-aca3-7d82f2382dd8', u'connection':
> u'ovc71.localdomain.local:/SHE_DOMAIN', u'iqn': u'', u'user': u'',
> u'protocol_version': u'3', u'tpgt': u'1', u'password': '********', u'port':
> u''}], options=None)
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,142::fileUtils::143::Storage.fileUtils::(createdir) Creating
> directory: /rhev/data-center/mnt/ovc71.localdomain.local:_SHE__DOMAIN mode:
> None
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,143::mount::229::Storage.Misc.excCmd::(_runcmd) /usr/bin/sudo -n
> /usr/bin/mount -t nfs -o soft,nosharecache,timeo=600,retrans=6,nfsvers=3
> ovc71.localdomain.local:/SHE_DOMAIN
> /rhev/data-center/mnt/ovc71.localdomain.local:_SHE__DOMAIN (cwd None)
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,199::hsm::2405::Storage.HSM::(__prefetchDomains) nfs local path:
> /rhev/data-center/mnt/ovc71.localdomain.local:_SHE__DOMAIN
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,201::hsm::2429::Storage.HSM::(__prefetchDomains) Found SD uuids:
> (u'9f1ec45d-0c32-4bfc-8b67-372d6f204fd1',)
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,202::hsm::2489::Storage.HSM::(connectStorageServer) knownSDs
> : {9f1ec45d-0c32-4bfc-8b67-372d6f204fd1: storage.nfsSD.findDomain,
> 3662a51f-39de-4533-97fe-d49bf98e2d43: storage.nfsSD.findDomain}
> Thread-1248::INFO::2015-10-28
> 14:59:00,202::logUtils::51::dispatcher::(wrapper) Run and protect:
> connectStorageServer, Return response: {'statuslist': [{'status': 0, 'id':
> u'189c29a5-6830-453c-aca3-7d82f2382dd8'}]}
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,202::task::1191::Storage.TaskManager.Task::(prepare)
> Task=`9ca908a0-45e2-41d5-802c-dc0bd2414a69`::finished: {'statuslist':
> [{'status': 0, 'id': u'189c29a5-6830-453c-aca3-7d82f2382dd8'}]}
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,202::task::595::Storage.TaskManager.Task::(_updateState)
> Task=`9ca908a0-45e2-41d5-802c-dc0bd2414a69`::moving from state preparing ->
> state finished
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,203::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources {}
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,203::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,203::task::993::Storage.TaskManager.Task::(_decref)
> Task=`9ca908a0-45e2-41d5-802c-dc0bd2414a69`::ref 0 aborting False
> Thread-1248::DEBUG::2015-10-28
> 14:59:00,203::__init__::533::jsonrpc.JsonRpcServer::(_serveRequest) Return
> 'StoragePool.connectStorageServer' in bridge with [{'status': 0, 'id':
> u'189c29a5-6830-453c-aca3-7d82f2382dd8'}]
> Thread-1249::DEBUG::2015-10-28
> 14:59:00,218::__init__::503::jsonrpc.JsonRpcServer::(_serveRequest) Calling
> 'StorageDomain.getInfo' in bridge with {u'storagedomainID':
> u'9f1ec45d-0c32-4bfc-8b67-372d6f204fd1'}
>
> Current filesystem layout on hypervisor, but still withou any qemu process
> for the hosted engine:
> [root at ovc71 log]# df -h
> Filesystem                           Size  Used Avail Use% Mounted on
> /dev/mapper/centos-root               27G  2.6G   24G  10% /
> devtmpfs                             4.9G     0  4.9G   0% /dev
> tmpfs                                4.9G  4.0K  4.9G   1% /dev/shm
> tmpfs                                4.9G  8.6M  4.9G   1% /run
> tmpfs                                4.9G     0  4.9G   0% /sys/fs/cgroup
> /dev/mapper/OVIRT_DOMAIN-NFS_DOMAIN   20G   36M   20G   1% /NFS_DOMAIN
> /dev/mapper/OVIRT_DOMAIN-SHE_DOMAIN   25G  2.9G   23G  12% /SHE_DOMAIN
> /dev/mapper/OVIRT_DOMAIN-ISO_DOMAIN  5.0G   33M  5.0G   1% /ISO_DOMAIN
> /dev/sda1                            497M  130M  368M  27% /boot
> ovc71.localdomain.local:/NFS_DOMAIN   20G   35M   20G   1%
> /rhev/data-center/mnt/ovc71.localdomain.local:_NFS__DOMAIN
> ovc71.localdomain.local:/SHE_DOMAIN   25G  2.9G   23G  12%
> /rhev/data-center/mnt/ovc71.localdomain.local:_SHE__DOMAIN
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20151028/55f9a93d/attachment-0001.html>


More information about the Users mailing list