[ovirt-users] Native Access on gluster storage domain
Stefano Danzi
s.danzi at hawai.it
Mon Sep 11 09:57:32 UTC 2017
Your suggestion solved the problem.
In the UI relative flag still missing, but now VMs are using gfapi.
Il 11/09/2017 05:23, Sahina Bose ha scritto:
> You could try to enable the config option for the 4.1 cluster level -
> using engine-config tool from the Hosted Engine VM. This will require
> a restart of the engine service and will enable gfapi access for all
> clusters at 4.1 level though - so try this option if this is acceptable.
>
> On Wed, Aug 30, 2017 at 8:02 PM, Stefano Danzi <s.danzi at hawai.it
> <mailto:s.danzi at hawai.it>> wrote:
>
> above the logs.
> PS cluster compatibility level is 4.1
>
> engine:
>
> 2017-08-30 16:26:07,928+02 INFO
> [org.ovirt.engine.core.bll.UpdateClusterCommand] (default task-8)
> [56d090c5-1097-4641-b745-74af8397d945] Lock Acquired to object
> 'EngineLock:{exclusiveLocks='[]', sharedLocks='[]'}'
> 2017-08-30 16:26:07,951+02 WARN
> [org.ovirt.engine.core.bll.UpdateClusterCommand] (default task-8)
> [56d090c5-1097-4641-b745-74af8397d945] Validation of action
> 'UpdateCluster' failed for user admin at internal. Reasons:
> VAR__TYPE__CLUSTER,VAR__ACTION__UPDATE,CLUSTER_CANNOT_UPDATE_SUPPORTED_FEATURES_WITH_LOWER_HOSTS
> 2017-08-30 16:26:07,952+02 INFO
> [org.ovirt.engine.core.bll.UpdateClusterCommand] (default task-8)
> [56d090c5-1097-4641-b745-74af8397d945] Lock freed to object
> 'EngineLock:{exclusiveLocks='[]', sharedLocks='[]'}'
>
> vdsm:
>
> 2017-08-30 16:29:23,310+0200 INFO (jsonrpc/0)
> [jsonrpc.JsonRpcServer] RPC call GlusterHost.list succeeded in
> 0.15 seconds (__init__:539)
> 2017-08-30 16:29:23,419+0200 INFO (jsonrpc/4)
> [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats succeeded in
> 0.01 seconds (__init__:539)
> 2017-08-30 16:29:23,424+0200 INFO (jsonrpc/3)
> [jsonrpc.JsonRpcServer] RPC call Host.getAllVmIoTunePolicies
> succeeded in 0.00 seconds (__init__:539)
> 2017-08-30 16:29:23,814+0200 INFO (jsonrpc/5)
> [jsonrpc.JsonRpcServer] RPC call GlusterHost.list succeeded in
> 0.15 seconds (__init__:539)
> 2017-08-30 16:29:24,011+0200 INFO (Reactor thread)
> [ProtocolDetector.AcceptorImpl] Accepted connection from ::1:51862
> (protocoldetector:72)
> 2017-08-30 16:29:24,023+0200 INFO (Reactor thread)
> [ProtocolDetector.Detector] Detected protocol stomp from ::1:51862
> (protocoldetector:127)
> 2017-08-30 16:29:24,024+0200 INFO (Reactor thread)
> [Broker.StompAdapter] Processing CONNECT request (stompreactor:103)
> 2017-08-30 16:29:24,031+0200 INFO (JsonRpc (StompReactor))
> [Broker.StompAdapter] Subscribe command received (stompreactor:130)
> 2017-08-30 16:29:24,287+0200 INFO (jsonrpc/2)
> [jsonrpc.JsonRpcServer] RPC call Host.getHardwareInfo succeeded in
> 0.01 seconds (__init__:539)
> 2017-08-30 16:29:24,443+0200 INFO (jsonrpc/7) [vdsm.api] START
> getSpmStatus(spUUID=u'00000002-0002-0002-0002-0000000001ef',
> options=None) from=::ffff:192.168.1.55,46502, flow_id=1f664a9,
> task_id=c856903a-0af1-4c0c-8a44-7971fee7dffa (api:46)
> 2017-08-30 16:29:24,446+0200 INFO (jsonrpc/7) [vdsm.api] FINISH
> getSpmStatus return={'spm_st': {'spmId': 1, 'spmStatus': 'SPM',
> 'spmLver': 1430L}} from=::ffff:192.168.1.55,46502,
> flow_id=1f664a9, task_id=c856903a-0af1-4c0c-8a44-7971fee7dffa (api:52)
> 2017-08-30 16:29:24,447+0200 INFO (jsonrpc/7)
> [jsonrpc.JsonRpcServer] RPC call StoragePool.getSpmStatus
> succeeded in 0.00 seconds (__init__:539)
> 2017-08-30 16:29:24,460+0200 INFO (jsonrpc/6)
> [jsonrpc.JsonRpcServer] RPC call GlusterHost.list succeeded in
> 0.16 seconds (__init__:539)
> 2017-08-30 16:29:24,467+0200 INFO (jsonrpc/1) [vdsm.api] START
> getStoragePoolInfo(spUUID=u'00000002-0002-0002-0002-0000000001ef',
> options=None) from=::ffff:192.168.1.55,46506, flow_id=1f664a9,
> task_id=029ec55e-9c47-4a20-be44-8c80fd1fd5ac (api:46)
>
>
> Il 30/08/2017 16:06, Shani Leviim ha scritto:
>> Hi Stefano,
>> Can you please attach your engine and vdsm logs?
>>
>> *Regards,
>> *
>> *Shani Leviim
>> *
>>
>> On Wed, Aug 30, 2017 at 12:46 PM, Stefano Danzi <s.danzi at hawai.it
>> <mailto:s.danzi at hawai.it>> wrote:
>>
>> Hello,
>> I have a test environment with a sigle host and self hosted
>> engine running oVirt Engine: 4.1.5.2-1.el7.centos
>>
>> I what to try the option "Native Access on gluster storage
>> domain" but I get an error because I have to put the
>> host in maintenance mode. I can't do that because I have a
>> single host so the hosted engine can't be migrated.
>>
>> There are a way to change this option but apply it at next
>> reboot?
>>
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org <mailto:Users at ovirt.org>
>> http://lists.ovirt.org/mailman/listinfo/users
>> <http://lists.ovirt.org/mailman/listinfo/users>
>>
>>
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org <mailto:Users at ovirt.org>
> http://lists.ovirt.org/mailman/listinfo/users
> <http://lists.ovirt.org/mailman/listinfo/users>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170911/c546ee3a/attachment.html>
More information about the Users
mailing list