[ovirt-users] [hosted-engine] Metadata too new error when adding 2nd host
Martin Sivak
msivak at redhat.com
Wed Apr 20 09:43:34 UTC 2016
> And we also do not clean on upgrades... Perhaps we can? Should? Optionally?
>
We can't. We do not execute any setup tool during upgrade and the
clean procedure
requires that all hosted engine tooling is shut down.
Martin
On Wed, Apr 20, 2016 at 11:40 AM, Yedidyah Bar David <didi at redhat.com> wrote:
> On Wed, Apr 20, 2016 at 11:40 AM, Martin Sivak <msivak at redhat.com> wrote:
>>> Doesn't cleaning sanlock lockspace require also to stop sanlock itself?
>>> I guess it's supposed to be able to handle this, but perhaps users want
>>> to clean the lockspace because dirt there causes also problems with
>>> sanlock, no?
>>
>> Sanlock can be up, but the lockspace has to be unused.
>>
>>> So the only tool we have to clean metadata is '–clean-metadata', which
>>> works one-by-one?
>>
>> Correct, it needs to acquire the lock first to make sure nobody is writing.
>>
>> The dirty disk issue should not be happening anymore, we added an
>> equivalent of the DD to hosted engine setup. But we might have a bug there
>> of course.
>
> And we also do not clean on upgrades... Perhaps we can? Should? Optionally?
>
>>
>> Martin
>>
>> On Wed, Apr 20, 2016 at 10:34 AM, Yedidyah Bar David <didi at redhat.com> wrote:
>>> On Wed, Apr 20, 2016 at 11:20 AM, Martin Sivak <msivak at redhat.com> wrote:
>>>>> after moving to global maintenance.
>>>>
>>>> Good point.
>>>>
>>>>> Martin - any advantage of this over '–reinitialize-lockspace'? Besides
>>>>> that it works also in older versions? Care to add this to the howto page?
>>>>
>>>> Reinitialize lockspace clears the sanlock lockspace, not the metadata
>>>> file. Those are two different places.
>>>
>>> So the only tool we have to clean metadata is '–clean-metadata', which
>>> works one-by-one?
>>>
>>> Doesn't cleaning sanlock lockspace require also to stop sanlock itself?
>>> I guess it's supposed to be able to handle this, but perhaps users want
>>> to clean the lockspace because dirt there causes also problems with
>>> sanlock, no?
>>>
>>>>
>>>>> Care to add this to the howto page?
>>>>
>>>> Yeah, I can do that.
>>>
>>> Thanks!
>>>
>>>>
>>>> Martin
>>>>
>>>> On Wed, Apr 20, 2016 at 10:17 AM, Yedidyah Bar David <didi at redhat.com> wrote:
>>>>> On Wed, Apr 20, 2016 at 11:11 AM, Martin Sivak <msivak at redhat.com> wrote:
>>>>>>> Assuming you never deployed a host with ID 52, this is likely a result of a
>>>>>>> corruption or dirt or something like that.
>>>>>>
>>>>>>> I see that you use FC storage. In previous versions, we did not clean such
>>>>>>> storage, so you might have dirt left.
>>>>>>
>>>>>> This is the exact reason for an error like yours. Using dirty block
>>>>>> storage. Please stop all hosted engine tooling (both agent and broker)
>>>>>> and fill the metadata drive with zeros.
>>>>>
>>>>> after moving to global maintenance.
>>>>>
>>>>> Martin - any advantage of this over '–reinitialize-lockspace'? Besides
>>>>> that it works also in older versions? Care to add this to the howto page?
>>>>> Thanks!
>>>>>
>>>>>>
>>>>>> You will have to find the proper hosted-engine.metadata file (which
>>>>>> will be a symlink) under /rhev:
>>>>>>
>>>>>> Example:
>>>>>>
>>>>>> [root at dev-03 rhev]# find . -name hosted-engine.metadata
>>>>>>
>>>>>> ./data-center/mnt/str-01.rhev.lab.eng.brq.redhat.com:_mnt_export_nfs_lv2_msivak/868a1a4e-9f94-42f5-af23-8f884b3c53d5/ha_agent/hosted-engine.metadata
>>>>>>
>>>>>> [root at dev-03 rhev]# ls -al
>>>>>> ./data-center/mnt/str-01:_mnt_export_nfs_lv2_msivak/868a1a4e-9f94-42f5-af23-8f884b3c53d5/ha_agent/hosted-engine.metadata
>>>>>>
>>>>>> lrwxrwxrwx. 1 vdsm kvm 201 Mar 15 15:00
>>>>>> ./data-center/mnt/str-01:_mnt_export_nfs_lv2_msivak/868a1a4e-9f94-42f5-af23-8f884b3c53d5/ha_agent/hosted-engine.metadata
>>>>>> -> /rhev/data-center/mnt/str-01:_mnt_export_nfs_lv2_msivak/868a1a4e-9f94-42f5-af23-8f884b3c53d5/images/6ab3f215-f234-4cd4-b9d4-8680767c3d99/dcbfa48d-8543-42d1-93dc-aa40855c4855
>>>>>>
>>>>>> And use (for example) dd if=/dev/zero of=/path/to/metadata bs=1M to
>>>>>> clean it - But be CAREFUL to not touch any other file or disk you
>>>>>> might find.
>>>>>>
>>>>>> Then restart the hosted engine tools and all should be fine.
>>>>>>
>>>>>>
>>>>>>
>>>>>> Martin
>>>>>>
>>>>>>
>>>>>> On Wed, Apr 20, 2016 at 8:20 AM, Yedidyah Bar David <didi at redhat.com> wrote:
>>>>>>> On Wed, Apr 20, 2016 at 7:15 AM, Wee Sritippho <wee.s at forest.go.th> wrote:
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> I used CentOS-7-x86_64-Minimal-1511.iso to install the hosts and the engine.
>>>>>>>>
>>>>>>>> The 1st host and the hosted-engine were installed successfully, but the 2nd
>>>>>>>> host failed with this error message:
>>>>>>>>
>>>>>>>> "Failed to execute stage 'Setup validation': Metadata version 2 from host 52
>>>>>>>> too new for this agent (highest compatible version: 1)"
>>>>>>>
>>>>>>> Assuming you never deployed a host with ID 52, this is likely a result of a
>>>>>>> corruption or dirt or something like that.
>>>>>>>
>>>>>>> What do you get on host 1 running 'hosted-engine --vm-status'?
>>>>>>>
>>>>>>> I see that you use FC storage. In previous versions, we did not clean such
>>>>>>> storage, so you might have dirt left. See also [1]. You can try cleaning
>>>>>>> using [2].
>>>>>>>
>>>>>>> [1] https://bugzilla.redhat.com/show_bug.cgi?id=1238823
>>>>>>> [2] https://www.ovirt.org/documentation/how-to/hosted-engine/#lockspace-corrupted-recovery-procedure
>>>>>>>
>>>>>>>>
>>>>>>>> Here is the package versions:
>>>>>>>>
>>>>>>>> [root at host02 ~]# rpm -qa | grep ovirt
>>>>>>>> libgovirt-0.3.3-1.el7_2.1.x86_64
>>>>>>>> ovirt-vmconsole-1.0.0-1.el7.centos.noarch
>>>>>>>> ovirt-vmconsole-host-1.0.0-1.el7.centos.noarch
>>>>>>>> ovirt-host-deploy-1.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-hosted-engine-ha-1.3.5.1-1.el7.centos.noarch
>>>>>>>> ovirt-hosted-engine-setup-1.3.4.0-1.el7.centos.noarch
>>>>>>>> ovirt-release36-007-1.noarch
>>>>>>>> ovirt-engine-sdk-python-3.6.3.0-1.el7.centos.noarch
>>>>>>>> ovirt-setup-lib-1.0.1-1.el7.centos.noarch
>>>>>>>>
>>>>>>>> [root at engine ~]# rpm -qa | grep ovirt
>>>>>>>> ovirt-engine-setup-base-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-setup-plugin-ovirt-engine-common-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-vmconsole-proxy-1.0.0-1.el7.centos.noarch
>>>>>>>> ovirt-engine-tools-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-vmconsole-proxy-helper-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-host-deploy-1.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-release36-007-1.noarch
>>>>>>>> ovirt-engine-sdk-python-3.6.3.0-1.el7.centos.noarch
>>>>>>>> ovirt-iso-uploader-3.6.0-1.el7.centos.noarch
>>>>>>>> ovirt-engine-extensions-api-impl-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-setup-lib-1.0.1-1.el7.centos.noarch
>>>>>>>> ovirt-host-deploy-java-1.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-cli-3.6.2.0-1.el7.centos.noarch
>>>>>>>> ovirt-engine-setup-plugin-websocket-proxy-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-vmconsole-1.0.0-1.el7.centos.noarch
>>>>>>>> ovirt-engine-backend-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-dbscripts-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-webadmin-portal-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-setup-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-setup-plugin-vmconsole-proxy-helper-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-guest-agent-common-1.0.11-1.el7.noarch
>>>>>>>> ovirt-engine-wildfly-8.2.1-1.el7.x86_64
>>>>>>>> ovirt-engine-wildfly-overlay-8.0.5-1.el7.noarch
>>>>>>>> ovirt-engine-websocket-proxy-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-restapi-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-userportal-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-engine-setup-plugin-ovirt-engine-3.6.4.1-1.el7.centos.noarch
>>>>>>>> ovirt-image-uploader-3.6.0-1.el7.centos.noarch
>>>>>>>> ovirt-engine-extension-aaa-jdbc-1.0.6-1.el7.noarch
>>>>>>>> ovirt-engine-lib-3.6.4.1-1.el7.centos.noarch
>>>>>>>>
>>>>>>>>
>>>>>>>> Here are the log files:
>>>>>>>> https://gist.github.com/weeix/1743f88d3afe1f405889a67ed4011141
>>>>>>>>
>>>>>>>> --
>>>>>>>> Wee
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list
>>>>>>>> Users at ovirt.org
>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Didi
>>>>>>> _______________________________________________
>>>>>>> Users mailing list
>>>>>>> Users at ovirt.org
>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Didi
>>>
>>>
>>>
>>> --
>>> Didi
>
>
>
> --
> Didi
More information about the Users
mailing list