Hi all,
For the records, I had to remove manually the conflicting directory and ts
respective gfid from the arbiter volume:
getfattr -m . -d -e hex e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
That gave me the gfid: 0x277c9caa9dce4a17a2a93775357befd5
Then cd .glusterfs/27/7c
rm -rf 277c9caa-9dce-4a17-a2a9-3775357befd5 (or move it out of there)
Triggerred heal: gluster volume heal engine
Then all ok:
gluster volume heal engine info
Brick gluster0:/gluster/engine/brick
Status: Connected
Number of entries: 0
Brick gluster1:/gluster/engine/brick
Status: Connected
Number of entries: 0
Brick gluster2:/gluster/engine/brick
Status: Connected
Number of entries: 0
Thanx.
On Fri, Jun 23, 2017 at 7:21 PM, Abi Askushi <rightkicktech(a)gmail.com>
wrote:
Hi Denis,
I receive permission denied as below:
gluster volume heal engine split-brain latest-mtime
/e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
Healing /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent failed:Operation
not permitted.
Volume heal failed.
When I shutdown host3 then no split brain is reported from the remaining
two hosts. When I power up host3 then I receive the mentioned split brain
and host3 logs the following at ovirt-hosted-engine-ha/agent.log
MainThread::INFO::2017-06-23 16:18:06,067::hosted_engine::
594::ovirt_hosted_engine_ha.agent.hosted_engine.
HostedEngine::(_initialize_broker) Failed set the storage domain: 'Failed
to set storage domain VdsmBackend, options {'hosted-engine.lockspace': '
7B22696D6167655F75756964223A202238323132626637382D663933332D
346465652D616333372D346265633734353035366235222C202270617468
223A206E756C6C2C2022766F6C756D655F75756964223A20223632373930
3162652D666261332D346263342D393037632D393931356138333632633537227D',
'sp_uuid': '00000000-0000-0000-0000-000000000000', 'dom_type':
'glusterfs', 'hosted-engine.metadata':
'7B22696D6167655F75756964223A20
2263353930633034372D613462322D346539312D613832362D6434386239
61643537323330222C202270617468223A206E756C6C2C2022766F6C756D
655F75756964223A202230353166653865612D333339632D346134302D38
3438382D386335313138666438373238227D', 'sd_uuid':
'e1c80750-b880-495e-9609-b8bc7760d101'}: Request failed: <type
'exceptions.OSError'>'. Waiting '5's before the next attempt
and the following at /var/log/messages:
Jun 23 16:19:43 v2 journal: vdsm root ERROR failed to retrieve Hosted
Engine HA info#012Traceback (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/vdsm/host/api.py", line 231, in
_getHaInfo#012 stats = instance.get_all_stats()#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py",
line 105, in get_all_stats#012 stats = broker.get_stats_from_storage(service)#012
File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
line 233, in get_stats_from_storage#012 result =
self._checked_communicate(request)#012 File "/usr/lib/python2.7/site-
packages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 261, in
_checked_communicate#012 .format(message or response))#012RequestError:
Request failed: failed to read metadata: [Errno 5] Input/output error:
'/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/e1c80750-b880-495e-
9609-b8bc7760d101/ha_agent/hosted-engine.metadata'
Thanx
On Fri, Jun 23, 2017 at 6:05 PM, Denis Chaplygin <dchaplyg(a)redhat.com>
wrote:
> Hello Abi,
>
> On Fri, Jun 23, 2017 at 4:47 PM, Abi Askushi <rightkicktech(a)gmail.com>
> wrote:
>
>> Hi All,
>>
>> I have a 3 node ovirt 4.1 setup. I lost one node due to raid controller
>> issues. Upon restoration I have the following split brain, although the
>> hosts have mounted the storage domains:
>>
>> gluster volume heal engine info split-brain
>> Brick gluster0:/gluster/engine/brick
>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>> Status: Connected
>> Number of entries in split-brain: 1
>>
>> Brick gluster1:/gluster/engine/brick
>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>> Status: Connected
>> Number of entries in split-brain: 1
>>
>> Brick gluster2:/gluster/engine/brick
>> /e1c80750-b880-495e-9609-b8bc7760d101/ha_agent
>> Status: Connected
>> Number of entries in split-brain: 1
>>
>>
>>
> It is definitely on gluster side. You could try to use
>
> gluster volume heal engine split-brain latest-mtime /e1c80750-b880-49
> 5e-9609-b8bc7760d101/ha_agent
>
>
> I also added gluster developers to that thread, so they may provide you
> with better advices.
>