Hi,
I have to update the engine again?
Thanks,
R
Tibor
----- 2018. máj.. 18., 6:47, Sahina Bose <sabose(a)redhat.com> írta:
On Thu, May 17, 2018 at 10:12 PM, Demeter Tibor < [
mailto:tdemeter@itsmart.hu |
tdemeter(a)itsmart.hu ] > wrote:
> Hi,
> 4.2.4-0.0.master.20180515183442.git00e1340.el7.centos
> Firstly, I did a yum update "ovirt-*-setup*"
> second, I have ran engine-setup to upgrade.
> I didn't remove the old repos, just installed the nightly
repo.
> Thank you again,
> Regards,
> Tibor
> ----- 2018. máj.. 17., 15:02, Sahina Bose < [
mailto:sabose@redhat.com |
> sabose(a)redhat.com ] > írta:
>> It doesn't look like the patch was applied. Still see the
same error in
>> engine.log
>> "Error while refreshing brick statuses for volume 'volume1' of
cluster 'C6220':
>> null"\
>> Did you use engine-setup to upgrade? What's the version
of ovirt-engine
>> currently installed?
>> On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor < [
mailto:tdemeter@itsmart.hu |
>> tdemeter(a)itsmart.hu ] > wrote:
>>
> Hi,
>>> sure,
>>> Thank you for your time!
>>>> R
>>
> Tibor
>>> ----- 2018. máj.. 17., 12:19, Sahina Bose < [
mailto:sabose@redhat.com |
>>> sabose(a)redhat.com ] > írta:
>>>> [+users]
>>>> Can you provide the engine.log to see why the
monitoring is not working here.
>>>> thanks!
>>>> On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor < [
mailto:tdemeter@itsmart.hu |
>>>> tdemeter(a)itsmart.hu ] > wrote:
>>>>
> Hi,
>>>>> Meanwhile, I did the upgrade engine, but the
gluster state is same on my first
>>>>> node.
>>>>> I've attached some screenshot of my problem.
>>>>> Thanks
>>>>
> Tibor
>>>>> ----- 2018. máj.. 16., 10:16, Demeter Tibor <
[ mailto:tdemeter@itsmart.hu |
>>>>> tdemeter(a)itsmart.hu ] > írta Hi,
>>>>>> If 4.3.4 will release, i just have to remove
the nightly repo and update to
>>>>>> stable?
>>>>>> I'm sorry for my terrible English, I try
to explain what was my problem with
>>>>>> update.
>>>>>> I'm upgraded from 4.1.8.
>>>>>> Maybe it need to update, because I had a lot
of question under upgrade and I was
>>>>>> not sure in all of necessary steps. For example, If I need to
installing the
>>>>>> new, 4.2 repo on the hosts, then need to remove the old repo from
that?
>>>>>> Why I need to do a" yum update -y" on hosts, meanwhile
there is an "Updatehost"
>>>>>> menu in the GUI? So, maybe it outdated.
>>>>>> Since upgrade hosted engine, and the first node, I have problems
with gluster.
>>>>>> It seems to working fine if you check it from console
"gluster volume status,
>>>>>> etc" but not on the Gui, because now it yellow, and the
brick reds in the first
>>>>>> node.
>>>>>> Previously I did a mistake with glusterfs, my
gluster config was wrong. I have
>>>>>> corrected them, but it did not helped to me,gluster bricks are
reds on my first
>>>>>> node yet....
>>>>>> Now I try to upgrade to nightly, but I'm
affraid, because it a living,
>>>>>> productive system, and I don't have downtime. I hope it will
help me.
>>>>>> Thanks for all,
>>>>>
> Regards,
>>>>>>> Tibor Demeter
>>>>>> ----- 2018. máj.. 16., 9:58, Sahina Bose <
[ mailto:sabose@redhat.com |
>>>>>> sabose(a)redhat.com ] > írta:
>>>>>>> On Wed, May 16, 2018 at 1:19 PM, Demeter
Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>> tdemeter(a)itsmart.hu ] > wrote:
>>>>>>>
> Hi,
>>>>>>>> is it a different, unstable repo? I
have a productive cluster, how is safe that?
>>>>>>>> I don't have any experience with nightly build. How
can I use this? It have to
>>>>>>>> install to the engine VM or all of my hosts?
>>>>>>>> Thanks in advance for help me..
>>>>>>> Only on the engine VM.
>>>>>>> Regarding stability - it passes CI so
relatively stable, beyond that there are
>>>>>>> no guarantees.
>>>>>>> What's the specific problem
you're facing with update? Can you elaborate?
>>>>>>>
> Regards,
>>>>>>>
> Tibor
>>>>>>>> ----- 2018. máj.. 15., 9:58, Demeter
Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>>> tdemeter(a)itsmart.hu ] > írta:
>>>>>>>>
> Hi,
>>>>>>>>> Could you explain how can I use
this patch?
>>>>>>>>>> R,
>>>>>>>>
> Tibor
>>>>>>>>> ----- 2018. máj.. 14., 11:18,
Demeter Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>>>> tdemeter(a)itsmart.hu ] > írta:
>>>>>>>>>
> Hi,
>>>>>>>>>> Sorry for my question, but
can you tell me please how can I use this patch?
>>>>>>>>>>> Thanks,
>>>>>>>>>
> Regards,
>>>>>>>>>
> Tibor
>>>>>>>>>>> ----- 2018. máj.. 14., 10:47, Sahina
Bose < [ mailto:sabose@redhat.com |
>>>>>>>>>>> sabose(a)redhat.com ] > írta:
>>>>>>>>>>> On Sat, May 12, 2018 at
1:14 PM, Demeter Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>>>>>> tdemeter(a)itsmart.hu ] > wrote:
>>>>>>>>>>>
> Hi,
>>>>>>>>>>>> Could someone help me
please ? I can't finish my upgrade process.
>>>>>>>>>>> [
https://gerrit.ovirt.org/91164 |
https://gerrit.ovirt.org/91164 ] should fix
>>>>>>>>>>> the error you're facing.
>>>>>>>>>>> Can you elaborate why
this is affecting the upgrade process?
>>>>>>>
>>>>> Thanks
>>>>>>>>>>>>> R
>>>>>>>>>>>
> Tibor
>>>>>>>>>>>> ----- 2018. máj..
10., 12:51, Demeter Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>>>>>>> tdemeter(a)itsmart.hu ] > írta:
>>>>>>>>>>>>
> Hi,
>>>>>>>>>>>>> I've attached
the vdsm and supervdsm logs. But I don't have engine.log here,
>>>>>>>>>>>>> because that is on hosted engine vm.
Should I send that ?
>>>>>>>>>>>>> Thank you
>>>>>>>>>>>>
> Regards,
>>>>>>>>>>>>
> Tibor
>>>>>>>>>>>>>> ----- 2018. máj.. 10.,
12:30, Sahina Bose < [ mailto:sabose@redhat.com |
>>>>>>>>>>>>>> sabose(a)redhat.com ] > írta:
>>>>>>>>>>>>>> There's a
bug here. Can you log one attaching this engine.log and also vdsm.log
>>>>>>>>>>>>>> & supervdsm.log from
n3.itsmart.cloud
>>>>>>>>>>>>>> On Thu, May
10, 2018 at 3:35 PM, Demeter Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>>>>>>>>> tdemeter(a)itsmart.hu ] > wrote:
>>>>>>>>>>>>>>
> Hi,
>>>>>>>>>>>>>>> I found
this:
>>>>>>>>>>>>>>>
2018-05-10 03:24:19,096+02 INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] FINISH,
>>>>>>>>>>>>>>>
GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>>>
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e,
>>>>>>>>>>>>>>> log id: 347435ae
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7)
>>>>>>>>>>>>>>> [43f4eaec] Error while
refreshing brick statuses for volume 'volume2' of
>>>>>>>>>>>>>>> cluster 'C6220':
null
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler8)
>>>>>>>>>>>>>>> [7715ceda] Failed to acquire
lock and wait lock
>>>>>>>>>>>>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,104+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] START,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>>>>>>>>>>>>> log id: 6908121d
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] Command
>>>>>>>>>>>>>>>
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] FINISH,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,107+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] START,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}),
>>>>>>>>>>>>>>> log id: 735c6a5f
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] Command
>>>>>>>>>>>>>>>
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] FINISH,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,110+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] START,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>> log id: 6f9e9f58
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] Command
>>>>>>>>>>>>>>>
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] FINISH,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,113+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] START,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>>>>>>>>>>>>> log id: 2ee46967
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,115+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] Command
>>>>>>>>>>>>>>>
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,116+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] FINISH,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967
>>>>>>>>>>>>>>> 2018-05-10 03:24:19,117+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] START,
>>>>>>>>>>>>>>>
GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n1.itsmart.cloud,
>>>>>>>>>>>>>>>
GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
>>>>>>>>>>>>>>>
volumeName='volume1'}), log id: 7550e5c
>>>>>>>>>>>>>>> 2018-05-10 03:24:20,748+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler7)
[43f4eaec] FINISH,
>>>>>>>>>>>>>>>
GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>>>
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f,
>>>>>>>>>>>>>>> log id: 7550e5c
>>>>>>>>>>>>>>> 2018-05-10 03:24:20,749+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7)
>>>>>>>>>>>>>>> [43f4eaec] Error while
refreshing brick statuses for volume 'volume1' of
>>>>>>>>>>>>>>> cluster 'C6220':
null
>>>>>>>>>>>>>>> 2018-05-10 03:24:20,750+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] START,
>>>>>>>>>>>>>>>
GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>> log id: 120cc68d
>>>>>>>>>>>>>>> 2018-05-10 03:24:20,930+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] FINISH, GlusterServersListVDSCommand,
>>>>>>>>>>>>>>> return: [ [
http://10.101.0.2/24:CONNECTED | 10.101.0.2/24:CONNECTED ] ,
>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED,
10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log
>>>>>>>>>>>>>>> id: 120cc68d
>>>>>>>>>>>>>>> 2018-05-10 03:24:20,949+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] START,
>>>>>>>>>>>>>>>
GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>
GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>> log id: 118aa264
>>>>>>>>>>>>>>> 2018-05-10 03:24:21,048+02
WARN
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] Could not associate brick
>>>>>>>>>>>>>>>
'10.104.0.1:/gluster/brick/brick1' of volume
>>>>>>>>>>>>>>>
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>>>>>>>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>> 2018-05-10 03:24:21,055+02
WARN
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] Could not associate brick
>>>>>>>>>>>>>>>
'10.104.0.1:/gluster/brick/brick2' of volume
>>>>>>>>>>>>>>>
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>>>>>>>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>> 2018-05-10 03:24:21,061+02
WARN
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] Could not associate brick
>>>>>>>>>>>>>>>
'10.104.0.1:/gluster/brick/brick3' of volume
>>>>>>>>>>>>>>>
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>>>>>>>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>> 2018-05-10 03:24:21,067+02
WARN
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] Could not associate brick
>>>>>>>>>>>>>>>
'10.104.0.1:/gluster2/brick/brick1' of volume
>>>>>>>>>>>>>>>
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>>>>>>>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>> 2018-05-10 03:24:21,074+02
WARN
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] Could not associate brick
>>>>>>>>>>>>>>>
'10.104.0.1:/gluster2/brick/brick2' of volume
>>>>>>>>>>>>>>>
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>>>>>>>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>> 2018-05-10 03:24:21,080+02
WARN
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] Could not associate brick
>>>>>>>>>>>>>>>
'10.104.0.1:/gluster2/brick/brick3' of volume
>>>>>>>>>>>>>>>
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>>>>>>>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>> 2018-05-10 03:24:21,081+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler8)
[7715ceda] FINISH, GlusterVolumesListVDSCommand,
>>>>>>>>>>>>>>> return:
>>>>>>>>>>>>>>>
{68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>>>>>>>>>>
e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g
>>>>>>>>>>>>>>>
luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264
>>>>>>>>>>>>>>>
2018-05-10 11:59:26,047+02 ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] Command
>>>>>>>>>>>>>>>
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] FINISH,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,048+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] START,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>>>>>>>>>>>>> log id: 28d9e255
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] Command
>>>>>>>>>>>>>>>
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] FINISH,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,052+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] START,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>> log id: 4a7b280e
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] Command
>>>>>>>>>>>>>>>
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] FINISH,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,055+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] START,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>>>>>>>>>>>>> log id: 18adc534
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] Command
>>>>>>>>>>>>>>>
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>>>>>>>>>>>>> execution failed: null
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] FINISH,
>>>>>>>>>>>>>>>
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534
>>>>>>>>>>>>>>> 2018-05-10 11:59:26,058+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] START,
>>>>>>>>>>>>>>>
GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n3.itsmart.cloud,
>>>>>>>>>>>>>>>
GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
>>>>>>>>>>>>>>>
volumeName='volume1'}), log id: 3451084f
>>>>>>>>>>>>>>> 2018-05-10 11:59:28,050+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire
lock and wait lock
>>>>>>>>>>>>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>> 2018-05-10 11:59:28,060+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire
lock and wait lock
>>>>>>>>>>>>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>> 2018-05-10 11:59:28,062+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire
lock and wait lock
>>>>>>>>>>>>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire
lock and wait lock
>>>>>>>>>>>>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire
lock and wait lock
>>>>>>>>>>>>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>> 2018-05-10 11:59:31,062+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire
lock and wait lock
>>>>>>>>>>>>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>> 2018-05-10 11:59:31,064+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire
lock and wait lock
>>>>>>>>>>>>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>>>>>>>>>>> sharedLocks=''}'
>>>>>>>>>>>>>>> 2018-05-10 11:59:31,465+02
INFO
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>>>>>>>>>>> (DefaultQuartzScheduler4)
[400fa486] FINISH,
>>>>>>>>>>>>>>>
GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>>>>>>>>>>>>
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43,
>>>>>>>>>>>>>>> log id: 3451084f
>>>>>>>>>>>>>>> 2018-05-10 11:59:31,466+02
ERROR
>>>>>>>>>>>>>>>
[org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler4)
>>>>>>>>>>>>>>> [400fa486] Error while
refreshing brick statuses for volume 'volume1' of
>>>>>>>>>>>>>>> cluster 'C6220': null
>>>>>>>>>>>>>>>> R
>>>>>>>>>>>>>>
> Tibor
>>>>>>>>>>>>>>> -----
2018. máj.. 10., 11:43, Sahina Bose < [ mailto:sabose@redhat.com |
>>>>>>>>>>>>>>> sabose(a)redhat.com ] >
írta:
>>>>>>>>>>>>>>>> This
doesn't affect the monitoring of state.
>>>>>>>>>>>>>>>> Any errors in vdsm.log?
>>>>>>>>>>>>>>>> Or errors in engine.log
of the form "Error while refreshing brick statuses for
>>>>>>>>>>>>>>>> volume"
>>>>>>>>>>>>>>>> On
Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>>>>>>>>>>> tdemeter(a)itsmart.hu ]
> wrote:
>>>>>>>>>>>>>>>>
> Hi,
>>>>>>>>>>>>>>>>>
Thank you for your fast reply :)
>>>>>>>>>>>>>>>>>
2018-05-10 11:01:51,574+02 INFO
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>>>>>>>>>>>>>
GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>>> log id: 39adbbb8
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,768+02 INFO
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand,
>>>>>>>>>>>>>>>>> return: [ [
http://10.101.0.2/24:CONNECTED | 10.101.0.2/24:CONNECTED ] ,
>>>>>>>>>>>>>>>>>
n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log
>>>>>>>>>>>>>>>>> id: 39adbbb8
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,788+02 INFO
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>>>>>>>>>>>>>
GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>>>>>>>>>>>>
GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>>>>>>>>>>>> log id: 738a7261
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,892+02 WARN
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>>>
'10.104.0.1:/gluster/brick/brick1' of volume
>>>>>>>>>>>>>>>>>
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>>>>>>>>>>>>>> network found in
cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,898+02 WARN
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>>>
'10.104.0.1:/gluster/brick/brick2' of volume
>>>>>>>>>>>>>>>>>
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>>>>>>>>>>>>>> network found in
cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,905+02 WARN
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>>>
'10.104.0.1:/gluster/brick/brick3' of volume
>>>>>>>>>>>>>>>>>
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>>>>>>>>>>>>>> network found in
cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,911+02 WARN
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>>>
'10.104.0.1:/gluster2/brick/brick1' of volume
>>>>>>>>>>>>>>>>>
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>>>>>>>>>>>>>> network found in
cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,917+02 WARN
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>>>
'10.104.0.1:/gluster2/brick/brick2' of volume
>>>>>>>>>>>>>>>>>
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>>>>>>>>>>>>>> network found in
cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,924+02 WARN
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>>>>>>>>>>>>
'10.104.0.1:/gluster2/brick/brick3' of volume
>>>>>>>>>>>>>>>>>
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>>>>>>>>>>>>>> network found in
cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>>>>>>>>>>>> 2018-05-10
11:01:51,925+02 INFO
>>>>>>>>>>>>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>>>>>>>>>>>>
(DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand,
>>>>>>>>>>>>>>>>> return:
>>>>>>>>>>>>>>>>>
{68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>>>>>>>>>>>>
e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>>>>>>>>>>>>>>>>> log id: 738a7261
>>>>>>>>>>>>>>>>>
This happening continuously.
>>>>>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>>>
> Tibor
>>>>>>>>>>>>>>>>>
----- 2018. máj.. 10., 10:56, Sahina Bose < [ mailto:sabose@redhat.com |
>>>>>>>>>>>>>>>>> sabose(a)redhat.com ]
> írta:
>>>>>>>>>>>>>>>>>>
Could you check the engine.log if there are errors related to getting
>>>>>>>>>>>>>>>>>>
GlusterVolumeAdvancedDetails ?
>>>>>>>>>>>>>>>>>>
On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>>>>>>>>>>>>>
tdemeter(a)itsmart.hu ] > wrote:
>>>>>>>>>>>>>>>>>>>
Dear Ovirt Users,
>>>>>>>>>>>>>>>>>>> I've
followed up the self-hosted-engine upgrade documentation, I upgraded my 4.1
>>>>>>>>>>>>>>>>>>> system to
4.2.3.
>>>>>>>>>>>>>>>>>>> I upgaded the
first node with yum upgrade, it seems working now fine. But since
>>>>>>>>>>>>>>>>>>> upgrade, the
gluster informations seems to displayed incorrect on the admin
>>>>>>>>>>>>>>>>>>> panel. The
volume yellow, and there are red bricks from that node.
>>>>>>>>>>>>>>>>>>> I've
checked in console, I think my gluster is not degraded:
>>>>>>>>>>>>>>>>>>>
root@n1 ~]# gluster volume list
>>>>>>>>>>>>>>>>>>> volume1
>>>>>>>>>>>>>>>>>>> volume2
>>>>>>>>>>>>>>>>>>> [root@n1 ~]#
gluster volume info
>>>>>>>>>>>>>>>>>>> Volume Name:
volume1
>>>>>>>>>>>>>>>>>>> Type:
Distributed-Replicate
>>>>>>>>>>>>>>>>>>> Volume ID:
e0f568fa-987c-4f5c-b853-01bce718ee27
>>>>>>>>>>>>>>>>>>> Status:
Started
>>>>>>>>>>>>>>>>>>> Snapshot
Count: 0
>>>>>>>>>>>>>>>>>>> Number of
Bricks: 3 x 3 = 9
>>>>>>>>>>>>>>>>>>>
Transport-type: tcp
>>>>>>>>>>>>>>>>>>> Bricks:
>>>>>>>>>>>>>>>>>>> Brick1:
10.104.0.1:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>>> Brick2:
10.104.0.2:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>>> Brick3:
10.104.0.3:/gluster/brick/brick1
>>>>>>>>>>>>>>>>>>> Brick4:
10.104.0.1:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>>> Brick5:
10.104.0.2:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>>> Brick6:
10.104.0.3:/gluster/brick/brick2
>>>>>>>>>>>>>>>>>>> Brick7:
10.104.0.1:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>>> Brick8:
10.104.0.2:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>>> Brick9:
10.104.0.3:/gluster/brick/brick3
>>>>>>>>>>>>>>>>>>> Options
Reconfigured:
>>>>>>>>>>>>>>>>>>>
transport.address-family: inet
>>>>>>>>>>>>>>>>>>>
performance.readdir-ahead: on
>>>>>>>>>>>>>>>>>>> nfs.disable:
on
>>>>>>>>>>>>>>>>>>>
storage.owner-uid: 36
>>>>>>>>>>>>>>>>>>>
storage.owner-gid: 36
>>>>>>>>>>>>>>>>>>>
performance.quick-read: off
>>>>>>>>>>>>>>>>>>>
performance.read-ahead: off
>>>>>>>>>>>>>>>>>>>
performance.io-cache: off
>>>>>>>>>>>>>>>>>>>
performance.stat-prefetch: off
>>>>>>>>>>>>>>>>>>>
performance.low-prio-threads: 32
>>>>>>>>>>>>>>>>>>>
network.remote-dio: enable
>>>>>>>>>>>>>>>>>>>
cluster.eager-lock: enable
>>>>>>>>>>>>>>>>>>>
cluster.quorum-type: auto
>>>>>>>>>>>>>>>>>>>
cluster.server-quorum-type: server
>>>>>>>>>>>>>>>>>>>
cluster.data-self-heal-algorithm: full
>>>>>>>>>>>>>>>>>>>
cluster.locking-scheme: granular
>>>>>>>>>>>>>>>>>>>
cluster.shd-max-threads: 8
>>>>>>>>>>>>>>>>>>>
cluster.shd-wait-qlength: 10000
>>>>>>>>>>>>>>>>>>>
features.shard: on
>>>>>>>>>>>>>>>>>>> user.cifs:
off
>>>>>>>>>>>>>>>>>>>
server.allow-insecure: on
>>>>>>>>>>>>>>>>>>> Volume Name:
volume2
>>>>>>>>>>>>>>>>>>> Type:
Distributed-Replicate
>>>>>>>>>>>>>>>>>>> Volume ID:
68cfb061-1320-4042-abcd-9228da23c0c8
>>>>>>>>>>>>>>>>>>> Status:
Started
>>>>>>>>>>>>>>>>>>> Snapshot
Count: 0
>>>>>>>>>>>>>>>>>>> Number of
Bricks: 3 x 3 = 9
>>>>>>>>>>>>>>>>>>>
Transport-type: tcp
>>>>>>>>>>>>>>>>>>> Bricks:
>>>>>>>>>>>>>>>>>>> Brick1:
10.104.0.1:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>>> Brick2:
10.104.0.2:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>>> Brick3:
10.104.0.3:/gluster2/brick/brick1
>>>>>>>>>>>>>>>>>>> Brick4:
10.104.0.1:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>>> Brick5:
10.104.0.2:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>>> Brick6:
10.104.0.3:/gluster2/brick/brick2
>>>>>>>>>>>>>>>>>>> Brick7:
10.104.0.1:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>>> Brick8:
10.104.0.2:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>>> Brick9:
10.104.0.3:/gluster2/brick/brick3
>>>>>>>>>>>>>>>>>>> Options
Reconfigured:
>>>>>>>>>>>>>>>>>>> nfs.disable:
on
>>>>>>>>>>>>>>>>>>>
performance.readdir-ahead: on
>>>>>>>>>>>>>>>>>>>
transport.address-family: inet
>>>>>>>>>>>>>>>>>>>
cluster.quorum-type: auto
>>>>>>>>>>>>>>>>>>>
network.ping-timeout: 10
>>>>>>>>>>>>>>>>>>> auth.allow:
*
>>>>>>>>>>>>>>>>>>>
performance.quick-read: off
>>>>>>>>>>>>>>>>>>>
performance.read-ahead: off
>>>>>>>>>>>>>>>>>>>
performance.io-cache: off
>>>>>>>>>>>>>>>>>>>
performance.stat-prefetch: off
>>>>>>>>>>>>>>>>>>>
performance.low-prio-threads: 32
>>>>>>>>>>>>>>>>>>>
network.remote-dio: enable
>>>>>>>>>>>>>>>>>>>
cluster.eager-lock: enable
>>>>>>>>>>>>>>>>>>>
cluster.server-quorum-type: server
>>>>>>>>>>>>>>>>>>>
cluster.data-self-heal-algorithm: full
>>>>>>>>>>>>>>>>>>>
cluster.locking-scheme: granular
>>>>>>>>>>>>>>>>>>>
cluster.shd-max-threads: 8
>>>>>>>>>>>>>>>>>>>
cluster.shd-wait-qlength: 10000
>>>>>>>>>>>>>>>>>>>
features.shard: on
>>>>>>>>>>>>>>>>>>> user.cifs:
off
>>>>>>>>>>>>>>>>>>>
storage.owner-uid: 36
>>>>>>>>>>>>>>>>>>>
storage.owner-gid: 36
>>>>>>>>>>>>>>>>>>>
server.allow-insecure: on
>>>>>>>>>>>>>>>>>>> [root@n1 ~]#
gluster volume status
>>>>>>>>>>>>>>>>>>> Status of
volume: volume1
>>>>>>>>>>>>>>>>>>> Gluster
process TCP Port RDMA Port Online Pid
>>>>>>>>>>>>>>>>>>>
------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520
>>>>>>>>>>>>>>>>>>> Self-heal
Daemon on localhost N/A N/A Y 54356
>>>>>>>>>>>>>>>>>>> Self-heal
Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>>>>>>>>>>>>> Self-heal
Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>>>>>>>>>>>>> Self-heal
Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>>>>>>>>>>>>> Task Status
of Volume volume1
>>>>>>>>>>>>>>>>>>>
------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>> There are no
active volume tasks
>>>>>>>>>>>>>>>>>>> Status of
volume: volume2
>>>>>>>>>>>>>>>>>>> Gluster
process TCP Port RDMA Port Online Pid
>>>>>>>>>>>>>>>>>>>
------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968
>>>>>>>>>>>>>>>>>>> Brick
10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541
>>>>>>>>>>>>>>>>>>> Self-heal
Daemon on localhost N/A N/A Y 54356
>>>>>>>>>>>>>>>>>>> Self-heal
Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>>>>>>>>>>>>> Self-heal
Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>>>>>>>>>>>>> Self-heal
Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>>>>>>>>>>>>> Task Status
of Volume volume2
>>>>>>>>>>>>>>>>>>>
------------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>> There are no
active volume tasks
>>>>>>>>>>>>>>>>>>> I think ovirt
can't read valid informations about gluster.
>>>>>>>>>>>>>>>>>>> I can't
contiune upgrade of other hosts until this problem exist.
>>>>>>>>>>>>>>>>>>>
Please help me:)
>>>>>>>>>>>>>>
>>>>> Thanks
>>>>>>>>>>>>>>>>>>
> Regards,
>>>>>>>>>>>>>>>>>>
> Tibor
>>>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>>>> Users mailing
list -- [ mailto:users@ovirt.org | users(a)ovirt.org ]
>>>>>>>>>>>>>>>>>>> To
unsubscribe send an email to [ mailto:users-leave@ovirt.org |
>>>>>>>>>>>>>>>>>>>
users-leave(a)ovirt.org ]
>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>> Users mailing list -- [
mailto:users@ovirt.org | users(a)ovirt.org ]
>>>>>>>>>>>>> To unsubscribe send an email to [
mailto:users-leave@ovirt.org |
>>>>>>>>>>>>> users-leave(a)ovirt.org ]
>>>>>>>>>>
_______________________________________________
>>>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org |
users(a)ovirt.org ]
>>>>>>>>>> To unsubscribe send an email to [
mailto:users-leave@ovirt.org |
>>>>>>>>>> users-leave(a)ovirt.org ]
>>>>>>>>>
_______________________________________________
>>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org |
users(a)ovirt.org ]
>>>>>>>>> To unsubscribe send an email to [
mailto:users-leave@ovirt.org |
>>>>>>>>> users-leave(a)ovirt.org ]
>>>>>>>>> oVirt Code of Conduct: [
>>>>>>>>>
https://www.ovirt.org/community/about/community-guidelines/ |
>>>>>>>>>
https://www.ovirt.org/community/about/community-guidelines/ ]
>>>>>>>>> List Archives: