Dear Sahina,

Is there any changes with this bug?

Still I haven't finish my upgrade process that i've started on 9th may:(

Please help me if you can.

Thanks

Tibor

----- 2018. máj.. 18., 9:29, Demeter Tibor <tdemeter@itsmart.hu> írta:
Hi,

I have to update the engine again? 

Thanks,

R
Tibor

----- 2018. máj.. 18., 6:47, Sahina Bose <sabose@redhat.com> írta:
Thanks for reporting this. https://gerrit.ovirt.org/91375 fixes this. I've re-opened bug https://bugzilla.redhat.com/show_bug.cgi?id=1574508

On Thu, May 17, 2018 at 10:12 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

4.2.4-0.0.master.20180515183442.git00e1340.el7.centos

Firstly, I did a yum update "ovirt-*-setup*"
second, I have ran engine-setup to upgrade.

I didn't remove the old repos, just installed the nightly repo.

Thank you again,

Regards,

Tibor

----- 2018. máj.. 17., 15:02, Sahina Bose <sabose@redhat.com> írta:
It doesn't look like the patch was applied. Still see the same error in engine.log
"Error while refreshing brick statuses for volume 'volume1' of cluster 'C6220': null"\

Did you use engine-setup to upgrade? What's the version of ovirt-engine currently installed?

On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

sure,

Thank you for your time!

R
Tibor

----- 2018. máj.. 17., 12:19, Sahina Bose <sabose@redhat.com> írta:
[+users]

Can you provide the engine.log to see why the monitoring is not working here. thanks!

On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

Meanwhile, I did the upgrade engine, but the gluster state is same on my first node.
I've attached some screenshot of my problem.

Thanks

Tibor



----- 2018. máj.. 16., 10:16, Demeter Tibor <tdemeter@itsmart.hu> írtaHi,

If 4.3.4 will release, i just have to remove the nightly repo and update to stable?

I'm sorry for my terrible English, I try to explain what was my problem with update.
I'm upgraded from 4.1.8.

I followed up the official hosted-engine update documentation, that was not clear me, because it has referenced to a lot of old thing (i think). 

Maybe it need to update, because I had a lot of question under upgrade and I was not sure in all of necessary steps. For example, If I need to installing the new, 4.2 repo on the hosts, then need to remove the old repo from that? 
Why I need to do a" yum update -y" on hosts, meanwhile there is an "Updatehost" menu in the GUI? So, maybe it outdated.
Since upgrade hosted engine, and the first node, I have problems with gluster. It seems to working fine if you check it from console "gluster volume status, etc" but not on the Gui, because now it yellow, and the brick reds in the first node.

Previously I did a mistake with glusterfs, my gluster config was wrong. I have corrected them, but it did not helped to me,gluster bricks are reds on my first node yet....


Now I try to upgrade to nightly, but I'm affraid, because it a living, productive system, and I don't have downtime. I hope it will help me.

Thanks for all,

Regards,
Tibor Demeter



----- 2018. máj.. 16., 9:58, Sahina Bose <sabose@redhat.com> írta:


On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

is it a different, unstable repo? I have a productive cluster, how is safe that?
I don't have any experience with nightly build. How can I use this? It have to install to the engine VM or all of my hosts? 
Thanks in advance for help me..

Only on the engine VM.

Regarding stability - it passes CI so relatively stable, beyond that there are no guarantees.

What's the specific problem you're facing with update? Can you elaborate?


Regards,

Tibor

----- 2018. máj.. 15., 9:58, Demeter Tibor <tdemeter@itsmart.hu> írta:
Hi,

Could you explain how can I use this patch?

R,
Tibor


----- 2018. máj.. 14., 11:18, Demeter Tibor <tdemeter@itsmart.hu> írta:
Hi,

Sorry for my question, but can you tell me please how can I use this patch?

Thanks,
Regards,
Tibor
----- 2018. máj.. 14., 10:47, Sahina Bose <sabose@redhat.com> írta:


On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

Could someone help me please ? I can't finish my upgrade process.

https://gerrit.ovirt.org/91164 should fix the error you're facing.

Can you elaborate why this is affecting the upgrade process?


Thanks
R
Tibor



----- 2018. máj.. 10., 12:51, Demeter Tibor <tdemeter@itsmart.hu> írta:
Hi,

I've attached the vdsm and supervdsm logs. But I don't have engine.log here, because that is on hosted engine vm. Should I send that ?

Thank you

Regards,

Tibor
----- 2018. máj.. 10., 12:30, Sahina Bose <sabose@redhat.com> írta:
There's a bug here. Can you log one attaching this engine.log and also vdsm.log & supervdsm.log from n3.itsmart.cloud

On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

I found this:


2018-05-10 03:24:19,096+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae
2018-05-10 03:24:19,097+02 ERROR [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses for volume 'volume2' of cluster 'C6220': null
2018-05-10 03:24:19,097+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 03:24:19,104+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d
2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: null
2018-05-10 03:24:19,106+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d
2018-05-10 03:24:19,107+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f
2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: null
2018-05-10 03:24:19,109+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f
2018-05-10 03:24:19,110+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58
2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: null
2018-05-10 03:24:19,112+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58
2018-05-10 03:24:19,113+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967
2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: null
2018-05-10 03:24:19,116+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967
2018-05-10 03:24:19,117+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', volumeName='volume1'}), log id: 7550e5c
2018-05-10 03:24:20,748+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c
2018-05-10 03:24:20,749+02 ERROR [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses for volume 'volume1' of cluster 'C6220': null
2018-05-10 03:24:20,750+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler8) [7715ceda] START, GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d
2018-05-10 03:24:20,930+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand, return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 120cc68d
2018-05-10 03:24:20,949+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler8) [7715ceda] START, GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264
2018-05-10 03:24:21,048+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,055+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,061+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,067+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,074+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,080+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,081+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand, return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g
luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264



2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: null
2018-05-10 11:59:26,047+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0
2018-05-10 11:59:26,048+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255
2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: null
2018-05-10 11:59:26,051+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255
2018-05-10 11:59:26,052+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e
2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: null
2018-05-10 11:59:26,054+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e
2018-05-10 11:59:26,055+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534
2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: null
2018-05-10 11:59:26,057+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534
2018-05-10 11:59:26,058+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler4) [400fa486] START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', volumeName='volume1'}), log id: 3451084f
2018-05-10 11:59:28,050+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:28,060+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:28,062+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,054+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,054+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,062+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,064+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,465+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f
2018-05-10 11:59:31,466+02 ERROR [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick statuses for volume 'volume1' of cluster 'C6220': null


R
Tibor

----- 2018. máj.. 10., 11:43, Sahina Bose <sabose@redhat.com> írta:
This doesn't affect the monitoring of state.
Any errors in vdsm.log?
Or errors in engine.log of the form "Error while refreshing brick statuses for volume"

On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

Thank you for your fast reply :)


2018-05-10 11:01:51,574+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler6) [7f01fc2d] START, GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8
2018-05-10 11:01:51,768+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand, return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8
2018-05-10 11:01:51,788+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler6) [7f01fc2d] START, GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261
2018-05-10 11:01:51,892+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,898+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,905+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,911+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,917+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,924+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,925+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand, return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, log id: 738a7261


This happening continuously.

Thanks! 
Tibor 



----- 2018. máj.. 10., 10:56, Sahina Bose <sabose@redhat.com> írta:
Could you check the engine.log if there are errors related to getting GlusterVolumeAdvancedDetails ?

On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Dear Ovirt Users,
I've followed up the self-hosted-engine upgrade documentation, I upgraded my 4.1 system to 4.2.3. 
I upgaded the first node with yum upgrade, it seems working now fine. But since upgrade, the gluster informations seems to displayed incorrect on the admin panel. The volume yellow, and there are red bricks from that node.
I've checked in console, I think my gluster is not degraded:

root@n1 ~]# gluster volume list
volume1
volume2
[root@n1 ~]# gluster volume info
 
Volume Name: volume1
Type: Distributed-Replicate
Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
Status: Started
Snapshot Count: 0
Number of Bricks: 3 x 3 = 9
Transport-type: tcp
Bricks:
Brick1: 10.104.0.1:/gluster/brick/brick1
Brick2: 10.104.0.2:/gluster/brick/brick1
Brick3: 10.104.0.3:/gluster/brick/brick1
Brick4: 10.104.0.1:/gluster/brick/brick2
Brick5: 10.104.0.2:/gluster/brick/brick2
Brick6: 10.104.0.3:/gluster/brick/brick2
Brick7: 10.104.0.1:/gluster/brick/brick3
Brick8: 10.104.0.2:/gluster/brick/brick3
Brick9: 10.104.0.3:/gluster/brick/brick3
Options Reconfigured:
transport.address-family: inet
performance.readdir-ahead: on
nfs.disable: on
storage.owner-uid: 36
storage.owner-gid: 36
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
performance.low-prio-threads: 32
network.remote-dio: enable
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
server.allow-insecure: on
 
Volume Name: volume2
Type: Distributed-Replicate
Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
Status: Started
Snapshot Count: 0
Number of Bricks: 3 x 3 = 9
Transport-type: tcp
Bricks:
Brick1: 10.104.0.1:/gluster2/brick/brick1
Brick2: 10.104.0.2:/gluster2/brick/brick1
Brick3: 10.104.0.3:/gluster2/brick/brick1
Brick4: 10.104.0.1:/gluster2/brick/brick2
Brick5: 10.104.0.2:/gluster2/brick/brick2
Brick6: 10.104.0.3:/gluster2/brick/brick2
Brick7: 10.104.0.1:/gluster2/brick/brick3
Brick8: 10.104.0.2:/gluster2/brick/brick3
Brick9: 10.104.0.3:/gluster2/brick/brick3
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
transport.address-family: inet
cluster.quorum-type: auto
network.ping-timeout: 10
auth.allow: *
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
performance.low-prio-threads: 32
network.remote-dio: enable
cluster.eager-lock: enable
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
storage.owner-uid: 36
storage.owner-gid: 36
server.allow-insecure: on
[root@n1 ~]# gluster volume status
Status of volume: volume1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick 10.104.0.1:/gluster/brick/brick1      49152     0          Y       3464 
Brick 10.104.0.2:/gluster/brick/brick1      49152     0          Y       68937
Brick 10.104.0.3:/gluster/brick/brick1      49161     0          Y       94506
Brick 10.104.0.1:/gluster/brick/brick2      49153     0          Y       3457 
Brick 10.104.0.2:/gluster/brick/brick2      49153     0          Y       68943
Brick 10.104.0.3:/gluster/brick/brick2      49162     0          Y       94514
Brick 10.104.0.1:/gluster/brick/brick3      49154     0          Y       3465 
Brick 10.104.0.2:/gluster/brick/brick3      49154     0          Y       68949
Brick 10.104.0.3:/gluster/brick/brick3      49163     0          Y       94520
Self-heal Daemon on localhost               N/A       N/A        Y       54356
Self-heal Daemon on 10.104.0.2              N/A       N/A        Y       962  
Self-heal Daemon on 10.104.0.3              N/A       N/A        Y       108977
Self-heal Daemon on 10.104.0.4              N/A       N/A        Y       61603
 
Task Status of Volume volume1
------------------------------------------------------------------------------
There are no active volume tasks
 
Status of volume: volume2
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick 10.104.0.1:/gluster2/brick/brick1     49155     0          Y       3852 
Brick 10.104.0.2:/gluster2/brick/brick1     49158     0          Y       68955
Brick 10.104.0.3:/gluster2/brick/brick1     49164     0          Y       94527
Brick 10.104.0.1:/gluster2/brick/brick2     49156     0          Y       3851 
Brick 10.104.0.2:/gluster2/brick/brick2     49159     0          Y       68961
Brick 10.104.0.3:/gluster2/brick/brick2     49165     0          Y       94533
Brick 10.104.0.1:/gluster2/brick/brick3     49157     0          Y       3883 
Brick 10.104.0.2:/gluster2/brick/brick3     49160     0          Y       68968
Brick 10.104.0.3:/gluster2/brick/brick3     49166     0          Y       94541
Self-heal Daemon on localhost               N/A       N/A        Y       54356
Self-heal Daemon on 10.104.0.2              N/A       N/A        Y       962  
Self-heal Daemon on 10.104.0.3              N/A       N/A        Y       108977
Self-heal Daemon on 10.104.0.4              N/A       N/A        Y       61603
 
Task Status of Volume volume2
------------------------------------------------------------------------------
There are no active volume tasks
 
I think ovirt can't read valid informations about gluster. 
I can't contiune upgrade of other hosts until this problem exist.

Please help me:)


Thanks

Regards,

Tibor


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org





_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org

_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives:





_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org