Hi,

I've attached the vdsm and supervdsm logs. But I don't have engine.log here, because that is on hosted engine vm. Should I send that ?

Thank you

Regards,

Tibor
----- 2018. máj.. 10., 12:30, Sahina Bose <sabose@redhat.com> írta:
There's a bug here. Can you log one attaching this engine.log and also vdsm.log & supervdsm.log from n3.itsmart.cloud

On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

I found this:


2018-05-10 03:24:19,096+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae
2018-05-10 03:24:19,097+02 ERROR [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses for volume 'volume2' of cluster 'C6220': null
2018-05-10 03:24:19,097+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 03:24:19,104+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d
2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: null
2018-05-10 03:24:19,106+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d
2018-05-10 03:24:19,107+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f
2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: null
2018-05-10 03:24:19,109+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f
2018-05-10 03:24:19,110+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58
2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: null
2018-05-10 03:24:19,112+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58
2018-05-10 03:24:19,113+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967
2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: null
2018-05-10 03:24:19,116+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967
2018-05-10 03:24:19,117+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', volumeName='volume1'}), log id: 7550e5c
2018-05-10 03:24:20,748+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) [43f4eaec] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c
2018-05-10 03:24:20,749+02 ERROR [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses for volume 'volume1' of cluster 'C6220': null
2018-05-10 03:24:20,750+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler8) [7715ceda] START, GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d
2018-05-10 03:24:20,930+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand, return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 120cc68d
2018-05-10 03:24:20,949+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler8) [7715ceda] START, GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264
2018-05-10 03:24:21,048+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,055+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,061+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,067+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,074+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,080+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,081+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand, return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g
luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264



2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: null
2018-05-10 11:59:26,047+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0
2018-05-10 11:59:26,048+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255
2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: null
2018-05-10 11:59:26,051+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255
2018-05-10 11:59:26,052+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e
2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: null
2018-05-10 11:59:26,054+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e
2018-05-10 11:59:26,055+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534
2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: null
2018-05-10 11:59:26,057+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534
2018-05-10 11:59:26,058+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler4) [400fa486] START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', volumeName='volume1'}), log id: 3451084f
2018-05-10 11:59:28,050+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:28,060+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:28,062+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,054+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,054+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,062+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,064+02 INFO  [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', sharedLocks=''}'
2018-05-10 11:59:31,465+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler4) [400fa486] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f
2018-05-10 11:59:31,466+02 ERROR [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick statuses for volume 'volume1' of cluster 'C6220': null


R
Tibor

----- 2018. máj.. 10., 11:43, Sahina Bose <sabose@redhat.com> írta:
This doesn't affect the monitoring of state.
Any errors in vdsm.log?
Or errors in engine.log of the form "Error while refreshing brick statuses for volume"

On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Hi,

Thank you for your fast reply :)


2018-05-10 11:01:51,574+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler6) [7f01fc2d] START, GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8
2018-05-10 11:01:51,768+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand, return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8
2018-05-10 11:01:51,788+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler6) [7f01fc2d] START, GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261
2018-05-10 11:01:51,892+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,898+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,905+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,911+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,917+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,924+02 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 11:01:51,925+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand, return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, log id: 738a7261


This happening continuously.

Thanks! 
Tibor 



----- 2018. máj.. 10., 10:56, Sahina Bose <sabose@redhat.com> írta:
Could you check the engine.log if there are errors related to getting GlusterVolumeAdvancedDetails ?

On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdemeter@itsmart.hu> wrote:
Dear Ovirt Users,
I've followed up the self-hosted-engine upgrade documentation, I upgraded my 4.1 system to 4.2.3. 
I upgaded the first node with yum upgrade, it seems working now fine. But since upgrade, the gluster informations seems to displayed incorrect on the admin panel. The volume yellow, and there are red bricks from that node.
I've checked in console, I think my gluster is not degraded:

root@n1 ~]# gluster volume list
volume1
volume2
[root@n1 ~]# gluster volume info
 
Volume Name: volume1
Type: Distributed-Replicate
Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
Status: Started
Snapshot Count: 0
Number of Bricks: 3 x 3 = 9
Transport-type: tcp
Bricks:
Brick1: 10.104.0.1:/gluster/brick/brick1
Brick2: 10.104.0.2:/gluster/brick/brick1
Brick3: 10.104.0.3:/gluster/brick/brick1
Brick4: 10.104.0.1:/gluster/brick/brick2
Brick5: 10.104.0.2:/gluster/brick/brick2
Brick6: 10.104.0.3:/gluster/brick/brick2
Brick7: 10.104.0.1:/gluster/brick/brick3
Brick8: 10.104.0.2:/gluster/brick/brick3
Brick9: 10.104.0.3:/gluster/brick/brick3
Options Reconfigured:
transport.address-family: inet
performance.readdir-ahead: on
nfs.disable: on
storage.owner-uid: 36
storage.owner-gid: 36
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
performance.low-prio-threads: 32
network.remote-dio: enable
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
server.allow-insecure: on
 
Volume Name: volume2
Type: Distributed-Replicate
Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
Status: Started
Snapshot Count: 0
Number of Bricks: 3 x 3 = 9
Transport-type: tcp
Bricks:
Brick1: 10.104.0.1:/gluster2/brick/brick1
Brick2: 10.104.0.2:/gluster2/brick/brick1
Brick3: 10.104.0.3:/gluster2/brick/brick1
Brick4: 10.104.0.1:/gluster2/brick/brick2
Brick5: 10.104.0.2:/gluster2/brick/brick2
Brick6: 10.104.0.3:/gluster2/brick/brick2
Brick7: 10.104.0.1:/gluster2/brick/brick3
Brick8: 10.104.0.2:/gluster2/brick/brick3
Brick9: 10.104.0.3:/gluster2/brick/brick3
Options Reconfigured:
nfs.disable: on
performance.readdir-ahead: on
transport.address-family: inet
cluster.quorum-type: auto
network.ping-timeout: 10
auth.allow: *
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
performance.low-prio-threads: 32
network.remote-dio: enable
cluster.eager-lock: enable
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
storage.owner-uid: 36
storage.owner-gid: 36
server.allow-insecure: on
[root@n1 ~]# gluster volume status
Status of volume: volume1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick 10.104.0.1:/gluster/brick/brick1      49152     0          Y       3464 
Brick 10.104.0.2:/gluster/brick/brick1      49152     0          Y       68937
Brick 10.104.0.3:/gluster/brick/brick1      49161     0          Y       94506
Brick 10.104.0.1:/gluster/brick/brick2      49153     0          Y       3457 
Brick 10.104.0.2:/gluster/brick/brick2      49153     0          Y       68943
Brick 10.104.0.3:/gluster/brick/brick2      49162     0          Y       94514
Brick 10.104.0.1:/gluster/brick/brick3      49154     0          Y       3465 
Brick 10.104.0.2:/gluster/brick/brick3      49154     0          Y       68949
Brick 10.104.0.3:/gluster/brick/brick3      49163     0          Y       94520
Self-heal Daemon on localhost               N/A       N/A        Y       54356
Self-heal Daemon on 10.104.0.2              N/A       N/A        Y       962  
Self-heal Daemon on 10.104.0.3              N/A       N/A        Y       108977
Self-heal Daemon on 10.104.0.4              N/A       N/A        Y       61603
 
Task Status of Volume volume1
------------------------------------------------------------------------------
There are no active volume tasks
 
Status of volume: volume2
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick 10.104.0.1:/gluster2/brick/brick1     49155     0          Y       3852 
Brick 10.104.0.2:/gluster2/brick/brick1     49158     0          Y       68955
Brick 10.104.0.3:/gluster2/brick/brick1     49164     0          Y       94527
Brick 10.104.0.1:/gluster2/brick/brick2     49156     0          Y       3851 
Brick 10.104.0.2:/gluster2/brick/brick2     49159     0          Y       68961
Brick 10.104.0.3:/gluster2/brick/brick2     49165     0          Y       94533
Brick 10.104.0.1:/gluster2/brick/brick3     49157     0          Y       3883 
Brick 10.104.0.2:/gluster2/brick/brick3     49160     0          Y       68968
Brick 10.104.0.3:/gluster2/brick/brick3     49166     0          Y       94541
Self-heal Daemon on localhost               N/A       N/A        Y       54356
Self-heal Daemon on 10.104.0.2              N/A       N/A        Y       962  
Self-heal Daemon on 10.104.0.3              N/A       N/A        Y       108977
Self-heal Daemon on 10.104.0.4              N/A       N/A        Y       61603
 
Task Status of Volume volume2
------------------------------------------------------------------------------
There are no active volume tasks
 
I think ovirt can't read valid informations about gluster. 
I can't contiune upgrade of other hosts until this problem exist.

Please help me:)


Thanks

Regards,

Tibor


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org