Hi,
Sorry for my question, but can you tell me please how can I use this patch?
Thanks,
Regards,
Tibor
----- 2018. máj.. 14., 10:47, Sahina Bose <sabose(a)redhat.com> írta:
On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor < [
mailto:tdemeter@itsmart.hu |
tdemeter(a)itsmart.hu ] > wrote:
> Hi,
> Could someone help me please ? I can't finish my upgrade
process.
Can you elaborate why this is affecting the upgrade process?
> Thanks
> R
> Tibor
> ----- 2018. máj.. 10., 12:51, Demeter Tibor < [
mailto:tdemeter@itsmart.hu |
> tdemeter(a)itsmart.hu ] > írta:
>
> Hi,
>> I've attached the vdsm and supervdsm logs. But I
don't have engine.log here,
>> because that is on hosted engine vm. Should I send that ?
>> Thank you
>> Regards,
>> Tibor
>> ----- 2018. máj.. 10., 12:30, Sahina Bose < [ mailto:sabose@redhat.com |
>> sabose(a)redhat.com ] > írta:
>>> There's a bug here. Can you log one attaching this
engine.log and also vdsm.log
>>> & supervdsm.log from n3.itsmart.cloud
>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor < [
mailto:tdemeter@itsmart.hu |
>>> tdemeter(a)itsmart.hu ] > wrote:
>>>
> Hi,
>>>> I found this:
>>>> 2018-05-10 03:24:19,096+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e,
>>>> log id: 347435ae
>>>> 2018-05-10 03:24:19,097+02 ERROR
>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
(DefaultQuartzScheduler7)
>>>> [43f4eaec] Error while refreshing brick statuses for volume
'volume2' of
>>>> cluster 'C6220': null
>>>> 2018-05-10 03:24:19,097+02 INFO
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
(DefaultQuartzScheduler8)
>>>> [7715ceda] Failed to acquire lock and wait lock
>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 03:24:19,104+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>> log id: 6908121d
>>>> 2018-05-10 03:24:19,106+02 ERROR
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
n4.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>> execution failed: null
>>>> 2018-05-10 03:24:19,106+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d
>>>> 2018-05-10 03:24:19,107+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}),
>>>> log id: 735c6a5f
>>>> 2018-05-10 03:24:19,109+02 ERROR
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
n1.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>> execution failed: null
>>>> 2018-05-10 03:24:19,109+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f
>>>> 2018-05-10 03:24:19,110+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>> log id: 6f9e9f58
>>>> 2018-05-10 03:24:19,112+02 ERROR
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
n2.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>> execution failed: null
>>>> 2018-05-10 03:24:19,112+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58
>>>> 2018-05-10 03:24:19,113+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>> log id: 2ee46967
>>>> 2018-05-10 03:24:19,115+02 ERROR
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
n3.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>> execution failed: null
>>>> 2018-05-10 03:24:19,116+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967
>>>> 2018-05-10 03:24:19,117+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n1.itsmart.cloud,
>>>>
GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
>>>> volumeName='volume1'}), log id: 7550e5c
>>>> 2018-05-10 03:24:20,748+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f,
>>>> log id: 7550e5c
>>>> 2018-05-10 03:24:20,749+02 ERROR
>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
(DefaultQuartzScheduler7)
>>>> [43f4eaec] Error while refreshing brick statuses for volume
'volume1' of
>>>> cluster 'C6220': null
>>>> 2018-05-10 03:24:20,750+02 INFO
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>> (DefaultQuartzScheduler8) [7715ceda] START,
>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>> log id: 120cc68d
>>>> 2018-05-10 03:24:20,930+02 INFO
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH,
GlusterServersListVDSCommand,
>>>> return: [ [
http://10.101.0.2/24:CONNECTED | 10.101.0.2/24:CONNECTED ] ,
>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED],
log
>>>> id: 120cc68d
>>>> 2018-05-10 03:24:20,949+02 INFO
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>> (DefaultQuartzScheduler8) [7715ceda] START,
>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>
GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>> log id: 118aa264
>>>> 2018-05-10 03:24:21,048+02 WARN
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>> '10.104.0.1:/gluster/brick/brick1' of volume
>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no
gluster
>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>> 2018-05-10 03:24:21,055+02 WARN
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>> '10.104.0.1:/gluster/brick/brick2' of volume
>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no
gluster
>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>> 2018-05-10 03:24:21,061+02 WARN
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>> '10.104.0.1:/gluster/brick/brick3' of volume
>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no
gluster
>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>> 2018-05-10 03:24:21,067+02 WARN
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>> '10.104.0.1:/gluster2/brick/brick1' of volume
>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no
gluster
>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>> 2018-05-10 03:24:21,074+02 WARN
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>> '10.104.0.1:/gluster2/brick/brick2' of volume
>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no
gluster
>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>> 2018-05-10 03:24:21,080+02 WARN
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>> '10.104.0.1:/gluster2/brick/brick3' of volume
>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no
gluster
>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>> 2018-05-10 03:24:21,081+02 INFO
>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH,
GlusterVolumesListVDSCommand,
>>>> return:
>>>>
{68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>
e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g
>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264
>>>> 2018-05-10 11:59:26,047+02 ERROR
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
n1.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>> execution failed: null
>>>> 2018-05-10 11:59:26,047+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0
>>>> 2018-05-10 11:59:26,048+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>> log id: 28d9e255
>>>> 2018-05-10 11:59:26,051+02 ERROR
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
n4.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>> execution failed: null
>>>> 2018-05-10 11:59:26,051+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255
>>>> 2018-05-10 11:59:26,052+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>> log id: 4a7b280e
>>>> 2018-05-10 11:59:26,054+02 ERROR
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
n2.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>> execution failed: null
>>>> 2018-05-10 11:59:26,054+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e
>>>> 2018-05-10 11:59:26,055+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>> log id: 18adc534
>>>> 2018-05-10 11:59:26,057+02 ERROR
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
n3.itsmart.cloud,
>>>>
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>> execution failed: null
>>>> 2018-05-10 11:59:26,057+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534
>>>> 2018-05-10 11:59:26,058+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n3.itsmart.cloud,
>>>>
GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
>>>> volumeName='volume1'}), log id: 3451084f
>>>> 2018-05-10 11:59:28,050+02 INFO
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
(DefaultQuartzScheduler3)
>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:28,060+02 INFO
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
(DefaultQuartzScheduler3)
>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:28,062+02 INFO
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
(DefaultQuartzScheduler3)
>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,054+02 INFO
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
(DefaultQuartzScheduler3)
>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,054+02 INFO
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
(DefaultQuartzScheduler3)
>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,062+02 INFO
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
(DefaultQuartzScheduler3)
>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,064+02 INFO
>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
(DefaultQuartzScheduler3)
>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>> sharedLocks=''}'
>>>> 2018-05-10 11:59:31,465+02 INFO
>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43,
>>>> log id: 3451084f
>>>> 2018-05-10 11:59:31,466+02 ERROR
>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
(DefaultQuartzScheduler4)
>>>> [400fa486] Error while refreshing brick statuses for volume
'volume1' of
>>>> cluster 'C6220': null
>>>> R
>>>> Tibor
>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose < [
mailto:sabose@redhat.com |
>>>> sabose(a)redhat.com ] > írta:
>>>>> This doesn't affect the monitoring of state.
>>>>> Any errors in vdsm.log?
>>>>> Or errors in engine.log of the form "Error while refreshing
brick statuses for
>>>>> volume"
>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor
< [ mailto:tdemeter@itsmart.hu |
>>>>> tdemeter(a)itsmart.hu ] > wrote:
>>>>>
> Hi,
>>>>>> Thank you for your fast reply :)
>>>>>> 2018-05-10 11:01:51,574+02 INFO
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>> log id: 39adbbb8
>>>>>> 2018-05-10 11:01:51,768+02 INFO
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH,
GlusterServersListVDSCommand,
>>>>>> return: [ [
http://10.101.0.2/24:CONNECTED |
10.101.0.2/24:CONNECTED ] ,
>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED,
10.104.0.4:CONNECTED], log
>>>>>> id: 39adbbb8
>>>>>> 2018-05-10 11:01:51,788+02 INFO
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>
GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>> log id: 738a7261
>>>>>> 2018-05-10 11:01:51,892+02 WARN
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume
>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct
network as no gluster
>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 11:01:51,898+02 WARN
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume
>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct
network as no gluster
>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 11:01:51,905+02 WARN
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume
>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct
network as no gluster
>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 11:01:51,911+02 WARN
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume
>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct
network as no gluster
>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 11:01:51,917+02 WARN
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume
>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct
network as no gluster
>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 11:01:51,924+02 WARN
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume
>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct
network as no gluster
>>>>>> network found in cluster
'59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 11:01:51,925+02 INFO
>>>>>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH,
GlusterVolumesListVDSCommand,
>>>>>> return:
>>>>>>
{68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>
e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>>>>>> log id: 738a7261
>>>>>> This happening continuously.
>>>>>> Thanks!
>>>>>> Tibor
>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose
< [ mailto:sabose@redhat.com |
>>>>>> sabose(a)redhat.com ] > írta:
>>>>>>> Could you check the engine.log if there
are errors related to getting
>>>>>>> GlusterVolumeAdvancedDetails ?
>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter
Tibor < [ mailto:tdemeter@itsmart.hu |
>>>>>>> tdemeter(a)itsmart.hu ] > wrote:
>>>>>>>> Dear Ovirt Users,
>>>>>>>> I've followed up the self-hosted-engine upgrade
documentation, I upgraded my 4.1
>>>>>>>> system to 4.2.3.
>>>>>>>> I upgaded the first node with yum upgrade, it seems
working now fine. But since
>>>>>>>> upgrade, the gluster informations seems to displayed
incorrect on the admin
>>>>>>>> panel. The volume yellow, and there are red bricks from
that node.
>>>>>>>> I've checked in console, I think my gluster is not
degraded:
>>>>>>>> root@n1 ~]# gluster volume list
>>>>>>>> volume1
>>>>>>>> volume2
>>>>>>>> [root@n1 ~]# gluster volume info
>>>>>>>> Volume Name: volume1
>>>>>>>> Type: Distributed-Replicate
>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>>>>>>> Status: Started
>>>>>>>> Snapshot Count: 0
>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>> Transport-type: tcp
>>>>>>>> Bricks:
>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>>>>>>> Options Reconfigured:
>>>>>>>> transport.address-family: inet
>>>>>>>> performance.readdir-ahead: on
>>>>>>>> nfs.disable: on
>>>>>>>> storage.owner-uid: 36
>>>>>>>> storage.owner-gid: 36
>>>>>>>> performance.quick-read: off
>>>>>>>> performance.read-ahead: off
>>>>>>>> performance.io-cache: off
>>>>>>>> performance.stat-prefetch: off
>>>>>>>> performance.low-prio-threads: 32
>>>>>>>> network.remote-dio: enable
>>>>>>>> cluster.eager-lock: enable
>>>>>>>> cluster.quorum-type: auto
>>>>>>>> cluster.server-quorum-type: server
>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>> cluster.locking-scheme: granular
>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>> features.shard: on
>>>>>>>> user.cifs: off
>>>>>>>> server.allow-insecure: on
>>>>>>>> Volume Name: volume2
>>>>>>>> Type: Distributed-Replicate
>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>>>>>>> Status: Started
>>>>>>>> Snapshot Count: 0
>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>> Transport-type: tcp
>>>>>>>> Bricks:
>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>>>>>>> Options Reconfigured:
>>>>>>>> nfs.disable: on
>>>>>>>> performance.readdir-ahead: on
>>>>>>>> transport.address-family: inet
>>>>>>>> cluster.quorum-type: auto
>>>>>>>> network.ping-timeout: 10
>>>>>>>> auth.allow: *
>>>>>>>> performance.quick-read: off
>>>>>>>> performance.read-ahead: off
>>>>>>>> performance.io-cache: off
>>>>>>>> performance.stat-prefetch: off
>>>>>>>> performance.low-prio-threads: 32
>>>>>>>> network.remote-dio: enable
>>>>>>>> cluster.eager-lock: enable
>>>>>>>> cluster.server-quorum-type: server
>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>> cluster.locking-scheme: granular
>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>> features.shard: on
>>>>>>>> user.cifs: off
>>>>>>>> storage.owner-uid: 36
>>>>>>>> storage.owner-gid: 36
>>>>>>>> server.allow-insecure: on
>>>>>>>> [root@n1 ~]# gluster volume status
>>>>>>>> Status of volume: volume1
>>>>>>>> Gluster process TCP Port RDMA Port Online Pid
>>>>>>>>
------------------------------------------------------------------------------
>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464
>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937
>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506
>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457
>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943
>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514
>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465
>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949
>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520
>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>> Task Status of Volume volume1
>>>>>>>>
------------------------------------------------------------------------------
>>>>>>>> There are no active volume tasks
>>>>>>>> Status of volume: volume2
>>>>>>>> Gluster process TCP Port RDMA Port Online Pid
>>>>>>>>
------------------------------------------------------------------------------
>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852
>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955
>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527
>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851
>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961
>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533
>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883
>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968
>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541
>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>> Task Status of Volume volume2
>>>>>>>>
------------------------------------------------------------------------------
>>>>>>>> There are no active volume tasks
>>>>>>>> I think ovirt can't read valid informations about
gluster.
>>>>>>>> I can't contiune upgrade of other hosts until this
problem exist.
>>>>>>>> Please help me:)
>>>>>>>> Thanks
>>>>>>
>> Regards,
>>>>>>>> Tibor
>>>>>>>>
_______________________________________________
>>>>>>>> Users mailing list -- [ mailto:users@ovirt.org |
users(a)ovirt.org ]
>>>>>>>> To unsubscribe send an email to [
mailto:users-leave@ovirt.org |
>>>>>>>> users-leave(a)ovirt.org ]
>> _______________________________________________
>> Users mailing list -- [ mailto:users@ovirt.org | users(a)ovirt.org ]
>> To unsubscribe send an email to [ mailto:users-leave@ovirt.org |
>> users-leave(a)ovirt.org ]