What about the gluster logs ?
Best Regards,
Strahil Nikolov
В четвъртък, 3 септември 2020 г., 20:57:05 Гринуич+3, souvaliotimaria(a)mail.com
<souvaliotimaria(a)mail.com> написа:
Thank you very much for your reply.
I checked the NTP and realized the service wasn't working properly on two of the three
nodes, but despite that the clocks seemed to have the correct time(date and hwdate). I
switched to chronyd and stopped the ntpd service and now it seems the servers' clocks
are synchronized.
The time in BIOS has different time than the systems. Does this affect the behaviour of
the overall performance?
This is what I could gather from the logs:
Node1:
[root@ov-no1 ~]# more /var/log/messages |egrep "(WARN|error)"|more
d
Aug 27 17:53:08 ov-no1 libvirtd: 2020-08-27 14:53:08.947+0000: 5613: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Aug 27 17:58:08 ov-no1 libvirtd: 2020-08-27 14:58:08.943+0000: 5613: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Aug 27 18:03:08 ov-no1 libvirtd: 2020-08-27 15:03:08.937+0000: 5614: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Aug 27 18:08:08 ov-no1 libvirtd: 2020-08-27 15:08:08.951+0000: 5617: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Aug 27 18:13:08 ov-no1 libvirtd: 2020-08-27 15:13:08.951+0000: 5616: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Aug 27 18:18:08 ov-no1 libvirtd: 2020-08-27 15:18:08.942+0000: 5618: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
.
.
[Around that time Node 3, who is the arbiter node, was placed in local Maintenance mode.
It was shut down for maintenance and when we boot it up again all seemed right. We removed
it from the maintenance mode and when the healing processes finished, Node 1 became
NonResponsive. Long story short, the VDSM agent sent Node1 a restart command. Node1
rebooted, HostedEngine was up on Node1 and the rest of the VMs that were hosted by Node1
had to be manually brought up. Since then everything seemed to be working as it should.
The HostedEngine VM shutdown with no apparent reason 5 days later, making us believe there
was no connection between the two incidents.]
.
.
.
.
.
Sep 1 05:53:30 ov-no1 vdsm[6706]: WARN Worker blocked: <Worker name=jsonrpc/3 running
<Task <JsonRpcTask {'params': {}, 'jsonrpc': '2.0',
'method':
u'GlusterVolume.geoRepSessionList', 'id':
u'e1f4db88-5d36-4ba1-8848-3855c1dbc6ee'} at 0x7f1ed7381650> timeout=60,
duration=60.00 at 0x7f1ed7381dd0> t
ask#=76268 at 0x7f1ebc0797d0>, traceback:#012File:
"/usr/lib64/python2.7/threading.py", line 785, in __bootstrap#012
self.__bootstrap_inner()#012Fil
e: "/usr/lib64/python2.7/threading.py", line 812, in __bootstrap_inner#012
self.run()#012File: "/usr/lib64/python2.7/threading.py", line 765, in run
#012 self.__target(*self.__args, **self.__kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/concurrent.py", line 195, in
run#012 ret =
func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 301, in _run#012
self._execute_task()#012File: "/usr/lib/p
ython2.7/site-packages/vdsm/executor.py", line 315, in _execute_task#012
task()#012File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 3
91, in __call__#012 self._callable()#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 262, in
__call__#012 self._handler(sel
f._ctx, self._req)#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 305, in
_serveRequest#012 response = self._handle_request
(req, ctx)#012File: "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py",
line 345, in _handle_request#012 res = method(**params)#012File: "/usr
/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 194, in _dynamicMethod#012
result = fn(*methodArgs)#012File: "/usr/lib/python2.7/site-package
s/vdsm/gluster/apiwrapper.py", line 237, in geoRepSessionList#012
remoteUserName)#012File: "/usr/lib/python2.7/site-packages/vdsm/gluster/api.py",
l
ine 93, in wrapper#012 rv = func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/gluster/api.py", line 551, in
volumeGeoRepSessionL
ist#012 remoteUserName,#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 56, in
__call__#012 return callMethod()#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 54, in
<lambda>#012 **kwargs)#012File: "<string>", line 2, in
glusterVolumeGeoRep
Status#012File: "/usr/lib64/python2.7/multiprocessing/managers.py", line 759, in
_callmethod#012 kind, result = conn.recv()
Sep 1 05:54:30 ov-no1 vdsm[6706]: WARN Worker blocked: <Worker name=jsonrpc/3 running
<Task <JsonRpcTask {'params': {}, 'jsonrpc': '2.0',
'method':
u'GlusterVolume.geoRepSessionList', 'id':
u'e1f4db88-5d36-4ba1-8848-3855c1dbc6ee'} at 0x7f1ed7381650> timeout=60,
duration=120.00 at 0x7f1ed7381dd0>
task#=76268 at 0x7f1ebc0797d0>, traceback:#012File:
"/usr/lib64/python2.7/threading.py", line 785, in __bootstrap#012
self.__bootstrap_inner()#012Fi
le: "/usr/lib64/python2.7/threading.py", line 812, in __bootstrap_inner#012
self.run()#012File: "/usr/lib64/python2.7/threading.py", line 765, in ru
n#012 self.__target(*self.__args, **self.__kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/concurrent.py", line 195, in
run#012 ret
= func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 301, in _run#012
self._execute_task()#012File: "/usr/lib/
python2.7/site-packages/vdsm/executor.py", line 315, in _execute_task#012
task()#012File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", line
391, in __call__#012 self._callable()#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 262, in
__call__#012 self._handler(se
lf._ctx, self._req)#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 305, in
_serveRequest#012 response = self._handle_reques
t(req, ctx)#012File: "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py",
line 345, in _handle_request#012 res = method(**params)#012File: "/us
r/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 194, in _dynamicMethod#012
result = fn(*methodArgs)#012File: "/usr/lib/python2.7/site-packag
es/vdsm/gluster/apiwrapper.py", line 237, in geoRepSessionList#012
remoteUserName)#012File: "/usr/lib/python2.7/site-packages/vdsm/gluster/api.py",
line 93, in wrapper#012 rv = func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/gluster/api.py", line 551, in
volumeGeoRepSession
List#012 remoteUserName,#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 56, in
__call__#012 return callMethod()#012File
: "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 54, in
<lambda>#012 **kwargs)#012File: "<string>", line 2, in
glusterVolumeGeoRe
pStatus#012File: "/usr/lib64/python2.7/multiprocessing/managers.py", line 759,
in _callmethod#012 kind, result = conn.recv()
[root@ov-no1 ~]# xzcat /var/log/vdsm/vdsm.log.57.xz |egrep "(WARN|error)" |more
2020-09-01 09:40:50,023+0300 INFO (jsonrpc/5) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:40:50,023+0300 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:41:00,061+0300 INFO (jsonrpc/0) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:41:00,061+0300 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:41:10,078+0300 INFO (jsonrpc/4) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:41:10,078+0300 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:41:20,086+0300 INFO (jsonrpc/0) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:41:20,086+0300 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:41:30,107+0300 INFO (jsonrpc/0) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:41:30,107+0300 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:41:38,710+0300 INFO (jsonrpc/4) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:41:38,711+0300 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:41:40,115+0300 INFO (jsonrpc/5) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:41:40,115+0300 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.01 seconds (__init__:312)
2020-09-01 09:41:44,541+0300 WARN (monitor/e2e7282) [storage.PersistentDict] Could not
parse line `
2020-09-01 09:41:44,544+0300 WARN (monitor/0a5e552) [storage.PersistentDict] Could not
parse line `
[root@ov-no1 ~]# xzcat /var/log/vdsm/supervdsm.log.29.xz |egrep
"(WARN|error|ERR)" |more
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,182::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,183::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,185::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,304::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,306::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,308::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,440::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,441::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,442::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,442::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,443::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,443::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,444::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,445::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,445::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::ERROR::2020-09-01 09:32:36,358::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-01 09:32:36,361::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-01 09:32:36,363::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-01 09:32:36,365::udev::53::blivet::(get_device) No
device at ''
[root@ov-no1 ~]# xzcat /var/log/vdsm/supervdsm.log.28.xz |egrep
"(WARN|error|ERR)" |more
MainProcess|jsonrpc/7::WARNING::2020-09-01
11:32:36,730::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/7::WARNING::2020-09-01
11:32:36,731::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/7::WARNING::2020-09-01
11:32:36,732::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/7::WARNING::2020-09-01
11:32:36,847::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/7::WARNING::2020-09-01
11:32:36,849::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/7::WARNING::2020-09-01
11:32:36,850::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,974::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,975::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,975::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,976::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,977::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,977::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,978::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,979::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/7::WARNING::2020-09-01 11:32:36,979::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/7::ERROR::2020-09-01 11:32:40,951::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/7::ERROR::2020-09-01 11:32:40,954::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/7::ERROR::2020-09-01 11:32:40,956::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/7::ERROR::2020-09-01 11:32:40,958::udev::53::blivet::(get_device) No
device at ''
[root@ov-no1 ~]# cat /var/log/ovirt-hosted-engine-ha/broker.log.2020-09-01 |egrep -i
"(WARN|ERR)" |more
[nothing in the output]
[root@ov-no1 ~]# gluster volume status
Status of volume: data
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 6070
Brick ov-no2.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 28860
Brick ov-no3.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 5752
Self-heal Daemon on localhost N/A N/A Y 6539
Self-heal Daemon on ov-no2.ariadne-t.local N/A N/A Y 813
Self-heal Daemon on ov-no3.ariadne-t.local N/A N/A Y 19353
Task Status of Volume data
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: engine
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/engine/engine 49153 0 Y 6098
Brick ov-no2.ariadne-t.local:/gluster_brick
s/engine/engine 49156 0 Y 28867
Brick ov-no3.ariadne-t.local:/gluster_brick
s/engine/engine 49153 0 Y 5789
Self-heal Daemon on localhost N/A N/A Y 6539
Self-heal Daemon on ov-no2.ariadne-t.local N/A N/A Y 813
Self-heal Daemon on ov-no3.ariadne-t.local N/A N/A Y 19353
Task Status of Volume engine
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: vmstore
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49154 0 Y 6102
Brick ov-no2.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49157 0 Y 28878
Brick ov-no3.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49154 0 Y 5772
Self-heal Daemon on localhost N/A N/A Y 6539
Self-heal Daemon on ov-no2.ariadne-t.local N/A N/A Y 813
Self-heal Daemon on ov-no3.ariadne-t.local N/A N/A Y 19353
Task Status of Volume vmstore
------------------------------------------------------------------------------
There are no active volume tasks
[root@ov-no1 ~]# gluster volume heal engine info
Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
[root@ov-no1 ~]# gluster volume heal engine info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no1 ~]# gluster volume heal data info
Brick ov-no1.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
[root@ov-no1 ~]# gluster volume heal data info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no1 ~]# gluster volume heal vmstore info
Brick ov-no1.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
[root@ov-no1 ~]# gluster volume heal vmstore info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no1 ~]# ps -A|grep gluster
5522 ? 00:01:40 glustereventsd
5591 ? 00:05:04 glusterd
6070 ? 02:16:28 glusterfsd
6098 ? 05:44:14 glusterfsd
6102 ? 00:23:48 glusterfsd
6285 ? 00:00:38 glustereventsd
6539 ? 01:32:19 glusterfs
10470 ? 00:34:10 glusterfs
17763 ? 00:50:18 glusterfs
18101 ? 00:10:23 glusterfs
Node2:
[root@ov-no2 ~]# more /var/log/messages |egrep "(WARN|error)"|more
Sep 1 07:58:30 ov-no2 vdsm[13586]: WARN Worker blocked: <Worker name=jsonrpc/7 running
<Task <JsonRpcTask {'params': {}, 'jsonrpc': '2.0',
'method':
u'GlusterVolume.geoRepSessionList', 'id':
u'ec8af6cc-6109-4ddb-a6fb-ad9f4fbc1e7b'} at 0x7f44602a7c90> timeout=60,
duration=120.00 at 0x7f44602a7510>
task#=5773350 at 0x7f44c00c71d0>, traceback:#012File:
"/usr/lib64/python2.7/threading.py", line 785, in __bootstrap#012
self.__bootstrap_inner()#01
2File: "/usr/lib64/python2.7/threading.py", line 812, in __bootstrap_inner#012
self.run()#012File: "/usr/lib64/python2.7/threading.py", line 765, in
run#012 self.__target(*self.__args, **self.__kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/concurrent.py", line 195, in
run#012 r
et = func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 301, in _run#012
self._execute_task()#012File: "/usr/l
ib/python2.7/site-packages/vdsm/executor.py", line 315, in _execute_task#012
task()#012File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", li
ne 391, in __call__#012 self._callable()#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 262, in
__call__#012 self._handler
(self._ctx, self._req)#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 305, in
_serveRequest#012 response = self._handle_req
uest(req, ctx)#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 345, in
_handle_request#012 res = method(**params)#012File: "
/usr/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 194, in
_dynamicMethod#012 result = fn(*methodArgs)#012File: "/usr/lib/python2.7/site-pac
kages/vdsm/gluster/apiwrapper.py", line 237, in geoRepSessionList#012
remoteUserName)#012File: "/usr/lib/python2.7/site-packages/vdsm/gluster/api.py
", line 93, in wrapper#012 rv = func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/gluster/api.py", line 551, in
volumeGeoRepSess
ionList#012 remoteUserName,#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 56, in
__call__#012 return callMethod()#012F
ile: "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 54, in
<lambda>#012 **kwargs)#012File: "<string>", line 2, in
glusterVolumeGe
oRepStatus#012File: "/usr/lib64/python2.7/multiprocessing/managers.py", line
759, in _callmethod#012 kind, result = conn.recv()
Sep 1 09:48:02 ov-no2 journal: ovirt-ha-agent
ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine ERROR Unhandled monitoring loop
exception#012T
raceback (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
line 449, in start_moni
toring#012 self._monitoring_loop()#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
line 468, in _monit
oring_loop#012 for old_state, state, delay in self.fsm:#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/fsm/machine.py",
li
ne 127, in next#012 new_data = self.refresh(self._state.data)#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/state_machi
ne.py", line 81, in refresh#012
stats.update(self.hosted_engine.collect_stats())#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_h
a/agent/hosted_engine.py", line 756, in collect_stats#012 all_stats =
self._broker.get_stats_from_storage()#012 File "/usr/lib/python2.7/site-pac
kages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 143, in
get_stats_from_storage#012 result = self._proxy.get_stats()#012 File
"/usr/lib64/py
thon2.7/xmlrpclib.py", line 1233, in __call__#012 return self.__send(self.__name,
args)#012 File "/usr/lib64/python2.7/xmlrpclib.py", line 1591,
in __request#012 verbose=self.__verbose#012 File
"/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request#012 return
self.single_request(hos
t, handler, request_body, verbose)#012 File
"/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request#012
self.send_content(h, request_bo
dy)#012 File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in
send_content#012 connection.endheaders(request_body)#012 File "/usr/lib64/python
2.7/httplib.py", line 1037, in endheaders#012 self._send_output(message_body)#012
File "/usr/lib64/python2.7/httplib.py", line 881, in _send_outp
ut#012 self.send(msg)#012 File "/usr/lib64/python2.7/httplib.py", line 843,
in send#012 self.connect()#012 File "/usr/lib/python2.7/site-pack
ages/ovirt_hosted_engine_ha/lib/unixrpc.py", line 52, in connect#012
self.sock.connect(base64.b16decode(self.host))#012 File "/usr/lib64/python2.
7/socket.py", line 224, in meth#012 return
getattr(self._sock,name)(*args)#012error: [Errno 2] No such file or directory
Sep 1 09:48:02 ov-no2 journal: ovirt-ha-agent ovirt_hosted_engine_ha.agent.agent.Agent
ERROR Traceback (most recent call last):#012 File "/usr/lib/
python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 131, in
_run_agent#012 return action(he)#012 File "/usr/lib/python2.7/site-p
ackages/ovirt_hosted_engine_ha/agent/agent.py", line 55, in action_proper#012
return he.start_monitoring()#012 File "/usr/lib/python2.7/site-pack
ages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 456, in
start_monitoring#012 self.publish(stopped)#012 File "/usr/lib/python2.7/site-pa
ckages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 356, in publish#012
self._push_to_storage(blocks)#012 File "/usr/lib/python2.7/site-p
ackages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 727, in
_push_to_storage#012 self._broker.put_stats_on_storage(self.host_id, blocks)#
012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
line 113, in put_stats_on_storage#012 self._proxy.put_stats
(host_id, xmlrpclib.Binary(data))#012 File "/usr/lib64/python2.7/xmlrpclib.py",
line 1233, in __call__#012 return self.__send(self.__name, args)#
012 File "/usr/lib64/python2.7/xmlrpclib.py", line 1591, in __request#012
verbose=self.__verbose#012 File "/usr/lib64/python2.7/xmlrpclib.py", l
ine 1273, in request#012 return self.single_request(host, handler, request_body,
verbose)#012 File "/usr/lib64/python2.7/xmlrpclib.py", line 1301
, in single_request#012 self.send_content(h, request_body)#012 File
"/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content#012 connec
tion.endheaders(request_body)#012 File "/usr/lib64/python2.7/httplib.py", line
1037, in endheaders#012 self._send_output(message_body)#012 File
"/usr/lib64/python2.7/httplib.py", line 881, in _send_output#012
self.send(msg)#012 File "/usr/lib64/python2.7/httplib.py", line 843, in
send#012
self.connect()#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/unixrpc.py", line
52, in connect#012 self.sock.connect(b
ase64.b16decode(self.host))#012 File "/usr/lib64/python2.7/socket.py", line
224, in meth#012 return getattr(self._sock,name)(*args)#012error: [Er
rno 2] No such file or directory
Sep 1 09:48:15 ov-no2 journal: ovirt-ha-broker
ovirt_hosted_engine_ha.broker.notifications.Notifications ERROR [Errno 113] No route to
host#012Trace
back (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/notifications.py",
line 26, in send_email#012
timeout=float(cfg["smtp-timeout"]))#012 File
"/usr/lib64/python2.7/smtplib.py", line 255, in __init__#012 (code, msg) =
self.connect(host, po
rt)#012 File "/usr/lib64/python2.7/smtplib.py", line 315, in connect#012
self.sock = self._get_socket(host, port, self.timeout)#012 File "/usr/l
ib64/python2.7/smtplib.py", line 290, in _get_socket#012 return
socket.create_connection((host, port), timeout)#012 File "/usr/lib64/python2.7/so
cket.py", line 571, in create_connection#012 raise err#012error: [Errno 113] No
route to host
Sep 1 09:48:25 ov-no2 journal: ovirt-ha-broker
ovirt_hosted_engine_ha.broker.notifications.Notifications ERROR [Errno 113] No route to
host#012Trace
back (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/notifications.py",
line 26, in send_email#012
timeout=float(cfg["smtp-timeout"]))#012 File
"/usr/lib64/python2.7/smtplib.py", line 255, in __init__#012 (code, msg) =
self.connect(host, po
rt)#012 File "/usr/lib64/python2.7/smtplib.py", line 315, in connect#012
self.sock = self._get_socket(host, port, self.timeout)#012 File "/usr/l
ib64/python2.7/smtplib.py", line 290, in _get_socket#012 return
socket.create_connection((host, port), timeout)#012 File "/usr/lib64/python2.7/so
cket.py", line 571, in create_connection#012 raise err#012error: [Errno 113] No
route to host
Sep 1 09:49:15 ov-no2 journal: ovirt-ha-broker
ovirt_hosted_engine_ha.broker.notifications.Notifications ERROR [Errno 113] No route to
host#012Trace
back (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/notifications.py",
line 26, in send_email#012
timeout=float(cfg["smtp-timeout"]))#012 File
"/usr/lib64/python2.7/smtplib.py", line 255, in __init__#012 (code, msg) =
self.connect(host, po
rt)#012 File "/usr/lib64/python2.7/smtplib.py", line 315, in connect#012
self.sock = self._get_socket(host, port, self.timeout)#012 File "/usr/l
ib64/python2.7/smtplib.py", line 290, in _get_socket#012 return
socket.create_connection((host, port), timeout)#012 File "/usr/lib64/python2.7/so
cket.py", line 571, in create_connection#012 raise err#012error: [Errno 113] No
route to host
Sep 1 09:49:16 ov-no2 journal: ovirt-ha-broker
ovirt_hosted_engine_ha.broker.notifications.Notifications ERROR [Errno 113] No route to
host#012Trace
back (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/notifications.py",
line 26, in send_email#012
timeout=float(cfg["smtp-timeout"]))#012 File
"/usr/lib64/python2.7/smtplib.py", line 255, in __init__#012 (code, msg) =
self.connect(host, po
rt)#012 File "/usr/lib64/python2.7/smtplib.py", line 315, in connect#012
self.sock = self._get_socket(host, port, self.timeout)#012 File "/usr/l
ib64/python2.7/smtplib.py", line 290, in _get_socket#012 return
socket.create_connection((host, port), timeout)#012 File "/usr/lib64/python2.7/so
cket.py", line 571, in create_connection#012 raise err#012error: [Errno 113] No
route to host
Sep 1 09:49:17 ov-no2 vdsm[13586]: WARN Attempting to add an existing net user:
ovirtmgmt/baa3d4d1-6918-4447-80fa-64b86dbcb1f9
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -D PREROUTING -i vnet0 -j libvirt-J-vnet0' f
ailed: Illegal target name 'libvirt-J-vnet0'.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -D POSTROUTING -o vnet0 -j libvirt-P-vnet0'
failed: Illegal target name 'libvirt-P-vnet0'.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -L libvirt-J-vnet0' failed: Chain 'libvirt-J
-vnet0' doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -L libvirt-P-vnet0' failed: Chain 'libvirt-P
-vnet0' doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -F libvirt-J-vnet0' failed: Chain 'libvirt-J
-vnet0' doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -X libvirt-J-vnet0' failed: Chain 'libvirt-J
-vnet0' doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -F libvirt-P-vnet0' failed: Chain 'libvirt-P
-vnet0' doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -X libvirt-P-vnet0' failed: Chain 'libvirt-P
-vnet0' doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -F J-vnet0-mac' failed: Chain 'J-vnet0-mac'
doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -X J-vnet0-mac' failed: Chain 'J-vnet0-mac'
doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -F J-vnet0-arp-mac' failed: Chain 'J-vnet0-a
rp-mac' doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ebtables
--concurrent -t nat -X J-vnet0-arp-mac' failed: Chain 'J-vnet0-a
rp-mac' doesn't exist.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -D libvirt-out -m physdev --physdev-is-bridged --physdev-
out vnet0 -g FO-vnet0' failed: iptables v1.4.21: goto 'FO-vnet0' is not a
chain#012#012Try `iptables -h' or 'iptables --help' for more information.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -D libvirt-out -m physdev --physdev-out vnet0 -g FO-vnet0
' failed: iptables v1.4.21: goto 'FO-vnet0' is not a chain#012#012Try
`iptables -h' or 'iptables --help' for more information.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -D libvirt-in -m physdev --physdev-in vnet0 -g FI-vnet0'
failed: iptables v1.4.21: goto 'FI-vnet0' is not a chain#012#012Try `iptables
-h' or 'iptables --help' for more information.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -D libvirt-host-in -m physdev --physdev-in vnet0 -g HI-vn
et0' failed: iptables v1.4.21: goto 'HI-vnet0' is not a chain#012#012Try
`iptables -h' or 'iptables --help' for more information.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -F FO-vnet0' failed: iptables: No chain/target/match by t
hat name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -X FO-vnet0' failed: iptables: No chain/target/match by t
hat name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -F FI-vnet0' failed: iptables: No chain/target/match by t
hat name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -X FI-vnet0' failed: iptables: No chain/target/match by t
hat name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -F HI-vnet0' failed: iptables: No chain/target/match by t
hat name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -X HI-vnet0' failed: iptables: No chain/target/match by t
hat name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -E FP-vnet0 FO-vnet0' failed: iptables: No chain/target/m
atch by that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -E FJ-vnet0 FI-vnet0' failed: iptables: No chain/target/m
atch by that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/iptables
-w2 -w -E HJ-vnet0 HI-vnet0' failed: iptables: No chain/target/m
atch by that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -D libvirt-out -m physdev --physdev-is-bridged --physdev
-out vnet0 -g FO-vnet0' failed: ip6tables v1.4.21: goto 'FO-vnet0' is not a
chain#012#012Try `ip6tables -h' or 'ip6tables --help' for more informatio
n.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -D libvirt-out -m physdev --physdev-out vnet0 -g FO-vnet
0' failed: ip6tables v1.4.21: goto 'FO-vnet0' is not a chain#012#012Try
`ip6tables -h' or 'ip6tables --help' for more information.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -D libvirt-in -m physdev --physdev-in vnet0 -g FI-vnet0'
failed: ip6tables v1.4.21: goto 'FI-vnet0' is not a chain#012#012Try `ip6tables
-h' or 'ip6tables --help' for more information.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -D libvirt-host-in -m physdev --physdev-in vnet0 -g HI-v
net0' failed: ip6tables v1.4.21: goto 'HI-vnet0' is not a chain#012#012Try
`ip6tables -h' or 'ip6tables --help' for more information.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -F FO-vnet0' failed: ip6tables: No chain/target/match by
that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -X FO-vnet0' failed: ip6tables: No chain/target/match by
that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -F FI-vnet0' failed: ip6tables: No chain/target/match by
that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -X FI-vnet0' failed: ip6tables: No chain/target/match by
that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -F HI-vnet0' failed: ip6tables: No chain/target/match by
that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -X HI-vnet0' failed: ip6tables: No chain/target/match by
that name.
Sep 1 09:49:19 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -E FP-vnet0 FO-vnet0' failed: ip6tables: No chain/target
/match by that name.
Sep 1 09:49:20 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -E FJ-vnet0 FI-vnet0' failed: ip6tables: No chain/target
/match by that name.
Sep 1 09:49:20 ov-no2 firewalld[2428]: WARNING: COMMAND_FAILED: '/usr/sbin/ip6tables
-w2 -w -E HJ-vnet0 HI-vnet0' failed: ip6tables: No chain/target
/match by that name.
.
.
.
.
.
Sep 1 09:49:22 ov-no2 vdsm[13586]: WARN File:
/var/lib/libvirt/qemu/channels/baa3d4d1-6918-4447-80fa-64b86dbcb1f9.ovirt-guest-agent.0
already remove
d
Sep 1 09:49:22 ov-no2 vdsm[13586]: WARN Attempting to remove a non existing net user:
ovirtmgmt/baa3d4d1-6918-4447-80fa-64b86dbcb1f9
Sep 1 09:49:22 ov-no2 vdsm[13586]: WARN File:
/var/lib/libvirt/qemu/channels/baa3d4d1-6918-4447-80fa-64b86dbcb1f9.org.qemu.guest_agent.0
already rem
oved
Sep 1 09:49:22 ov-no2 vdsm[13586]: WARN File:
/var/run/ovirt-vmconsole-console/baa3d4d1-6918-4447-80fa-64b86dbcb1f9.sock already
removed
Sep 1 09:49:27 ov-no2 journal: ovirt-ha-broker
ovirt_hosted_engine_ha.broker.notifications.Notifications ERROR [Errno 113] No route to
host#012Trace
back (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/notifications.py",
line 26, in send_email#012
timeout=float(cfg["smtp-timeout"]))#012 File
"/usr/lib64/python2.7/smtplib.py", line 255, in __init__#012 (code, msg) =
self.connect(host, po
rt)#012 File "/usr/lib64/python2.7/smtplib.py", line 315, in connect#012
self.sock = self._get_socket(host, port, self.timeout)#012 File "/usr/l
ib64/python2.7/smtplib.py", line 290, in _get_socket#012 return
socket.create_connection((host, port), timeout)#012 File "/usr/lib64/python2.7/so
cket.py", line 571, in create_connection#012 raise err#012error: [Errno 113] No
route to host
Sep 1 09:49:27 ov-no2 journal: ovirt-ha-broker
ovirt_hosted_engine_ha.broker.notifications.Notifications ERROR [Errno 113] No route to
host#012Trace
back (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/notifications.py",
line 26, in send_email#012
timeout=float(cfg["smtp-timeout"]))#012 File
"/usr/lib64/python2.7/smtplib.py", line 255, in __init__#012 (code, msg) =
self.connect(host, po
rt)#012 File "/usr/lib64/python2.7/smtplib.py", line 315, in connect#012
self.sock = self._get_socket(host, port, self.timeout)#012 File "/usr/l
ib64/python2.7/smtplib.py", line 290, in _get_socket#012 return
socket.create_connection((host, port), timeout)#012 File "/usr/lib64/python2.7/so
cket.py", line 571, in create_connection#012 raise err#012error: [Errno 113] No
route to host
Sep 1 09:49:37 ov-no2 journal: ovirt-ha-broker
ovirt_hosted_engine_ha.broker.notifications.Notifications ERROR [Errno 113] No route to
host#012Trace
back (most recent call last):#012 File
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/notifications.py",
line 26, in send_email#012
timeout=float(cfg["smtp-timeout"]))#012 File
"/usr/lib64/python2.7/smtplib.py", line 255, in __init__#012 (code, msg) =
self.connect(host, po
rt)#012 File "/usr/lib64/python2.7/smtplib.py", line 315, in connect#012
self.sock = self._get_socket(host, port, self.timeout)#012 File "/usr/l
ib64/python2.7/smtplib.py", line 290, in _get_socket#012 return
socket.create_connection((host, port), timeout)#012 File "/usr/lib64/python2.7/so
cket.py", line 571, in create_connection#012 raise err#012error: [Errno 113] No
route to host
Sep 1 10:01:31 ov-no2 vdsm[13586]: WARN Worker blocked: <Worker name=jsonrpc/4 running
<Task <JsonRpcTask {'params': {}, 'jsonrpc': '2.0',
'method':
u'GlusterVolume.geoRepSessionList', 'id':
u'7f800636-f107-4b98-a35f-2175ec72b7b8'} at 0x7f44805aa110> timeout=60,
duration=60.00 at 0x7f44805aad10>
task#=5775347 at 0x7f44d81c4110>, traceback:#012File:
"/usr/lib64/python2.7/threading.py", line 785, in __bootstrap#012
self.__bootstrap_inner()#012
File: "/usr/lib64/python2.7/threading.py", line 812, in __bootstrap_inner#012
self.run()#012File: "/usr/lib64/python2.7/threading.py", line 765, in
run#012 self.__target(*self.__args, **self.__kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/concurrent.py", line 195, in
run#012 re
t = func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 301, in _run#012
self._execute_task()#012File: "/usr/li
b/python2.7/site-packages/vdsm/executor.py", line 315, in _execute_task#012
task()#012File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", lin
e 391, in __call__#012 self._callable()#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 262, in
__call__#012 self._handler(
self._ctx, self._req)#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 305, in
_serveRequest#012 response = self._handle_requ
est(req, ctx)#012File: "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py",
line 345, in _handle_request#012 res = method(**params)#012File: "/
usr/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 194, in _dynamicMethod#012
result = fn(*methodArgs)#012File: "/usr/lib/python2.7/site-pack
ages/vdsm/gluster/apiwrapper.py", line 237, in geoRepSessionList#012
remoteUserName)#012File: "/usr/lib/python2.7/site-packages/vdsm/gluster/api.py"
, line 93, in wrapper#012 rv = func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/gluster/api.py", line 551, in
volumeGeoRepSessi
onList#012 remoteUserName,#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 56, in
__call__#012 return callMethod()#012Fi
le: "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 54, in
<lambda>#012 **kwargs)#012File: "<string>", line 2, in
glusterVolumeGeo
RepStatus#012File: "/usr/lib64/python2.7/multiprocessing/managers.py", line 759,
in _callmethod#012 kind, result = conn.recv()
[root@ov-no2 ~]# xzcat /var/log/vdsm/vdsm.log.57.xz |egrep "(WARN|error)" |more
2020-09-01 09:24:06,889+0300 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:24:16,910+0300 INFO (jsonrpc/6) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:24:16,911+0300 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:24:26,932+0300 INFO (jsonrpc/5) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:24:26,933+0300 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.01 seconds (__init__:312)
2020-09-01 09:24:30,375+0300 WARN (monitor/0a5e552) [storage.PersistentDict] Could not
parse line `
2020-09-01 09:24:30,379+0300 WARN (monitor/e2e7282) [storage.PersistentDict] Could not
parse line `
.
.
.
.
2020-09-01 09:47:37,008+0300 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:47:47,030+0300 INFO (jsonrpc/5) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:47:47,030+0300 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:48:01,712+0300 WARN (jsonrpc/6) [storage.PersistentDict] Could not parse
line `
2020-09-01 09:48:01,743+0300 WARN (jsonrpc/1) [storage.PersistentDict] Could not parse
line `
2020-09-01 09:48:02,220+0300 WARN (jsonrpc/0) [storage.PersistentDict] Could not parse
line `
2020-09-01 09:48:05,192+0300 WARN (periodic/22) [storage.PersistentDict] Could not parse
line `
2020-09-01 09:48:11,745+0300 WARN (monitor/e2e7282) [storage.PersistentDict] Could not
parse line `
2020-09-01 09:48:13,711+0300 INFO (jsonrpc/7) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:48:13,711+0300 INFO (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:48:23,591+0300 INFO (jsonrpc/0) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:48:23,591+0300 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:48:23,731+0300 INFO (jsonrpc/5) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:48:23,731+0300 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:48:33,754+0300 INFO (jsonrpc/4) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:48:33,754+0300 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:48:43,787+0300 INFO (jsonrpc/2) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:48:43,787+0300 INFO (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:48:53,808+0300 INFO (jsonrpc/4) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:48:53,809+0300 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:49:03,830+0300 INFO (jsonrpc/4) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:49:03,831+0300 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:49:13,852+0300 INFO (jsonrpc/0) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:49:13,853+0300 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.01 seconds (__init__:312)
2020-09-01 09:49:16,042+0300 INFO (jsonrpc/4) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:49:16,042+0300 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.01 seconds (__init__:312)
2020-09-01 09:49:16,405+0300 INFO (jsonrpc/0) [api] FINISH getStats error=Virtual machine
does not exist: {'vmId': u'baa3d4d1-6918-4447-80fa-64b86db
cb1f9'} (api:129)
2020-09-01 09:49:16,405+0300 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call
VM.getStats failed (error 1) in 0.00 seconds (__init__:312)
2020-09-01 09:49:16,797+0300 INFO (jsonrpc/7) [api.virt] START
create(vmParams={u'xml': u'<?xml version=\'1.0\'
encoding=\'UTF-8\'?>\n<domain xmlns:
ovirt-tune="http://ovirt.org/vm/tune/1.0"
xmlns:ovirt-vm="http://ovirt.org/vm/1.0"
type="kvm"><name>HostedEngine</name><uuid>baa3d4d1-6918-4447-80fa-
64b86dbcb1f9</uuid><memory>16777216</memory><currentMemory>16777216</currentMemory><iothreads>1</iothreads><maxMemory
slots="16">67108864</maxMemory>
<vcpu current="4">64</vcpu><sysinfo
type="smbios"><system><entry
name="manufacturer">oVirt</entry><entry
name="product">OS-NAME:</entry><entry name="
version">OS-VERSION:</entry><entry
name="serial">HOST-SERIAL:</entry><entry
name="uuid">baa3d4d1-6918-4447-80fa-64b86dbcb1f9</entry></system></sysinf
o><clock offset="variable" adjustment="0"><timer
name="rtc" tickpolicy="catchup"/><timer name="pit"
tickpolicy="delay"/><timer name="hpet" present="n
o"/></clock><features><acpi/></features><cpu
match="exact"><model>Westmere</model><feature
name="pcid" policy="require"/><feature
name="spec-ctrl" po
licy="require"/><feature name="ssbd"
policy="require"/><topology cores="4" threads="1"
sockets="16"/><numa><cell id="0" cpus="0,1,2,3"
memory="167772
16"/></numa></cpu><cputune/><devices><input
type="tablet" bus="usb"/><channel
type="unix"><target type="virtio"
name="ovirt-guest-agent.0"/><source m
ode="bind"
path="/var/lib/libvirt/qemu/channels/baa3d4d1-6918-4447-80fa-64b86dbcb1f9.ovirt-guest-agent.0"/></channel><channel
type="unix"><target typ
e="virtio" name="org.qemu.guest_agent.0"/><source
mode="bind"
path="/var/lib/libvirt/qemu/channels/baa3d4d1-6918-4447-80fa-64b86dbcb1f9.org.qemu.gues
t_agent.0"/></channel><controller type="usb"
model="piix3-uhci" index="0"><address bus="0x00"
domain="0x0000" function="0x2" slot="0x01"
type="pci"/>
</controller><console type="unix"><source
path="/var/run/ovirt-vmconsole-console/baa3d4d1-6918-4447-80fa-64b86dbcb1f9.sock"
mode="bind"/><target type
="serial" port="0"/><alias
name="ua-228f9c56-c195-4f1e-8cd2-5f7ef60a78f1"/></console><memballoon
model="virtio"><stats period="5"/><alias name="ua-42
10968c-acf1-46cc-8a35-6e29e3c58b38"/><address bus="0x00"
domain="0x0000" function="0x0" slot="0x08"
type="pci"/></memballoon><controller type="scsi"
model="virtio-scsi" index="0"><driver
iothread="1"/><alias
name="ua-473cb0f7-09da-4077-a0e2-8eb5a5efdbe4"/><address
bus="0x00" domain="0x0000" functi
on="0x0" slot="0x05"
type="pci"/></controller><sound model="ich6"><alias
name="ua-496f9a72-5484-4dc0-83f4-caa8225100ae"/><address
bus="0x00" domain="
0x0000" function="0x0" slot="0x04"
type="pci"/></sound><graphics type="vnc"
port="-1" autoport="yes" passwd="*****"
passwdValidTo="1970-01-01T00:00:0
1" keymap="en-us"><listen type="network"
network="vdsm-ovirtmgmt"/></graphics><controller
type="ide" index="0"><address bus="0x00"
domain="0x0000" fu
nction="0x1" slot="0x01"
type="pci"/></controller><graphics type="spice"
port="-1" autoport="yes" passwd="*****"
passwdValidTo="1970-01-01T00:00:01"
tlsPort="-1"><channel name="main"
mode="secure"/><channel name="inputs"
mode="secure"/><channel name="cursor"
mode="secure"/><channel name="playback"
mode="secure"/><channel name="record"
mode="secure"/><channel name="display"
mode="secure"/><channel name="smartcard"
mode="secure"/><channel name="
usbredir" mode="secure"/><listen type="network"
network="vdsm-ovirtmgmt"/></graphics><rng
model="virtio"><backend
model="random">/dev/urandom</backen
d><alias
name="ua-9edfa822-3b11-424e-83f8-4846c6221a6e"/></rng><video><model
type="qxl" vram="32768" heads="1" ram="65536"
vgamem="16384"/><alias nam
e="ua-a4ad75db-c91d-45df-a767-33a71009587f"/><address bus="0x00"
domain="0x0000" function="0x0" slot="0x02"
type="pci"/></video><controller type="vir
tio-serial" index="0" ports="16"><alias
name="ua-f60f1742-07fe-464e-a160-1a16c94db61a"/><address
bus="0x00" domain="0x0000" function="0x0" slot="0x06
" type="pci"/></controller><serial
type="unix"><source
path="/var/run/ovirt-vmconsole-console/baa3d4d1-6918-4447-80fa-64b86dbcb1f9.sock"
mode="bind"/
<target port="0"/></serial><channel
type="spicevmc"><target type="virtio"
name="com.redhat.spice.0"/></channel><interface
type="bridge"><model type=
"virtio"/><link
state="up"/><source bridge="ovirtmgmt"/><driver
queues="4" name="vhost"/><alias
name="ua-e97c94c0-ffc2-46dd-b5cd-9d9c123f4db0"/><addr
ess bus="0x00" domain="0x0000" function="0x0"
slot="0x03" type="pci"/><mac
address="00:16:3e:27:6e:ef"/><mtu size="1500"/><filterref
filter="vdsm-no-
mac-spoofing"/><bandwidth/></interface><disk type="file"
device="cdrom" snapshot="no"><driver name="qemu"
type="raw" error_policy="report"/><source f
ile="" startupPolicy="optional"><seclabel model="dac"
type="none" relabel="no"/></source><target
dev="hdc" bus="ide"/><readonly/><alias name="ua-9034
9138-d20e-4306-b0e2-271e6baaad4a"/><address bus="1"
controller="0" unit="0" type="drive"
target="0"/></disk><disk snapshot="no"
type="file" device="d
isk"><target dev="vda" bus="virtio"/><source
file="/rhev/data-center/00000000-0000-0000-0000-000000000000/80f6e393-9718-4738-a14a-64cf43c3d8c2/images
/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7"><seclabel
model="dac" type="none"
relabel="no"/></source><driver name="qe
mu" iothread="1" io="threads" type="raw"
error_policy="stop" cache="none"/><alias
name="ua-d5de54b6-9f8e-4fba-819b-ebf6780757d2"/><address
bus="0x00"
domain="0x0000" function="0x0" slot="0x07"
type="pci"/><serial>d5de54b6-9f8e-4fba-819b-ebf6780757d2</serial></disk><lease><key>a48555f4-be23-4467-8a
54-400ae7baf9d7</key><lockspace>80f6e393-9718-4738-a14a-64cf43c3d8c2</lockspace><target
offset="LEASE-OFFSET:a48555f4-be23-4467-8a54-400ae7baf9d7:80f
6e393-9718-4738-a14a-64cf43c3d8c2"
path="LEASE-PATH:a48555f4-be23-4467-8a54-400ae7baf9d7:80f6e393-9718-4738-a14a-64cf43c3d8c2"/></lease></devices><pm
<suspend-to-disk enabled="no"/><suspend-to-mem
enabled="no"/></pm><os><type arch="x86_64"
machine="pc-i440fx-rhel7.6.0">hvm</type><smbios mode="sysi
nfo"/><bios
useserial="yes"/></os><metadata><ovirt-tune:qos/><ovirt-vm:vm><ovirt-vm:minGuaranteedMemoryMb
type="int">1024</ovirt-vm:minGuaranteedMemo
ryMb><ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion><ovirt-vm:custom/><ovirt-vm:device
mac_address="00:16:3e:27:6e:ef"><ovirt-vm:custom/></ovi
rt-vm:device><ovirt-vm:device devtype="disk"
name="vda"><ovirt-vm:poolID>00000000-0000-0000-0000-000000000000</ovirt-vm:poolID><ovirt-vm:volumeID>a48
555f4-be23-4467-8a54-400ae7baf9d7</ovirt-vm:volumeID><ovirt-vm:shared>exclusive</ovirt-vm:shared><ovirt-vm:imageID>d5de54b6-9f8e-4fba-819b-ebf6780757
d2</ovirt-vm:imageID><ovirt-vm:domainID>80f6e393-9718-4738-a14a-64cf43c3d8c2</ovirt-vm:domainID></ovirt-vm:device><ovirt-vm:launchPaused>false</ovirt
-vm:launchPaused><ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior></ovirt-vm:vm></metadata></domain>'})
from=::1,36826, vmId=baa3d4d1-69
18-4447-80fa-64b86dbcb1f9 (api:48)
[root@ov-no2 ~]# xzcat /var/log/vdsm/supervdsm.log.30.xz |egrep
"(WARN|error|ERR)" |more
MainProcess|jsonrpc/4::ERROR::2020-09-01
06:56:30,652::supervdsm_server::103::SuperVdsm.ServerCallback::(wrapper) Error in
volumeGeoRepStatus
MainProcess|jsonrpc/3::WARNING::2020-09-01
07:32:27,206::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/3::WARNING::2020-09-01
07:32:27,213::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/3::WARNING::2020-09-01
07:32:27,217::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/3::WARNING::2020-09-01
07:32:27,565::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/3::WARNING::2020-09-01
07:32:27,571::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/3::WARNING::2020-09-01
07:32:27,573::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,735::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,735::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,736::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,736::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,737::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,738::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,740::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,740::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/3::WARNING::2020-09-01 07:32:27,741::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/3::ERROR::2020-09-01 07:32:31,889::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/3::ERROR::2020-09-01 07:32:31,893::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/3::ERROR::2020-09-01 07:32:31,894::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/3::ERROR::2020-09-01 07:32:31,896::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/7::ERROR::2020-09-01
07:58:30,879::supervdsm_server::103::SuperVdsm.ServerCallback::(wrapper) Error in
volumeGeoRepStatus
[root@ov-no2 ~]# xzcat /var/log/vdsm/supervdsm.log.29.xz |egrep
"(WARN|error|ERR)" |more
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,179::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,180::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,181::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,390::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,393::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01
09:32:32,395::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,510::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,510::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,511::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,511::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,512::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,512::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,513::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,513::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-01 09:32:32,514::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::ERROR::2020-09-01 09:32:36,440::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-01 09:32:36,444::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-01 09:32:36,446::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-01 09:32:36,448::udev::53::blivet::(get_device) No
device at ''
[root@ov-no2 ~]# cat /var/log/vdsm/supervdsm.log |egrep "(WARN|error|ERR)"
|more
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,056::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,057::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,059::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,177::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,179::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,180::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,292::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,292::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,293::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,294::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,294::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,295::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,295::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,296::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,296::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::ERROR::2020-09-03 19:34:59,217::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-03 19:34:59,220::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-03 19:34:59,222::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-03 19:34:59,223::udev::53::blivet::(get_device) No
device at ''
[root@ov-no2 ~]# cat /var/log/ovirt-hosted-engine-ha/agent.log.2020-09-01 |egrep -i
"(WARN|ERR)" |more
MainThread::ERROR::2020-09-01
09:48:02,766::hosted_engine::452::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Unhandled
monitoring loop exception
error: [Errno 2] No such file or directory
MainThread::ERROR::2020-09-01
09:48:02,786::agent::144::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Traceback
(most recent call last):
error: [Errno 2] No such file or directory
MainThread::ERROR::2020-09-01
09:48:02,786::agent::145::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Trying to
restart agent
MainThread::INFO::2020-09-01
09:49:16,925::hosted_engine::854::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_start_engine_vm)
stderr: Co
mmand VM.getStats with args {'vmID':
'baa3d4d1-6918-4447-80fa-64b86dbcb1f9'} failed:
MainThread::INFO::2020-09-01
09:49:27,616::hosted_engine::931::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_stop_engine_vm)
stderr: Com
mand VM.destroy with args {'vmID': 'baa3d4d1-6918-4447-80fa-64b86dbcb1f9'}
failed:
MainThread::ERROR::2020-09-01
09:49:27,616::hosted_engine::936::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_stop_engine_vm)
Failed to
stop engine vm with /usr/sbin/hosted-engine --vm-poweroff: Command VM.destroy with args
{'vmID': 'baa3d4d1-6918-4447-80fa-64b86dbcb1f9'} failed:
MainThread::ERROR::2020-09-01
09:49:27,616::hosted_engine::942::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_stop_engine_vm)
Failed to
stop engine VM: Command VM.destroy with args {'vmID':
'baa3d4d1-6918-4447-80fa-64b86dbcb1f9'} failed:
[root@ov-no2 ~]# cat /var/log/ovirt-hosted-engine-ha/broker.log.2020-09-01 |egrep -i
"(WARN|ERR)" |more
StatusStorageThread::ERROR::2020-09-01
09:47:52,786::status_broker::90::ovirt_hosted_engine_ha.broker.status_broker.StatusBroker.Update::(run)
Failed
to update state.
raise ex.HostIdNotLockedError("Host id is not set")
HostIdNotLockedError: Host id is not set
StatusStorageThread::ERROR::2020-09-01
09:47:52,787::status_broker::70::ovirt_hosted_engine_ha.broker.status_broker.StatusBroker.Update::(trigger_res
tart) Trying to restart the broker
Listener::ERROR::2020-09-01
09:48:15,543::notifications::39::ovirt_hosted_engine_ha.broker.notifications.Notifications::(send_email)
[Errno 113] No r
oute to host
raise err
error: [Errno 113] No route to host
Listener::ERROR::2020-09-01
09:48:25,768::notifications::39::ovirt_hosted_engine_ha.broker.notifications.Notifications::(send_email)
[Errno 113] No r
oute to host
raise err
error: [Errno 113] No route to host
Listener::ERROR::2020-09-01
09:49:15,838::notifications::39::ovirt_hosted_engine_ha.broker.notifications.Notifications::(send_email)
[Errno 113] No r
oute to host
raise err
The Gluster status and info I could gather for the present are the following:
[root@ov-no2 ~]# gluster volume status
Status of volume: data
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 6070
Brick ov-no2.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 28860
Brick ov-no3.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 5752
Self-heal Daemon on localhost N/A N/A Y 813
Self-heal Daemon on ov-no1.ariadne-t.local N/A N/A Y 6539
Self-heal Daemon on ov-no3.ariadne-t.local N/A N/A Y 19353
Task Status of Volume data
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: engine
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/engine/engine 49153 0 Y 6098
Brick ov-no2.ariadne-t.local:/gluster_brick
s/engine/engine 49156 0 Y 28867
Brick ov-no3.ariadne-t.local:/gluster_brick
s/engine/engine 49153 0 Y 5789
Self-heal Daemon on localhost N/A N/A Y 813
Self-heal Daemon on ov-no1.ariadne-t.local N/A N/A Y 6539
Self-heal Daemon on ov-no3.ariadne-t.local N/A N/A Y 19353
Task Status of Volume engine
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: vmstore
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49154 0 Y 6102
Brick ov-no2.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49157 0 Y 28878
Brick ov-no3.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49154 0 Y 5772
Self-heal Daemon on localhost N/A N/A Y 813
Self-heal Daemon on ov-no1.ariadne-t.local N/A N/A Y 6539
Self-heal Daemon on ov-no3.ariadne-t.local N/A N/A Y 19353
Task Status of Volume vmstore
------------------------------------------------------------------------------
There are no active volume tasks
[root@ov-no2 ~]# gluster volume heal engine info
Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
[root@ov-no2 ~]# gluster volume heal engine info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no2 ~]# gluster volume heal data info
Brick ov-no1.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
[root@ov-no2 ~]# gluster volume heal data info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no2 ~]# gluster volume heal vmstore info
Brick ov-no1.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
[root@ov-no2 ~]# gluster volume heal vmstore info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no2 ~]# ps -A |grep gluster
813 ? 02:18:00 glusterfs
14183 ? 1-12:02:26 glusterfs
17944 ? 01:19:03 glustereventsd
17995 ? 00:31:50 glustereventsd
28722 ? 00:02:09 glusterd
28860 ? 02:19:48 glusterfsd
28867 ? 05:23:00 glusterfsd
28878 ? 00:20:15 glusterfsd
32187 ? 1-14:42:54 glusterfs
32309 ? 09:59:54 glusterfs
Node3:
[root@ov-no3 ~]# virsh qemu-agent-command HostedEngine
'{"execute":"guest-info"}'
Please enter your authentication name: root
Please enter your password:
{"return":{"version":"2.12.0","supported_commands":[{"enabled":true,"name":"guest-get-osinfo","success-response":true},{"enabled":true,"name":"guest-get-timezone","success-response":true},{"enabled":true,"name":"guest-get-users","success-response":true},{"enabled":true,"name":"guest-get-host-name","success-response":true},{"enabled":false,"name":"guest-exec","success-response":true},{"enabled":false,"name":"guest-exec-status","success-response":true},{"enabled":true,"name":"guest-get-memory-block-info","success-response":true},{"enabled":true,"name":"guest-set-memory-blocks","success-response":true},{"enabled":true,"name":"guest-get-memory-blocks","success-response":true},{"enabled":true,"name":"guest-set-user-password","success-response":true},{"enabled":true,"name":"guest-get-fsinfo","success-response":true},{"enabled":true,"name":"guest-set-vcpus","success-response":true},{"enabled":true,"name":"guest-get-vcpus","success-response":true},{"enabled":true,"name":"guest-network-get-in
terfaces","success-response":true},{"enabled":true,"name":"guest-suspend-hybrid","success-response":false},{"enabled":true,"name":"guest-suspend-ram","success-response":false},{"enabled":true,"name":"guest-suspend-disk","success-response":false},{"enabled":true,"name":"guest-fstrim","success-response":true},{"enabled":true,"name":"guest-fsfreeze-thaw","success-response":true},{"enabled":true,"name":"guest-fsfreeze-freeze-list","success-response":true},{"enabled":true,"name":"guest-fsfreeze-freeze","success-response":true},{"enabled":true,"name":"guest-fsfreeze-status","success-response":true},{"enabled":false,"name":"guest-file-flush","success-response":true},{"enabled":false,"name":"guest-file-seek","success-response":true},{"enabled":false,"name":"guest-file-write","success-response":true},{"enabled":false,"name":"guest-file-read","success-response":true},{"enabled":false,"name":"guest-file-close","success-response":true},{"enabled":false,"name":"guest-file-open","success-response
":true},{"enabled":true,"name":"guest-shutdown","success-response":false},{"enabled":true,"name":"guest-info","success-response":true},{"enabled":true,"name":"guest-set-time","success-response":true},{"enabled":true,"name":"guest-get-time","success-response":true},{"enabled":true,"name":"guest-ping","success-response":true},{"enabled":true,"name":"guest-sync","success-response":true},{"enabled":true,"name":"guest-sync-delimited","success-response":true}]}}
[root@ov-no3 ~]# cat /var/log/messages|egrep -i "(WARN|ERR)" |more
Sep 1 08:55:17 ov-no3 libvirtd: 2020-09-01 05:55:17.816+0000: 5284: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 08:59:30 ov-no3 vdsm[6112]: WARN Worker blocked: <Worker name=jsonrpc/3 running
<Task <JsonRpcTask {'params': {}, 'jsonrpc': '2.0',
'method':
u'GlusterVolume.geoRepSessionList', 'id':
u'8298ea63-eadc-47b8-b8b4-d54982200010'} at 0x7f3c14739d90> timeout=60,
duration=60.00 at 0x7f3c14739dd0> t
ask#=72655 at 0x7f3c34306550>, traceback:#012File:
"/usr/lib64/python2.7/threading.py", line 785, in __bootstrap#012
self.__bootstrap_inner()#012Fil
e: "/usr/lib64/python2.7/threading.py", line 812, in __bootstrap_inner#012
self.run()#012File: "/usr/lib64/python2.7/threading.py", line 765, in run
#012 self.__target(*self.__args, **self.__kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/concurrent.py", line 195, in
run#012 ret =
func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 301, in _run#012
self._execute_task()#012File: "/usr/lib/p
ython2.7/site-packages/vdsm/executor.py", line 315, in _execute_task#012
task()#012File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", line 3
91, in __call__#012 self._callable()#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 262, in
__call__#012 self._handler(sel
f._ctx, self._req)#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 305, in
_serveRequest#012 response = self._handle_request
(req, ctx)#012File: "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py",
line 345, in _handle_request#012 res = method(**params)#012File: "/usr
/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 194, in _dynamicMethod#012
result = fn(*methodArgs)#012File: "/usr/lib/python2.7/site-package
s/vdsm/gluster/apiwrapper.py", line 237, in geoRepSessionList#012
remoteUserName)#012File: "/usr/lib/python2.7/site-packages/vdsm/gluster/api.py",
l
ine 93, in wrapper#012 rv = func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/gluster/api.py", line 551, in
volumeGeoRepSessionL
ist#012 remoteUserName,#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 56, in
__call__#012 return callMethod()#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 54, in
<lambda>#012 **kwargs)#012File: "<string>", line 2, in
glusterVolumeGeoRep
Status#012File: "/usr/lib64/python2.7/multiprocessing/managers.py", line 759, in
_callmethod#012 kind, result = conn.recv()
Sep 1 09:00:17 ov-no3 libvirtd: 2020-09-01 06:00:17.817+0000: 5281: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:00:30 ov-no3 vdsm[6112]: WARN Worker blocked: <Worker name=jsonrpc/3 running
<Task <JsonRpcTask {'params': {}, 'jsonrpc': '2.0',
'method':
u'GlusterVolume.geoRepSessionList', 'id':
u'8298ea63-eadc-47b8-b8b4-d54982200010'} at 0x7f3c14739d90> timeout=60,
duration=120.00 at 0x7f3c14739dd0>
task#=72655 at 0x7f3c34306550>, traceback:#012File:
"/usr/lib64/python2.7/threading.py", line 785, in __bootstrap#012
self.__bootstrap_inner()#012Fi
le: "/usr/lib64/python2.7/threading.py", line 812, in __bootstrap_inner#012
self.run()#012File: "/usr/lib64/python2.7/threading.py", line 765, in ru
n#012 self.__target(*self.__args, **self.__kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/concurrent.py", line 195, in
run#012 ret
= func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/executor.py", line 301, in _run#012
self._execute_task()#012File: "/usr/lib/
python2.7/site-packages/vdsm/executor.py", line 315, in _execute_task#012
task()#012File: "/usr/lib/python2.7/site-packages/vdsm/executor.py", line
391, in __call__#012 self._callable()#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 262, in
__call__#012 self._handler(se
lf._ctx, self._req)#012File:
"/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py", line 305, in
_serveRequest#012 response = self._handle_reques
t(req, ctx)#012File: "/usr/lib/python2.7/site-packages/yajsonrpc/__init__.py",
line 345, in _handle_request#012 res = method(**params)#012File: "/us
r/lib/python2.7/site-packages/vdsm/rpc/Bridge.py", line 194, in _dynamicMethod#012
result = fn(*methodArgs)#012File: "/usr/lib/python2.7/site-packag
es/vdsm/gluster/apiwrapper.py", line 237, in geoRepSessionList#012
remoteUserName)#012File: "/usr/lib/python2.7/site-packages/vdsm/gluster/api.py",
line 93, in wrapper#012 rv = func(*args, **kwargs)#012File:
"/usr/lib/python2.7/site-packages/vdsm/gluster/api.py", line 551, in
volumeGeoRepSession
List#012 remoteUserName,#012File:
"/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 56, in
__call__#012 return callMethod()#012File
: "/usr/lib/python2.7/site-packages/vdsm/common/supervdsm.py", line 54, in
<lambda>#012 **kwargs)#012File: "<string>", line 2, in
glusterVolumeGeoRe
pStatus#012File: "/usr/lib64/python2.7/multiprocessing/managers.py", line 759,
in _callmethod#012 kind, result = conn.recv()
Sep 1 09:05:17 ov-no3 libvirtd: 2020-09-01 06:05:17.816+0000: 5283: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:10:17 ov-no3 libvirtd: 2020-09-01 06:10:17.823+0000: 5282: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:15:17 ov-no3 libvirtd: 2020-09-01 06:15:17.817+0000: 5281: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:20:17 ov-no3 libvirtd: 2020-09-01 06:20:17.820+0000: 5282: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:25:17 ov-no3 libvirtd: 2020-09-01 06:25:17.818+0000: 5280: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:30:17 ov-no3 libvirtd: 2020-09-01 06:30:17.820+0000: 5282: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:35:17 ov-no3 libvirtd: 2020-09-01 06:35:17.815+0000: 5281: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:40:17 ov-no3 libvirtd: 2020-09-01 06:40:17.823+0000: 5284: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:45:17 ov-no3 libvirtd: 2020-09-01 06:45:17.819+0000: 5280: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:50:17 ov-no3 libvirtd: 2020-09-01 06:50:17.824+0000: 5280: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 09:55:17 ov-no3 libvirtd: 2020-09-01 06:55:17.818+0000: 5284: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 10:00:17 ov-no3 libvirtd: 2020-09-01 07:00:17.822+0000: 5284: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
Sep 1 10:05:17 ov-no3 libvirtd: 2020-09-01 07:05:17.824+0000: 5282: error :
qemuDomainAgentAvailable:9133 : Guest agent is not responding: QEMU gues
t agent is not connected
.
.
.
.
[which continues as so 'till now]
[root@ov-no3 ~]# xzcat /var/log/vdsm/vdsm.log.57.xz |egrep "(WARN|error)" |more
2020-09-01 09:31:35,023+0300 WARN (monitor/0a5e552) [storage.PersistentDict] Could not
parse line ``. (persistent:241)
2020-09-01 09:31:35,026+0300 WARN (monitor/e2e7282) [storage.PersistentDict] Could not
parse line ``. (persistent:241)
2020-09-01 09:35:19,868+0300 WARN (monitor/80f6e39) [storage.PersistentDict] Could not
parse line ``. (persistent:241)
2020-09-01 09:35:22,210+0300 WARN (qgapoller/1) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:35:32,207+0300 WARN (qgapoller/3) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:35:42,213+0300 WARN (qgapoller/1) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:35:52,212+0300 WARN (qgapoller/2) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:36:02,207+0300 WARN (qgapoller/1) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:36:12,215+0300 WARN (qgapoller/2) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:36:35,276+0300 WARN (monitor/0a5e552) [storage.PersistentDict] Could not
parse line ``. (persistent:241)
2020-09-01 09:36:35,281+0300 WARN (monitor/e2e7282) [storage.PersistentDict] Could not
parse line ``. (persistent:241)
2020-09-01 09:40:18,494+0300 WARN (qgapoller/3) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c6803ff50> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:40:18,534+0300 WARN (qgapoller/1) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c6803fe60> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:40:20,244+0300 WARN (monitor/80f6e39) [storage.PersistentDict] Could not
parse line ``. (persistent:241)
2020-09-01 09:40:22,209+0300 WARN (qgapoller/3) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:40:32,207+0300 WARN (qgapoller/1) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:40:42,214+0300 WARN (qgapoller/3) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
2020-09-01 09:40:52,215+0300 WARN (qgapoller/2) [virt.periodic.VmDispatcher] could not
run <function <lambda> at 0x7f3c680470c8> on
['5836fc52-b585-46f8-8780-2cc5accd5847'] (periodic:289)
[root@ov-no3 ~]# xzcat /var/log/vdsm/supervdsm.log.29.xz|egrep "(WARN|ERR)"
|more
MainProcess|jsonrpc/4::WARNING::2020-09-01
09:32:32,165::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/4::WARNING::2020-09-01
09:32:32,166::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/4::WARNING::2020-09-01
09:32:32,168::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/4::WARNING::2020-09-01
09:32:32,285::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/4::WARNING::2020-09-01
09:32:32,287::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/4::WARNING::2020-09-01
09:32:32,288::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,396::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,397::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,397::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,398::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,398::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,399::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,399::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,400::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/4::WARNING::2020-09-01 09:32:32,400::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/4::ERROR::2020-09-01 09:32:36,337::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/4::ERROR::2020-09-01 09:32:36,341::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/4::ERROR::2020-09-01 09:32:36,342::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/4::ERROR::2020-09-01 09:32:36,344::udev::53::blivet::(get_device) No
device at ''
[root@ov-no3 ~]# cat /var/log/vdsm/supervdsm.log|egrep "(WARN|ERR)" |more
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,041::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,043::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,044::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,158::util::196::blivet::(get_sysfs_attr) queue/logical_block_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,160::util::196::blivet::(get_sysfs_attr) queue/optimal_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03
19:34:55,161::util::196::blivet::(get_sysfs_attr) queue/minimum_io_size is not a valid
attribute
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,265::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,266::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,266::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,267::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,267::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,268::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,269::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Read-only locking type set. Write loc
ks are prohibited.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,269::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Recovery of standalone physical volum
es failed.
MainProcess|jsonrpc/1::WARNING::2020-09-03 19:34:55,270::lvm::360::blivet::(pvinfo)
ignoring pvs output line: Cannot process standalone physical vo
lumes
MainProcess|jsonrpc/1::ERROR::2020-09-03 19:34:59,399::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-03 19:34:59,402::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-03 19:34:59,404::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/1::ERROR::2020-09-03 19:34:59,406::udev::53::blivet::(get_device) No
device at ''
MainProcess|jsonrpc/7::ERROR::2020-09-03
19:54:43,983::supervdsm_server::103::SuperVdsm.ServerCallback::(wrapper) Error in
volumeGeoRepStatus
[root@ov-no3 ~]# cat /var/log/ovirt-hosted-engine-ha/agent.log.2020-09-01 |egrep -i
"(WARN|ERR)" |more
[nothing in the output]
[root@ov-no3 ~]# cat /var/log/ovirt-hosted-engine-ha/broker.log.2020-09-01 |egrep -i
"(WARN|ERR)" |more
Thread-1::WARNING::2020-09-02 12:46:59,657::network::92::network.Network::(action) Failed
to verify network status, (4 out of 5)
[root@ov-no3 ~]# gluster volume status
Status of volume: data
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 6070
Brick ov-no2.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 28860
Brick ov-no3.ariadne-t.local:/gluster_brick
s/data/data 49152 0 Y 5752
Self-heal Daemon on localhost N/A N/A Y 19353
Self-heal Daemon on ov-no1.ariadne-t.local N/A N/A Y 6539
Self-heal Daemon on ov-no2.ariadne-t.local N/A N/A Y 813
Task Status of Volume data
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: engine
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/engine/engine 49153 0 Y 6098
Brick ov-no2.ariadne-t.local:/gluster_brick
s/engine/engine 49156 0 Y 28867
Brick ov-no3.ariadne-t.local:/gluster_brick
s/engine/engine 49153 0 Y 5789
Self-heal Daemon on localhost N/A N/A Y 19353
Self-heal Daemon on ov-no1.ariadne-t.local N/A N/A Y 6539
Self-heal Daemon on ov-no2.ariadne-t.local N/A N/A Y 813
Task Status of Volume engine
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: vmstore
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick ov-no1.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49154 0 Y 6102
Brick ov-no2.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49157 0 Y 28878
Brick ov-no3.ariadne-t.local:/gluster_brick
s/vmstore/vmstore 49154 0 Y 5772
Self-heal Daemon on localhost N/A N/A Y 19353
Self-heal Daemon on ov-no1.ariadne-t.local N/A N/A Y 6539
Self-heal Daemon on ov-no2.ariadne-t.local N/A N/A Y 813
Task Status of Volume vmstore
------------------------------------------------------------------------------
There are no active volume tasks
[root@ov-no3 ~]# gluster volume heal engine info
Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0
[root@ov-no3 ~]# gluster volume heal engine info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no3 ~]# gluster volume heal data info
Brick ov-no1.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0
[root@ov-no3 ~]# gluster volume heal data info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/data/data
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no3 ~]# gluster volume heal vmstore info
Brick ov-no1.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Number of entries: 0
[root@ov-no3 ~]# gluster volume heal vmstore info summary
Brick ov-no1.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no2.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
Brick ov-no3.ariadne-t.local:/gluster_bricks/vmstore/vmstore
Status: Connected
Total Number of entries: 0
Number of entries in heal pending: 0
Number of entries in split-brain: 0
Number of entries possibly healing: 0
[root@ov-no3 ~]# ps -A|grep gluster
5143 ? 00:01:31 glustereventsd
5752 ? 02:01:08 glusterfsd
5772 ? 00:17:26 glusterfsd
5789 ? 05:12:24 glusterfsd
5906 ? 00:00:38 glustereventsd
6382 ? 02:37:00 glusterfs
9843 ? 00:02:18 glusterd
10613 ? 00:22:08 glusterfs
10750 ? 00:13:24 glusterfs
19353 ? 01:20:36 glusterfs
Hosted Engine:
[root@ov-eng ~]# grep ERROR /var/log/ovirt-engine/engine.log |more
2020-09-02 03:36:35,199+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler8) [1a87b58
d] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 04:38:35,429+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler2) [15fadfa
2] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 05:40:35,655+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler9) [50fa85f
0] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 06:42:35,885+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler7) [3092fdc
c] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 07:44:36,119+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler6) [696ffca
7] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 08:46:36,341+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler9) [180bb50
] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 09:48:36,562+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler6) [3a45e1b
3] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 10:50:36,788+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler4) [7da1197
3] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 11:52:37,011+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler6) [76b3cb6
5] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 12:54:37,234+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler5) [6a0d904
9] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 13:56:37,459+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler4) [1e25ecf
f] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
2020-09-02 14:58:37,674+03 ERROR [org.ovirt.engine.core.bll.gluster.GlusterGeoRepSyncJob]
(DefaultQuartzScheduler1) [64144e3
4] VDS error Command execution failed: rc=1 out='Error : Request timed
out\ngeo-replication command failed\n' err=''
How can I find why the HE failover happened? I believe it's the gluster geo
replication, but I haven't configured geo replication in the enviroment. Can you help
me how to proceed?
Thank you all in advance,
Maria Souvalioti
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IZOXEF6MGTZ...