Hi,
Sorry for the delayed response...
After we migrated the gluster bricks to already upgraded nodes, everything was fine and it
seems the cluster is healthy again.
Thanks for the help!
Regards,
Bertjan
On Fri, Feb 07, 2020 at 03:56:35PM +0200, Strahil Nikolov wrote:
> On February 7, 2020 11:28:17 AM GMT+02:00, "Goorkate, B.J."
<b.j.goorkate(a)umcutrecht.nl> wrote:
> >Hi,
> >
> >There is just one unsynced entry which comes back every time: the
> >dom_md/ids file.
> >
> >When I add/delete a VM it's gone for a short while, but then it
> >reappears.
> >
> >I did a brick replace to a node which is already upgraded. After that
> >I'll do the same with the brick on the remaining not-yet-upgraded node,
> >just to be sure. Luckily we have the hardware for it.
> >
> >We're migrating from 4.2.8 (I think) to 4.3.7.2.
> >
> >Regards,
> >
> >Bertjan
> >
> >
> >On Thu, Feb 06, 2020 at 12:43:48PM +0200, Strahil Nikolov wrote:
> >> On February 6, 2020 9:42:13 AM GMT+02:00, "Goorkate, B.J."
> ><b.j.goorkate(a)umcutrecht.nl> wrote:
> >> >Hi Strahil,
> >> >
> >> >The 0-xlator-message still occurs, but not frequent. Yesterdag a
> >couple
> >> >of times, but the 4th there were no entries at all.
> >> >
> >> >What I did find out, was that the unsynced entries belonged to
> >> >VM-images which were on specific hosts.
> >> >
> >> >Yesterday I migrated them all to other hosts and the unsynced
> >entries
> >> >were gone except for 3. After a 'stat' of those
files/directories,
> >they
> >> >were gone too.
> >> >
> >> >I think I can migrate the remaining hosts now. An option would be to
> >> >move the bricks of the not-yet-upgraded hosts to upgraded hosts. I
> >have
> >> >spare disks. What do you think?
> >> >
> >> >Regards,
> >> >
> >> >Bertjan
> >> >
> >> >
> >> >
> >> >
> >> >On Tue, Feb 04, 2020 at 06:40:21PM +0000, Strahil Nikolov wrote:
> >> >> Did you manage to fix the issue with the 0-xlator ? If yes,
> >most
> >> >probably
> >> >> the issue will be OK. Yet 'probably' doesn't meant
that they
> >> >'will' be OK.
> >> >> If it was my lab - I would go ahead only if the 0-xlator issue
> >is
> >> >> over.Yet, a lab is a different thing than prod - so it is your
> >> >sole
> >> >> decision.
> >> >> Did you test the upgrade prior moving to Prod ?
> >> >> About the hooks - I had such issue before and I had to
> >reinstall
> >> >the
> >> >> gluster rpms to solve it.
> >> >> Best Regards,
> >> >> Strahil Nikolov
> >> >> В вторник, 4 февруари 2020 г., 16:35:27 ч. Гринуич+2, Goorkate,
> >> >B.J.
> >> >> <b.j.goorkate(a)umcutrecht.nl> написа:
> >> >> Hi Strahil,
> >> >>
> >> >> Thanks for your time so far!
> >> >>
> >> >> The packages seem fine on all of the 3 nodes.
> >> >> Only /var/lib/glusterd/glusterd.info is modified and on the not
> >> >yet
> >> >> upgraded nodes these files are missing:
> >> >>
> >> >> missing /var/lib/glusterd/hooks/1/set/post/S30samba-set.sh
> >> >> missing /var/lib/glusterd/hooks/1/start/post/S29CTDBsetup.sh
> >> >> missing
> >/var/lib/glusterd/hooks/1/start/post/S30samba-start.sh
> >> >> missing
> >/var/lib/glusterd/hooks/1/stop/pre/S29CTDB-teardown.sh
> >> >> missing /var/lib/glusterd/hooks/1/stop/pre/S30samba-stop.sh
> >> >>
> >> >> But that doesn't seem relevant...
> >> >>
> >> >> I stopped a couple of virtual machines with image files with
> >> >unsynched
> >> >> entries. When they were turned off, I couldn't find them
> >anymore
> >> >in the
> >> >> unsyced entries list. When I turned them on again, some
> >> >re-appeared, some
> >> >> didn't.
> >> >>
> >> >> I really don't know where to look next.
> >> >>
> >> >> The big question is: will the problems be resolved when I
> >upgrade
> >> >the two
> >> >> remaining nodes, or will it get worse?
> >> >>
> >> >> Regards,
> >> >>
> >> >> Bertjan
> >> >>
> >> >> On Sun, Feb 02, 2020 at 08:06:58PM +0000, Strahil Nikolov
> >wrote:
> >> >> > It seems that something went "off".
> >> >> > That '0-xlator: dlsym(xlator_api) missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api' is really worrisome.
> >> >> > I think that you might have a bad package and based on
the
> >> >info it
> >> >> could
> >> >> > be glusterfs-libs which should provide
> >> >/usr/lib64/libgfrpc.so.0 . I'm
> >> >> > currently on gluster v7.0 and I can't check it on
my
> >> >installation.
> >> >> > Run a for loop to check the rpms:
> >> >> > for i in $(rpm -qa | grep gluster) ; do echo "$i
:" ; rpm
> >-V
> >> >$i';
> >> >> > echo;echo; done
> >> >> > Most probably you can safely redo (on the upgraded
node)
> >the
> >> >last yum
> >> >> > transaction:
> >> >> > yum history
> >> >> > yum history info <ID> -> verify the gluster
packages were
> >> >installed
> >> >> here
> >> >> > yum history redo <ID>
> >> >> > If kernel, glibc, systemd were not in this transaction
,
> >you
> >> >can stop
> >> >> > gluster and start it again:
> >> >> > Node in maintenance in oVirt
> >> >> > systemctl stop ovirt-ha-agent ovirt-ha-broker vdsmd
> >> >supervdsmd
> >> >> > systemctl stop sanlock
> >> >> > systemctl stop glusterd
> >> >> >
/usr/share/glusterfs/scripts/stop-all-gluster-processes.sh
> >> >> > And then power up again:
> >> >> > systemctl start glusterd
> >> >> > gluster volume status -> check all connected
> >> >> > systemctl start sanlock supervdsmd vdsmd
> >> >> > systemctl start ovirt-ha-broker ovirt-ha-agent
> >> >> > Check situation.
> >> >> > Yet, you need to make gluster stop complaining , before
> >you
> >> >can take
> >> >> care
> >> >> > of the heal. Usually 'rsync' is my best friend -
but this
> >is
> >> >when
> >> >> gluster
> >> >> > is working normally - and your case is far from normal.
> >> >> > If redo doesn't work for you -> try the "yum
history
> >> >rollback" to
> >> >> recover
> >> >> > to last good state.
> >> >> > I think that 'BOOM Boot Manager' is best in such
cases.
> >> >> > Note: Never take any of my words for granted. I'm
not
> >running
> >> >oVirt
> >> >> in
> >> >> > production and some of my methods might not be OK for
your
> >> >> environment.
> >> >> > Best Regards,
> >> >> > Strahil Nikolov
> >> >> > В неделя, 2 февруари 2020 г., 08:56:01 ч. Гринуич+2,
Mark
> >> >Lamers
> >> >> > <mark.r.lamers(a)gmail.com> написа:
> >> >> >
> >> >> > Hi Strahil,
> >> >> >
> >> >> > Bertjan is not in the office today, so I will reply if
> >okay
> >> >with you.
> >> >> >
> >> >> > First I like to describe the status of our network
> >> >> >
> >> >> > There are three bricks:
> >> >> >
> >> >> > and 5 nodes:
> >> >> >
> >> >> > Every host has a management and migrate vlan iface on
a
> >> >different
> >> >> bond
> >> >> > iface.
> >> >> >
> >> >> > The last octet of the ipaddress is similar
> >> >> >
> >> >> > The output from 'gluster volume heal <volname>
info' gives
> >a
> >> >long
> >> >> list of
> >> >> > shards from all three nodes, the file is attached as
> >> >> > 'gluster_volume_heal_info.txt'. The node without
shards in
> >> >the list
> >> >> is the
> >> >> > updated node to gluster6, oVirt 4.3. That is curious i
> >think.
> >> >> >
> >> >> > Further more I find the following errors ' E '
in the
> >> >glusterd.log
> >> >> of the
> >> >> > upgrated host:
> >> >> >
> >> >> > [2020-01-26 17:40:46.147730] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-26 22:47:16.655651] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-27 07:07:51.815490] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-27 18:28:14.953974] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-27 18:58:22.629457] E [MSGID: 106118]
> >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-27 18:58:22.629595] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-27 18:58:22.756430] E [MSGID: 106117]
> >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock]
> >0-management:
> >> >Unable to
> >> >> > release lock for vmstore1
> >> >> > [2020-01-27 18:58:22.756581] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-28 05:31:52.427196] E [MSGID: 106118]
> >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 05:31:52.427315] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-28 05:31:52.537799] E [MSGID: 106115]
> >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors]
> >0-management:
> >> >Locking
> >> >> > failed on 10.13.250.14. Please check log file for
details.
> >> >> > [2020-01-28 05:31:52.537973] E [MSGID: 106150]
> >> >> > [glusterd-syncop.c:1931:gd_sync_task_begin]
0-management:
> >> >Locking
> >> >> Peers
> >> >> > Failed.
> >> >> > [2020-01-28 05:31:52.539620] E [MSGID: 106117]
> >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock]
> >0-management:
> >> >Unable to
> >> >> > release lock for vmstore1
> >> >> > [2020-01-28 05:31:52.539759] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-28 05:31:52.539937] E [MSGID: 106115]
> >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors]
> >0-management:
> >> >> Unlocking
> >> >> > failed on 10.13.250.14. Please check log file for
details.
> >> >> > [2020-01-28 05:31:52.540446] E [MSGID: 106151]
> >> >> > [glusterd-syncop.c:1616:gd_unlock_op_phase]
0-management:
> >> >Failed to
> >> >> unlock
> >> >> > on some peer(s)
> >> >> > [2020-01-28 08:51:45.638694] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 08:52:45.709950] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 08:54:12.455555] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> > [2020-01-28 08:54:16.214779] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 08:57:59.496842] E [MSGID: 106118]
> >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 08:57:59.496905] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-28 08:57:59.505119] E
> >> >[rpcsvc.c:1577:rpcsvc_submit_generic]
> >> >> > 0-rpc-service: failed to submit message (XID: 0x2,
> >Program:
> >> >GlusterD
> >> >> svc
> >> >> > cli read-only, ProgVers: 2, Proc: 5) to rpc-transport
> >> >> (socket.management)
> >> >> > [2020-01-28 08:57:59.505135] E [MSGID: 106430]
> >> >> > [glusterd-utils.c:558:glusterd_submit_reply]
0-glusterd:
> >> >Reply
> >> >> submission
> >> >> > failed
> >> >> > [2020-01-28 08:57:59.647456] E [MSGID: 106117]
> >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock]
> >0-management:
> >> >Unable to
> >> >> > release lock for vmstore1
> >> >> > [2020-01-28 08:57:59.647508] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: 1
> >> >> > [2020-01-28 08:57:59.654929] E
> >> >[rpcsvc.c:1577:rpcsvc_submit_generic]
> >> >> > 0-rpc-service: failed to submit message (XID: 0x2,
> >Program:
> >> >GlusterD
> >> >> svc
> >> >> > cli, ProgVers: 2, Proc: 27) to rpc-transport
> >> >(socket.management)
> >> >> > [2020-01-28 08:57:59.654943] E [MSGID: 106430]
> >> >> > [glusterd-utils.c:558:glusterd_submit_reply]
0-glusterd:
> >> >Reply
> >> >> submission
> >> >> > failed
> >> >> > [2020-01-28 09:07:34.941350] E [MSGID: 106118]
> >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 09:07:34.941391] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-28 09:07:35.042466] E [MSGID: 106117]
> >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock]
> >0-management:
> >> >Unable to
> >> >> > release lock for vmstore1
> >> >> > [2020-01-28 09:07:35.042510] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-28 09:13:32.329172] E [MSGID: 106244]
> >> >[glusterd.c:1785:init]
> >> >> > 0-management: creation of 1 listeners failed,
continuing
> >with
> >> >> succeeded
> >> >> > transport
> >> >> > [2020-01-28 09:13:44.024431] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> > [2020-01-28 09:45:46.347499] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 09:46:45.837466] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 09:47:45.976186] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 09:48:24.976568] E
> >> >[rpc-clnt.c:346:saved_frames_unwind]
> >> >> (-->
> >> >> >
> >> >/lib64/libglusterfs.so.0(_gf_log_callingfn+0x13b)[0x7f61c7c07adb]
> >> >> (-->
> >> >> > /lib64/libgfrpc.so.0(+0xd7e4)[0x7f61c79ae7e4] (-->
> >> >> > /lib64/libgfrpc.so.0(+0xd8fe)[0x7f61c79ae8fe] (-->
> >> >> >
> >> >>
> >>
> >>/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x97)[0x7f61c79af987]
> >> >> > (--> /lib64/libgfrpc.so.0(+0xf518)[0x7f61c79b0518]
)))))
> >> >> 0-management:
> >> >> > forced unwinding frame type(glusterd mgmt v3) op(--(1))
> >> >called at
> >> >> > 2020-01-28 09:44:45.743365 (xid=0x52)
> >> >> > [2020-01-28 09:48:24.976640] E [MSGID: 106115]
> >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors]
> >0-management:
> >> >Locking
> >> >> > failed on 10.13.250.11. Please check log file for
details.
> >> >> > [2020-01-28 09:48:24.976743] E [MSGID: 106150]
> >> >> > [glusterd-syncop.c:1931:gd_sync_task_begin]
0-management:
> >> >Locking
> >> >> Peers
> >> >> > Failed.
> >> >> > [2020-01-28 09:48:24.977361] E
> >> >[rpc-clnt.c:346:saved_frames_unwind]
> >> >> (-->
> >> >> >
> >> >/lib64/libglusterfs.so.0(_gf_log_callingfn+0x13b)[0x7f61c7c07adb]
> >> >> (-->
> >> >> > /lib64/libgfrpc.so.0(+0xd7e4)[0x7f61c79ae7e4] (-->
> >> >> > /lib64/libgfrpc.so.0(+0xd8fe)[0x7f61c79ae8fe] (-->
> >> >> >
> >> >>
> >>
> >>/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x97)[0x7f61c79af987]
> >> >> > (--> /lib64/libgfrpc.so.0(+0xf518)[0x7f61c79b0518]
)))))
> >> >> 0-management:
> >> >> > forced unwinding frame type(glusterd mgmt) op(--(3))
> >called
> >> >at
> >> >> 2020-01-28
> >> >> > 09:46:03.586874 (xid=0x53)
> >> >> > [2020-01-28 09:48:24.977417] E [MSGID: 106152]
> >> >> > [glusterd-syncop.c:104:gd_collate_errors] 0-glusterd:
> >Staging
> >> >failed
> >> >> on
> >> >> > 10.13.250.11. Please check log file for details.
> >> >> > [2020-01-28 09:48:24.977631] E
> >> >[rpcsvc.c:1577:rpcsvc_submit_generic]
> >> >> > 0-rpc-service: failed to submit message (XID: 0x2,
> >Program:
> >> >GlusterD
> >> >> svc
> >> >> > cli, ProgVers: 2, Proc: 22) to rpc-transport
> >> >(socket.management)
> >> >> > [2020-01-28 09:48:24.977664] E [MSGID: 106430]
> >> >> > [glusterd-utils.c:558:glusterd_submit_reply]
0-glusterd:
> >> >Reply
> >> >> submission
> >> >> > failed
> >> >> > [2020-01-28 09:48:24.977744] E [MSGID: 106115]
> >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors]
> >0-management:
> >> >> Unlocking
> >> >> > failed on 10.13.250.13. Please check log file for
details.
> >> >> > [2020-01-28 09:48:24.977861] E [MSGID: 106115]
> >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors]
> >0-management:
> >> >> Unlocking
> >> >> > failed on 10.13.250.15. Please check log file for
details.
> >> >> > [2020-01-28 09:48:25.012037] E [MSGID: 106115]
> >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors]
> >0-management:
> >> >> Unlocking
> >> >> > failed on 10.13.250.14. Please check log file for
details.
> >> >> > [2020-01-28 09:48:25.012148] E [MSGID: 106151]
> >> >> > [glusterd-syncop.c:1616:gd_unlock_op_phase]
0-management:
> >> >Failed to
> >> >> unlock
> >> >> > on some peer(s)
> >> >> > [2020-01-28 09:48:25.012315] E [MSGID: 106117]
> >> >> > [glusterd-syncop.c:1640:gd_unlock_op_phase]
0-management:
> >> >Unable to
> >> >> > release lock for vmstore1
> >> >> > [2020-01-28 09:48:25.012425] E
> >> >[rpcsvc.c:1577:rpcsvc_submit_generic]
> >> >> > 0-rpc-service: failed to submit message (XID: 0x2,
> >Program:
> >> >GlusterD
> >> >> svc
> >> >> > cli, ProgVers: 2, Proc: 27) to rpc-transport
> >> >(socket.management)
> >> >> > [2020-01-28 09:48:25.012452] E [MSGID: 106430]
> >> >> > [glusterd-utils.c:558:glusterd_submit_reply]
0-glusterd:
> >> >Reply
> >> >> submission
> >> >> > failed
> >> >> > [2020-01-28 09:50:46.105492] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> > [2020-01-28 10:13:16.199982] E [MSGID: 106118]
> >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 10:13:16.200078] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-28 10:13:16.345918] E [MSGID: 106117]
> >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock]
> >0-management:
> >> >Unable to
> >> >> > release lock for vmstore1
> >> >> > [2020-01-28 10:13:16.346012] E [MSGID: 106376]
> >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management:
> >handler
> >> >> returned: -1
> >> >> > [2020-01-28 12:42:46.212783] E [MSGID: 106118]
> >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin]
0-management:
> >> >Unable to
> >> >> > acquire lock for vmstore1
> >> >> > [2020-01-28 13:25:49.857368] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> > [2020-01-29 13:26:03.031179] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> > [2020-01-30 12:33:18.590063] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> > [2020-01-30 13:34:57.027468] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> > [2020-01-30 15:08:10.814931] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> > [2020-01-31 09:40:22.725825] E [MSGID: 101097]
> >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator:
> >> >dlsym(xlator_api)
> >> >> missing:
> >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so:
> >undefined
> >> >symbol:
> >> >> > xlator_api
> >> >> >
> >> >> > I 'm curious for your opinion of the data sofar.
> >> >> >
> >> >> > Love to here from you (or ather members of the list)
> >> >> >
> >> >> > Regards Mark Lamers
> >> >> >
> >> >> > _______________________________________________
> >> >> > Users mailing list -- users(a)ovirt.org
> >> >> > To unsubscribe send an email to users-leave(a)ovirt.org
> >> >> > Privacy Statement:
> >https://www.ovirt.org/site/privacy-policy/
> >> >> > oVirt Code of Conduct:
> >> >> >
> >https://www.ovirt.org/community/about/community-guidelines/
> >> >> > List Archives:
> >> >> >
> >> >>
> >>
>
>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/VUBWMDUN3GMKGOT63J2Y2A5X7VXXW3H3/
> >> >>
> >> >> > _______________________________________________
> >> >> > Users mailing list -- users(a)ovirt.org
> >> >> > To unsubscribe send an email to users-leave(a)ovirt.org
> >> >> > Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> >> >> > oVirt Code of Conduct:
> >> >>
https://www.ovirt.org/community/about/community-guidelines/
> >> >> > List Archives:
> >> >>
> >>
>
>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/PF2ONI7EHIXWXRBCJIZCVCFHYYKCYZSI/
> >> >>
> >> >>
> >>
>
>>------------------------------------------------------------------------------
> >> >>
> >> >> De informatie opgenomen in dit bericht kan vertrouwelijk zijn
> >en
> >> >is
> >> >> uitsluitend bestemd voor de geadresseerde. Indien u dit bericht
> >> >onterecht
> >> >> ontvangt, wordt u verzocht de inhoud niet te gebruiken en de
> >> >afzender
> >> >> direct
> >> >> te informeren door het bericht te retourneren. Het Universitair
> >> >Medisch
> >> >> Centrum Utrecht is een publiekrechtelijke rechtspersoon in de
> >zin
> >> >van de
> >> >> W.H.W.
> >> >> (Wet Hoger Onderwijs en Wetenschappelijk Onderzoek) en staat
> >> >geregistreerd
> >> >> bij
> >> >> de Kamer van Koophandel voor Midden-Nederland onder nr.
> >30244197.
> >> >>
> >> >> Denk s.v.p aan het milieu voor u deze e-mail afdrukt.
> >> >>
> >> >>
> >>
>
>>------------------------------------------------------------------------------
> >> >>
> >> >> This message may contain confidential information and is
> >intended
> >> >> exclusively
> >> >> for the addressee. If you receive this message unintentionally,
> >> >please do
> >> >> not
> >> >> use the contents but notify the sender immediately by return
> >> >e-mail.
> >> >> University
> >> >> Medical Center Utrecht is a legal person by public law and is
> >> >registered
> >> >> at
> >> >> the Chamber of Commerce for Midden-Nederland under no.
> >30244197.
> >> >>
> >> >> Please consider the environment before printing this e-mail.
> >>
> >> Hi Bertjan,
> >>
> >> If you healed everything - I think there is no need to migrate
> >bricks.
> >> Just a precaution - keep a test VM for testing power off and power
> >on after each node is upgraded.
> >>
> >> What version are you upgrading to ?
> >>
> >> Best Regards,
> >> Strahil Nikolov
> >_______________________________________________
> >Users mailing list -- users(a)ovirt.org
> >To unsubscribe send an email to users-leave(a)ovirt.org
> >Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> >oVirt Code o
>
> You can check on the bricks which has the newest (usually the diff is in the
timestamp inside) and copy that on the rest of the bricks.
> Then a heal can clear the pending heal entry.
>
> Sadly this happens for some time and I never had such issues with oVirt 4.2 on
gluster v3.
>
> Best Regards,
> Strahil Nikolov