Re: ovirt 4.3.7 geo-replication not working

Hi Adrian, Have you checked the passwordless rsync between master and slave volume nodes ? Best Regards, Strahil NikolovOn Dec 11, 2019 22:36, adrianquintero@gmail.com wrote:
Hi, I am trying to setup geo-replication between 2 sites, but I keep getting: [root@host1 ~]# gluster vol geo-rep geo-master slave1.mydomain2.com::geo-slave status
MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- host1.mydomain1.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A host2.mydomain2.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A vmm11.virt.iad3p geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A
oVirt GUI has an icon in the volume that says "volume data is being geo-replicated" but we know that is not the case From the logs i can see: [2019-12-11 19:57:48.441557] I [fuse-bridge.c:6810:fini] 0-fuse: Unmounting '/tmp/gsyncd-aux-mount-5WaCmt'. [2019-12-11 19:57:48.441578] I [fuse-bridge.c:6815:fini] 0-fuse: Closing fuse connection to '/tmp/gsyncd-aux-mount-5WaCmt'
and [2019-12-11 19:45:14.785758] I [monitor(monitor):278:monitor] Monitor: worker died in startup phase brick=/gluster_bricks/geo-master/geo-master
thoughts?
thanks,
Adrian _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/TPTAODQ3Q4ZDKJ...

+Sunny Kumar <sunkumar@redhat.com> On Thu, Dec 12, 2019 at 6:33 AM Strahil <hunter86_bg@yahoo.com> wrote:
Hi Adrian,
Have you checked the passwordless rsync between master and slave volume nodes ?
Best Regards, Strahil NikolovOn Dec 11, 2019 22:36, adrianquintero@gmail.com wrote:
Hi, I am trying to setup geo-replication between 2 sites, but I keep
getting:
[root@host1 ~]# gluster vol geo-rep geo-master slave1.mydomain2.com::geo-slave status
MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED
--------------------------------------------------------------------------------------------------------------------------------------------------------------------------
host1.mydomain1.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A host2.mydomain2.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A vmm11.virt.iad3p geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A
oVirt GUI has an icon in the volume that says "volume data is being geo-replicated" but we know that is not the case From the logs i can see: [2019-12-11 19:57:48.441557] I [fuse-bridge.c:6810:fini] 0-fuse: Unmounting '/tmp/gsyncd-aux-mount-5WaCmt'. [2019-12-11 19:57:48.441578] I [fuse-bridge.c:6815:fini] 0-fuse: Closing fuse connection to '/tmp/gsyncd-aux-mount-5WaCmt'
and [2019-12-11 19:45:14.785758] I [monitor(monitor):278:monitor] Monitor: worker died in startup phase brick=/gluster_bricks/geo-master/geo-master
thoughts?
thanks,
Adrian _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/TPTAODQ3Q4ZDKJ...
Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZAN3VFGL347RJZ...

Hi Adrian, If possible please share geo-rep logs, it will help to root cause. /sunny On Thu, Dec 12, 2019 at 5:43 AM Sahina Bose <sabose@redhat.com> wrote:
+Sunny Kumar
On Thu, Dec 12, 2019 at 6:33 AM Strahil <hunter86_bg@yahoo.com> wrote:
Hi Adrian,
Have you checked the passwordless rsync between master and slave volume nodes ?
Best Regards, Strahil NikolovOn Dec 11, 2019 22:36, adrianquintero@gmail.com wrote:
Hi, I am trying to setup geo-replication between 2 sites, but I keep getting: [root@host1 ~]# gluster vol geo-rep geo-master slave1.mydomain2.com::geo-slave status
MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- host1.mydomain1.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A host2.mydomain2.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A vmm11.virt.iad3p geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A
oVirt GUI has an icon in the volume that says "volume data is being geo-replicated" but we know that is not the case From the logs i can see: [2019-12-11 19:57:48.441557] I [fuse-bridge.c:6810:fini] 0-fuse: Unmounting '/tmp/gsyncd-aux-mount-5WaCmt'. [2019-12-11 19:57:48.441578] I [fuse-bridge.c:6815:fini] 0-fuse: Closing fuse connection to '/tmp/gsyncd-aux-mount-5WaCmt'
and [2019-12-11 19:45:14.785758] I [monitor(monitor):278:monitor] Monitor: worker died in startup phase brick=/gluster_bricks/geo-master/geo-master
thoughts?
thanks,
Adrian _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/TPTAODQ3Q4ZDKJ...
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZAN3VFGL347RJZ...

Hi Sunny, Thanks for replying, the issue was solved and added the comments to the thread https://lists.ovirt.org/archives/list/users@ovirt.org/thread/ZAN3VFGL347RJZS... really appreciate you looking into it. regards, Adrian On Thu, Dec 12, 2019 at 4:50 AM Sunny Kumar <sunkumar@redhat.com> wrote:
Hi Adrian,
If possible please share geo-rep logs, it will help to root cause.
/sunny
On Thu, Dec 12, 2019 at 5:43 AM Sahina Bose <sabose@redhat.com> wrote:
+Sunny Kumar
On Thu, Dec 12, 2019 at 6:33 AM Strahil <hunter86_bg@yahoo.com> wrote:
Hi Adrian,
Have you checked the passwordless rsync between master and slave
volume nodes ?
Best Regards, Strahil NikolovOn Dec 11, 2019 22:36, adrianquintero@gmail.com wrote:
Hi, I am trying to setup geo-replication between 2 sites, but I keep
getting:
[root@host1 ~]# gluster vol geo-rep geo-master slave1.mydomain2.com::geo-slave status
MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED
host1.mydomain1.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A host2.mydomain2.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A vmm11.virt.iad3p geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave N/A Faulty N/A N/A
oVirt GUI has an icon in the volume that says "volume data is being geo-replicated" but we know that is not the case From the logs i can see: [2019-12-11 19:57:48.441557] I [fuse-bridge.c:6810:fini] 0-fuse: Unmounting '/tmp/gsyncd-aux-mount-5WaCmt'. [2019-12-11 19:57:48.441578] I [fuse-bridge.c:6815:fini] 0-fuse: Closing fuse connection to '/tmp/gsyncd-aux-mount-5WaCmt'
and [2019-12-11 19:45:14.785758] I [monitor(monitor):278:monitor] Monitor: worker died in startup phase brick=/gluster_bricks/geo-master/geo-master
thoughts?
thanks,
Adrian _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/TPTAODQ3Q4ZDKJ...
Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZAN3VFGL347RJZ...
-- Adrian Quintero

Hi Sahina/Strahil, We followed the recommended setup from gluster documentation however one of my colleagues noticed a python entry in the logs, turns out it is a missing sym link to a library We created the following symlink to all the master servers (cluster 1 oVirt 1) and slave servers (Cluster 2, oVirt2) and geo-sync started working: /lib64/libgfchangelog.so -> /lib64/libgfchangelog.so.0 -------------------------------------------------------------------------------------------------------------------------------------------------------- MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- host1.mydomain1.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave slave1.mydomain2.com Active Changelog Crawl 2019-12-12 05:22:56 host2.mydomain1.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave slave2.mydomain2.com Passive N/A N/A host3.mydomain1.com geo-master /gluster_bricks/geo-master/geo-master root slave1.mydomain2.com::geo-slave slave3.mydomain2.com Passive N/A N/A -------------------------------------------------------------------------------------------------------------------------------------------------------- we still require a bit more testing but at least it is syncing now. I am trying to find good documentation on how to achieve geo-replication for oVirt, is that something you can point me to? basically looking for a way to do Geo-replication from site A to Site B, but the Geo-Replication pop up window from ovirt does not seem to have the functionality to connect to a slave server from another oVirt setup.... As a side note, from the oVirt WEB UI the "cancel button" for the "New Geo-Replication" does not seem to work: storage > volumes > "select your volume" > "click 'Geo-Replication' Any good documentation you can point me to is welcome. thank you for the swift assistance. Regards, Adrian

Forgot to add the log entry that lead us to the solution for our particular case: Log = /var/log/glusterfs/geo-replication/geo-master_slave1.mydomain2.com_geo-slave/gsyncd.log --------------------------------------------------------------------------------- [2019-12-11 20:37:27.831976] E [syncdutils(worker /gluster_bricks/geo-master/geo-master):339:log_raise_exception] <top>: FAIL: Traceback (most recent call last): File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 330, in main func(args) File "/usr/libexec/glusterfs/python/syncdaemon/subcmds.py", line 82, in subcmd_worker local.service_loop(remote) File "/usr/libexec/glusterfs/python/syncdaemon/resource.py", line 1267, in service_loop changelog_agent.init() File "/usr/libexec/glusterfs/python/syncdaemon/repce.py", line 233, in __call__ return self.ins(self.meth, *a) File "/usr/libexec/glusterfs/python/syncdaemon/repce.py", line 215, in __call__ raise res OSError: libgfchangelog.so: cannot open shared object file: No such file or directory ---------------------------------------------------------------------------------
participants (5)
-
Adrian Quintero
-
adrianquintero@gmail.com
-
Sahina Bose
-
Strahil
-
Sunny Kumar