[ovirt-users] Importing gluster SD from geo-replicated volume
Sahina Bose
sabose at redhat.com
Fri Nov 20 05:33:44 UTC 2015
On 11/19/2015 11:21 PM, Ivan Bulatovic wrote:
> Hi, I have two DC's (both initialized), two-node each, and on the
> first one I have a replica 2 gluster storage domain that is
> geo-replicating on a replica 2 slave volume on the second DC (managed
> within the same engine). When I stop the replication (volumes are
> synced) and try to import the gluster storage domain that resides on
> the slave, import storage domain dialog throws a general exception.
>
> Exception is raised when vdsm loads the list of backup servers so that
> the backup-volfile-servers mount option could get populated. If I
> override that in storageServer.py, so that it always return blank, or
> when I manually enter this option in the import storage domain dialog,
> then everything works as expected.
This problem is addressed in Ala's patch -
https://gerrit.ovirt.org/#/c/48308/
Are there multiple interfaces configured for gluster at the slave cluster ?
>
> Maybe it's worth mentioning that I have a dedicated gluster network
> and hostnames for all nodes in both DC's (node hostname, and hostname
> I use for gluster on that node are different), and that all attempts
> to import a storage domain were on the second DC.
>
> Btw, setting up gluster geo-replication from oVirt was a breeze, easy
> and straightforward. Importing domain based on slave gluster volume
> works when gluster storage domain that resides on master volume gets
> removed from the first DC. This is something that we could improve, if
> I don't detach and remove original gluster sd, import storage dialog
> just shows up again after a short "running circle", but it should
> provide a warning that there is another storage domain already
> active/registered in the engine with the same ID/name and that the
> domain should be removed (or the engine can do it for us). I get this
> warning only when I've already removed storage domain on a master
> volume from the first DC (which doesn't make sense to me).
Glad to know geo-rep setup was easy.
Regarding the import of storage domain, Nir could help
>
> I can open bug reports for both issues if needed, just want to check
> if the rationale behind the process is correct or not.
>
> vdsm-gluster-4.17.10.1-0.el7
> ovirt-engine-webadmin-portal-3.6.1-0.0.master.20151117185807.git529d3d2.el7
>
>
> engine.log
>
> 2015-11-19 07:33:15,245 ERROR
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (default task-23) [34886be8] Correlation ID: null, Call Stack: null,
> Custom Event ID: -1, Message: The error message for connection
> hostname:/volname returned by VDSM was: General Exception
> 2015-11-19 07:33:15,245 ERROR
> [org.ovirt.engine.core.bll.storage.BaseFsStorageHelper] (default
> task-23) [34886be8] The connection with details 'hostname:/volname'
> failed because of error code '100' and error message is: general
> exception
>
> vdsm.log
>
> Thread-38::ERROR::2015-11-19
> 07:33:15,237::hsm::2465::Storage.HSM::(connectStorageServer) Could not
> connect to storageServer
> Traceback (most recent call last):
> File "/usr/share/vdsm/storage/hsm.py", line 2462, in
> connectStorageServer
> conObj.connect()
> File "/usr/share/vdsm/storage/storageServer.py", line 224, in connect
> self._mount.mount(self.options, self._vfsType, cgroup=self.CGROUP)
> File "/usr/share/vdsm/storage/storageServer.py", line 324, in options
> backup_servers_option = self._get_backup_servers_option()
> File "/usr/share/vdsm/storage/storageServer.py", line 341, in
> _get_backup_servers_option
> servers.remove(self._volfileserver)
> ValueError: list.remove(x): x not in list
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
More information about the Users
mailing list