[ovirt-users] Importing gluster SD from geo-replicated volume

Sahina Bose sabose at redhat.com
Fri Nov 20 05:33:44 UTC 2015



On 11/19/2015 11:21 PM, Ivan Bulatovic wrote:
> Hi, I have two DC's (both initialized), two-node each, and on the 
> first one I have a replica 2 gluster storage domain that is 
> geo-replicating on a replica 2 slave volume on the second DC (managed 
> within the same engine). When I stop the replication (volumes are 
> synced) and try to import the gluster storage domain that resides on 
> the slave, import storage domain dialog throws a general exception.
>
> Exception is raised when vdsm loads the list of backup servers so that 
> the backup-volfile-servers mount option could get populated. If I 
> override that in storageServer.py, so that it always return blank, or 
> when I manually enter this option in the import storage domain dialog, 
> then everything works as expected.

This problem is addressed in Ala's patch - 
https://gerrit.ovirt.org/#/c/48308/
Are there multiple interfaces configured for gluster at the slave cluster ?

>
> Maybe it's worth mentioning that I have a dedicated gluster network 
> and hostnames for all nodes in both DC's (node hostname, and hostname 
> I use for gluster on that node are different), and that all attempts 
> to import a storage domain were on the second DC.
>
> Btw, setting up gluster geo-replication from oVirt was a breeze, easy 
> and straightforward. Importing domain based on slave gluster volume 
> works when gluster storage domain that resides on master volume gets 
> removed from the first DC. This is something that we could improve, if 
> I don't detach and remove original gluster sd, import storage dialog 
> just shows up again after a short "running circle", but it should 
> provide a warning that there is another storage domain already 
> active/registered in the engine with the same ID/name and that the 
> domain should be removed (or the engine can do it for us). I get this 
> warning only when I've already removed storage domain on a master 
> volume from the first DC (which doesn't make sense to me).

Glad to know geo-rep setup was easy.
Regarding the import of storage domain, Nir could help


>
> I can open bug reports for both issues if needed, just want to check 
> if the rationale behind the process is correct or not.
>
> vdsm-gluster-4.17.10.1-0.el7
> ovirt-engine-webadmin-portal-3.6.1-0.0.master.20151117185807.git529d3d2.el7 
>
>
> engine.log
>
> 2015-11-19 07:33:15,245 ERROR 
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
> (default task-23) [34886be8] Correlation ID: null, Call Stack: null, 
> Custom Event ID: -1, Message: The error message for connection 
> hostname:/volname returned by VDSM was: General Exception
> 2015-11-19 07:33:15,245 ERROR 
> [org.ovirt.engine.core.bll.storage.BaseFsStorageHelper] (default 
> task-23) [34886be8] The connection with details 'hostname:/volname' 
> failed because of error code '100' and error message is: general 
> exception
>
> vdsm.log
>
> Thread-38::ERROR::2015-11-19 
> 07:33:15,237::hsm::2465::Storage.HSM::(connectStorageServer) Could not 
> connect to storageServer
> Traceback (most recent call last):
>   File "/usr/share/vdsm/storage/hsm.py", line 2462, in 
> connectStorageServer
>     conObj.connect()
>   File "/usr/share/vdsm/storage/storageServer.py", line 224, in connect
>     self._mount.mount(self.options, self._vfsType, cgroup=self.CGROUP)
>   File "/usr/share/vdsm/storage/storageServer.py", line 324, in options
>     backup_servers_option = self._get_backup_servers_option()
>   File "/usr/share/vdsm/storage/storageServer.py", line 341, in 
> _get_backup_servers_option
>     servers.remove(self._volfileserver)
> ValueError: list.remove(x): x not in list
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users




More information about the Users mailing list