[ovirt-users] Importing gluster SD from geo-replicated volume
Ivan Bulatovic
combuster at gmail.com
Thu Nov 19 17:51:26 UTC 2015
Hi, I have two DC's (both initialized), two-node each, and on the first
one I have a replica 2 gluster storage domain that is geo-replicating on
a replica 2 slave volume on the second DC (managed within the same
engine). When I stop the replication (volumes are synced) and try to
import the gluster storage domain that resides on the slave, import
storage domain dialog throws a general exception.
Exception is raised when vdsm loads the list of backup servers so that
the backup-volfile-servers mount option could get populated. If I
override that in storageServer.py, so that it always return blank, or
when I manually enter this option in the import storage domain dialog,
then everything works as expected.
Maybe it's worth mentioning that I have a dedicated gluster network and
hostnames for all nodes in both DC's (node hostname, and hostname I use
for gluster on that node are different), and that all attempts to import
a storage domain were on the second DC.
Btw, setting up gluster geo-replication from oVirt was a breeze, easy
and straightforward. Importing domain based on slave gluster volume
works when gluster storage domain that resides on master volume gets
removed from the first DC. This is something that we could improve, if I
don't detach and remove original gluster sd, import storage dialog just
shows up again after a short "running circle", but it should provide a
warning that there is another storage domain already active/registered
in the engine with the same ID/name and that the domain should be
removed (or the engine can do it for us). I get this warning only when
I've already removed storage domain on a master volume from the first DC
(which doesn't make sense to me).
I can open bug reports for both issues if needed, just want to check if
the rationale behind the process is correct or not.
vdsm-gluster-4.17.10.1-0.el7
ovirt-engine-webadmin-portal-3.6.1-0.0.master.20151117185807.git529d3d2.el7
engine.log
2015-11-19 07:33:15,245 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(default task-23) [34886be8] Correlation ID: null, Call Stack: null,
Custom Event ID: -1, Message: The error message for connection
hostname:/volname returned by VDSM was: General Exception
2015-11-19 07:33:15,245 ERROR
[org.ovirt.engine.core.bll.storage.BaseFsStorageHelper] (default
task-23) [34886be8] The connection with details 'hostname:/volname'
failed because of error code '100' and error message is: general exception
vdsm.log
Thread-38::ERROR::2015-11-19
07:33:15,237::hsm::2465::Storage.HSM::(connectStorageServer) Could not
connect to storageServer
Traceback (most recent call last):
File "/usr/share/vdsm/storage/hsm.py", line 2462, in connectStorageServer
conObj.connect()
File "/usr/share/vdsm/storage/storageServer.py", line 224, in connect
self._mount.mount(self.options, self._vfsType, cgroup=self.CGROUP)
File "/usr/share/vdsm/storage/storageServer.py", line 324, in options
backup_servers_option = self._get_backup_servers_option()
File "/usr/share/vdsm/storage/storageServer.py", line 341, in
_get_backup_servers_option
servers.remove(self._volfileserver)
ValueError: list.remove(x): x not in list
More information about the Users
mailing list