On Wed, Mar 9, 2016 at 7:54 AM, Bond, Darryl <dbond(a)nrggos.com.au> wrote:
I have a 3 node 3.6.3 hosted engine cluster (Default) with a number
of VMs. The hosted engine is stored on gluster.
Adding an additional server to the Default cluster that isn't a hosted-engine ha
server fails.
Looking at the vdsm.log, the host attempts to mount the gluster as NFS with the gluster
options which fails.
Please add the logs above the log you posted, with the string "Run and
protect, connectStorageServer"
This log contains the arguments received from engine, revealing what
is going on.
jsonrpc.Executor/6::DEBUG::2016-03-09
15:10:01,022::fileUtils::143::Storage.fileUtils::(createdir) Creating directory:
/rhev/data-center/mnt/glusterSD/ovirt36-h1:_hosted-engine mode: None
jsonrpc.Executor/6::DEBUG::2016-03-09
15:10:01,022::storageServer::357::Storage.StorageServer.MountConnection::(_get_backup_servers_option)
Using bricks: ['ovirt36-h1', 'ovirt36-h2', 'ovirt36-h3']
jsonrpc.Executor/6::DEBUG::2016-03-09
15:10:01,022::mount::229::Storage.Misc.excCmd::(_runcmd) /usr/bin/taskset --cpu-list 0-11
/usr/bin/sudo -n /usr/bin/systemd-run --scope --slice=vdsm-glusterfs /usr/bin/mount -o
backup-volfile-servers=ovirt36-h2:ovirt36-h3 ovirt36-h1:/hosted-engine
/rhev/data-center/mnt/glusterSD/ovirt36-h1:_hosted-engine (cwd None)
-t glusterfs is missing here
This line can be generated only by the GlusterfFSConnection, used when
connecting
to gluster storage domain, but this connection type ads the "glustefs" type.
jsonrpc.Executor/6::ERROR::2016-03-09
15:10:01,042::hsm::2473::Storage.HSM::(connectStorageServer) Could not connect to
storageServer
Traceback (most recent call last):
File "/usr/share/vdsm/storage/hsm.py", line 2470, in connectStorageServer
conObj.connect()
File "/usr/share/vdsm/storage/storageServer.py", line 236, in connect
six.reraise(t, v, tb)
File "/usr/share/vdsm/storage/storageServer.py", line 228, in connect
self._mount.mount(self.options, self._vfsType, cgroup=self.CGROUP)
File "/usr/share/vdsm/storage/mount.py", line 225, in mount
return self._runcmd(cmd, timeout)
File "/usr/share/vdsm/storage/mount.py", line 241, in _runcmd
raise MountError(rc, ";".join((out, err)))
MountError: (32, ';Running scope as unit run-18808.scope.\nmount.nfs: an incorrect
mount option was specified\n')
I noticed the hosted-engine servers perform the same mount but pass the -t glusterfs
correctly.
A bug or am I doing something wrong??
I do not want to create a new datacentre without the hosted engine storage as I want to
use the same storage domains.
Same storage domains? Maybe you mean same bricks?
You cannot use the same storage domain from different DC. You can
create new gluster volume
using the same bricks.
Nir