2016-09-23 13:17 GMT+02:00 Sahina Bose <sabose(a)redhat.com>:
What are the stability issues you're facing? Data volume if used as a data
storage domain should be a replica 3 volume as well.
Basically that after the first host installation+deploy (from the CLI),
after I enable the gluster management in the cluster, I have to manually
restart vdsmd on host1 to be able to install the other hosts. But maybe I
should just wait more time for vdmsd catch up with everything, I don't know.
Then I have some other problem like a ghost VM stuck on one host after
moving the host to maintenance and the VM (the hosted-engine, the only one
running in the whole cluster) being correctly migrated to another host,
solved only by a manual reboot of the whole host (and consequent HE fencing
of the host). I must say that that particular host is giving ECC correction
errors in one DIMM, so maybe it could just be an HW related problem.
>> Deploy the hosted-engine on the first VM (with the engine volume) froom
>> the CLI, then log in Ovirt admin, enable gluster support, install *and
>> deploy* from the GUI host2 and host3 (where the engine bricks are) and then
>> install host4 without deploying. This should get you the 4 hosts online,
>> but the engine will run only on the first 3
>
> Right. You can add the 4th node
to the cluster, but not have any bricks on
> this volume in which case VMs will be run on this node but will access data
> from the other 3 nodes.
Well, actually I *do* have data bricks on the 4th host, it's just the
engine volume that's not present there (but that host is not HE eligible
anyway). Am I doing something wrong?