On Fri, Aug 19, 2016 at 6:20 PM, Nicolas Ecarnot <nicolas@ecarnot.net> wrote:
Le 19/08/2016 à 13:43, Sahina Bose a écrit :

Or are you adding the 3 nodes to your existing cluster? If so, I suggest you try adding this to a new cluster
OK, I tried and succeed to create a new cluster.
In this new cluster, I was ABLE to add the first new host, using its mgmt DNS name.
This first host still has to have its NICs configured, and (using Chrome or FF) the access to the network settings window is stalling the browser (I tried to restart even the engine, to no avail). Thus, I can not setup this first node NICs.

Thus, I can not add any further host because oVirt relies on a first host to validate the further ones.


Network team should be able to help you here.


OK, there were no mean I could continue this way (browser crash), so I tried and succeed doing so :
- remove the newly created host and cluster
- create a new DATACENTER
- create a new cluster in this DC
- add the first new host : OK
- add the 2 other new hosts : OK

Now, I can smoothly configure their NICs.

Doing all this, I saw that oVirt detected there already was existing gluster cluster and volume, and integrated it in oVirt.

Then, I was able to create a new storage domain in this new DC and cluster, using one of the *gluster* FQDN's host. It went nicely.

BUT, when viewing the volume tab and brick details, the displayed brick names are the host DNS name, and NOT the host GLUSTER DNS names.

I'm worrying about this, confirmed by what I read in the logs :

2016-08-19 14:46:30,484 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturnForXmlRpc] (DefaultQuartzScheduler_Worker-100) [107dc2e3] Could not associate brick 'serv-vm-al04-data.sdis.isere.fr:/gluster/data/brick04
' of volume '35026521-e76e-4774-8ddf-0a701b9eb40c' with correct network as no gluster network found in cluster '1c8e75a0-af3f-4e97-a8fb-2f7ef3ed9f30'
2016-08-19 14:46:30,492 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturnForXmlRpc] (DefaultQuartzScheduler_Worker-100) [107dc2e3] Could not associate brick 'serv-vm-al05-data.sdis.isere.fr:/gluster/data/brick04
' of volume '35026521-e76e-4774-8ddf-0a701b9eb40c' with correct network as no gluster network found in cluster '1c8e75a0-af3f-4e97-a8fb-2f7ef3ed9f30'
2016-08-19 14:46:30,500 WARN  [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturnForXmlRpc] (DefaultQuartzScheduler_Worker-100) [107dc2e3] Could not associate brick 'serv-vm-al06-data.sdis.isere.fr:/gluster/data/brick04
' of volume '35026521-e76e-4774-8ddf-0a701b9eb40c' with correct network as no gluster network found in cluster '1c8e75a0-af3f-4e97-a8fb-2f7ef3ed9f30'

[oVirt shell (connected)]# list clusters

id         : 00000001-0001-0001-0001-000000000045
name       : cluster51
description: Cluster d'alerte de test

id         : 1c8e75a0-af3f-4e97-a8fb-2f7ef3ed9f30
name       : cluster52
description: Cluster d'alerte de test

[oVirt shell (connected)]#

"cluster52" is the recent cluster, and I do have a dedicated gluster network, marked as gluster network, in the correct DC and cluster.
The only point is that :
- Each host has its name ("serv-vm-al04") and a second name for gluster ("serv-vm-al04-data").
- Using blahblahblah-data is correct on a gluster point of view
- Maybe oVirt is disturb not to be able to ping the gluster FQDN (not routed) and then throwing this error?


We do have a limitation currently that if you use multiple FQDNs, oVirt cannot associate it to the gluster brick correctly. This will be a problem only when you try brick management from oVirt - i.e try to remove or replace brick from oVirt. For monitoring brick status and detecting bricks - this is not an issue, and you can ignore the error in logs.

Adding Ramesh who has a patch to fix this .

--
Nicolas ECARNOT