From previous thread about gluster issues, things seem to be running much better than
before, and it has raised a few questions that I can't seem to find any answers to:
Setup:
3 hyperconverged nodes
each node has 1x 1tb SSD and 1x 500gb nVME drive
each node is connected via ethernet and also by a 40gb infiniband connection for the
gluster replication.
Questions:
1. I created a 3tb VDO drive on the SSD and a 1.3tb VDO Drive on the nVME drive with a 1Gb
cache on each server and I enabled the RDMA transport.
a. Did I lose anything by doing the whole process manually? Since I did it like
that, things seem to run MUCH better so far, rebooting nodes the gluster resyncs almost
instantly and storage seems faster also.
b. Is there a way to split out the gluster traffic from the normal ovirt traffic (VM
and cluster communication)? originally I had separate names for each node for the gluster
network and it WAS split, but if I ever reset a gluster node, the name got changed. What
I did now is just put in a hosts file on each node that sends all traffic over the
infiniband, but I don't feel like that's optimal. I was unable to add in the
separate gluster names in the configuration as they were not part of the existing
cluster.
c. Why does the hyper-connverged wizard not let you adjust the VDO cache size, the
transport type, or use VDO to "over-subscribe" the drive sizes like I did
manually?
2. I wanted to add more space in the near future, would it be better to create a RAID0
with the new drives, or just use them as another separate storage location?