I have a gluster volume that was created using an older version 4.x of oVirt. This was so long ago I am not sure I remember how I did it except that I used the wizard and the built in ansible job. I recently rebuilt my system that is hyperconverged. The
storage raid array (hardware backed) is in my hypervisor. The engine is a stand alone external device. I am now at ver 4.5.7 of Ovirt. I want to add my gluster brick back into the setup. I have reviewed the wiki notes but they presume a new build, not
my scenario. From what I read and researched, I think I need to create a distributed brick. How do you do that with just a single system when there is no replication to other hosts? Can anyone share a general procedure?
##DIAGNOSTICS
[root@vmh /]# mount
...
/dev/mapper/gluster_vg_sda-gluster_lv_vmstore on /gluster_bricks/vmstore type xfs (rw,relatime,seclabel,attr2,inode64,logbufs=8,logbsize=256k,sunit=512,swidth=2048,noquota)
[root@vmh /]# tail -n 1 /etc/fstab
/dev/mapper/gluster_vg_sda-gluster_lv_vmstore /gluster_bricks/vmstore xfs defaults 0 0
[root@vmh /]# gluster volume create vmstore vmh.tourville.org:/gluster_bricks/vmstore
volume create: vmstore: failed: The brick vmh.tourville.org:/gluster_bricks/vmstore is a mount point. Please create a sub-directory under the mount point and use that as the brick directory. Or use 'force' at the end of the command if you want to override this
behavior.
[root@vmh /]# gluster volume create vmstore vmh.tourville.org:/gluster_bricks/vmstore/vmstore
volume create: vmstore: failed: /gluster_bricks/vmstore/vmstore is already part of a volume
[root@vmh /]# tree /gluster_bricks/
/gluster_bricks/
vmstore
vmstore
42c329c3-8eca-4748-af7a-f48aca494d72
dom_md
ids
inbox
leases
metadata
outbox
xleases
images
019f828b-8f5d-47cf-bccd-7044402e7dd8
...
Do I need to force creation at this point? I just want to proceed cautiously at this point as I am not super confident with gluster.
IE- [root@vmh /]# gluster volume create vmstore vmh.tourville.org:/gluster_bricks/vmstore force