I got a little further but just wanted to confirm a scenario. I realized part of my issue was due to having the volume mounted. I unmounted it and the message I get changed.
cd /
umount /gluster_bricks/vmstore
now comment out /etc/fstab, systemctl daemon-reload and mount -a to confirm it is no longer mounted.
[root@vmh /]# gluster volume create vmstore vmh.tourville.org:/gluster_bricks/vmstore
volume create: vmstore: failed: The brick vmh.tourville.org:/gluster_bricks/vmstore is being created in the root partition. It is recommended that you don't use the system's root partition for storage backend. Or use 'force' at the end of the command if you
want to override this behavior.
## DISK LAYOUT
[root@vmh /]# vgs
VG #PV #LV #SN Attr VSize VFree
gluster_vg_sda 1 2 0 wz--n- <5.46t 0
rl_vmh 1 10 0 wz--n- <219.18g <43.84g
[root@vmh /]# lvs
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
gluster_lv_vmstore gluster_vg_sda Vwi-a-t--- 4.39t gluster_thinpool_gluster_vg_sda 48.48
gluster_thinpool_gluster_vg_sda gluster_vg_sda twi-aot--- <5.43t 39.26 0.51
home rl_vmh Vwi-aotz-- 70.00g pool00 0.09
pool00 rl_vmh twi-aotz-- <171.01g 4.36 11.17
root rl_vmh Vwi-aotz-- 20.00g pool00 22.29
swap rl_vmh -wi-ao---- 4.00g
tmp rl_vmh Vwi-aotz-- 6.00g pool00 1.20
var rl_vmh Vwi-aotz-- 30.00g pool00 7.56
var_crash rl_vmh Vwi-aotz-- 15.00g pool00 0.42
var_log rl_vmh Vwi-aotz-- 16.00g pool00 0.84
var_log_audit rl_vmh Vwi-aotz-- 4.00g pool00 1.70
var_tmp rl_vmh Vwi-aotz-- 10.00g pool00 3.35
[root@vmh /]# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS
sda 8:0 0 5.5T 0 disk
3600605b00a2faca2233aa64a0966bb09 253:14 0 5.5T 0 mpath
gluster_vg_sda-gluster_thinpool_gluster_vg_sda_tmeta 253:16 0 15.8G 0 lvm
gluster_vg_sda-gluster_thinpool_gluster_vg_sda-tpool 253:18 0 5.4T 0 lvm
gluster_vg_sda-gluster_thinpool_gluster_vg_sda 253:19 0 5.4T 1 lvm
gluster_vg_sda-gluster_lv_vmstore 253:20 0 4.4T 0 lvm
gluster_vg_sda-gluster_thinpool_gluster_vg_sda_tdata 253:17 0 5.4T 0 lvm
gluster_vg_sda-gluster_thinpool_gluster_vg_sda-tpool 253:18 0 5.4T 0 lvm
gluster_vg_sda-gluster_thinpool_gluster_vg_sda 253:19 0 5.4T 1 lvm
gluster_vg_sda-gluster_lv_vmstore 253:20 0 4.4T 0 lvm
sdb 8:16 0 223.6G 0 disk
sdb1 8:17 0 219.3G 0 part
md127 9:127 0 219.2G 0 raid1
rl_vmh-pool00_tmeta 253:0 0 172M 0 lvm
rl_vmh-pool00-tpool 253:2 0 171G 0 lvm
rl_vmh-root 253:3 0 20G 0 lvm /
rl_vmh-pool00 253:5 0 171G 1 lvm
rl_vmh-home 253:6 0 70G 0 lvm /home
rl_vmh-var_crash 253:7 0 15G 0 lvm /var/crash
rl_vmh-tmp 253:8 0 6G 0 lvm /tmp
rl_vmh-var_tmp 253:9 0 10G 0 lvm /var/tmp
rl_vmh-var_log_audit 253:10 0 4G 0 lvm /var/log/audit
rl_vmh-var_log 253:11 0 16G 0 lvm /var/log
rl_vmh-var 253:12 0 30G 0 lvm /var
rl_vmh-pool00_tdata 253:1 0 171G 0 lvm
rl_vmh-pool00-tpool 253:2 0 171G 0 lvm
rl_vmh-root 253:3 0 20G 0 lvm /
rl_vmh-pool00 253:5 0 171G 1 lvm
rl_vmh-home 253:6 0 70G 0 lvm /home
rl_vmh-var_crash 253:7 0 15G 0 lvm /var/crash
rl_vmh-tmp 253:8 0 6G 0 lvm /tmp
rl_vmh-var_tmp 253:9 0 10G 0 lvm /var/tmp
rl_vmh-var_log_audit 253:10 0 4G 0 lvm /var/log/audit
rl_vmh-var_log 253:11 0 16G 0 lvm /var/log
rl_vmh-var 253:12 0 30G 0 lvm /var
rl_vmh-swap 253:4 0 4G 0 lvm [SWAP]
sdc 8:32 0 223.6G 0 disk
sdc1 8:33 0 600M 0 part /boot/efi
sdc2 8:34 0 1M 0 part
sdc3 8:35 0 1G 0 part /boot
sdc4 8:36 0 219.3G 0 part
md127 9:127 0 219.2G 0 raid1
rl_vmh-pool00_tmeta 253:0 0 172M 0 lvm
rl_vmh-pool00-tpool 253:2 0 171G 0 lvm
rl_vmh-root 253:3 0 20G 0 lvm /
rl_vmh-pool00 253:5 0 171G 1 lvm
rl_vmh-home 253:6 0 70G 0 lvm /home
rl_vmh-var_crash 253:7 0 15G 0 lvm /var/crash
rl_vmh-tmp 253:8 0 6G 0 lvm /tmp
rl_vmh-var_tmp 253:9 0 10G 0 lvm /var/tmp
rl_vmh-var_log_audit 253:10 0 4G 0 lvm /var/log/audit
rl_vmh-var_log 253:11 0 16G 0 lvm /var/log
rl_vmh-var 253:12 0 30G 0 lvm /var
rl_vmh-pool00_tdata 253:1 0 171G 0 lvm
rl_vmh-pool00-tpool 253:2 0 171G 0 lvm
rl_vmh-root 253:3 0 20G 0 lvm /
rl_vmh-pool00 253:5 0 171G 1 lvm
rl_vmh-home 253:6 0 70G 0 lvm /home
rl_vmh-var_crash 253:7 0 15G 0 lvm /var/crash
rl_vmh-tmp 253:8 0 6G 0 lvm /tmp
rl_vmh-var_tmp 253:9 0 10G 0 lvm /var/tmp
rl_vmh-var_log_audit 253:10 0 4G 0 lvm /var/log/audit
rl_vmh-var_log 253:11 0 16G 0 lvm /var/log
rl_vmh-var 253:12 0 30G 0 lvm /var
rl_vmh-swap 253:4 0 4G 0 lvm [SWAP]
sdd 8:48 0 596.2G 0 disk
Hitachi_HTS547564A9E384_J21B0053G8E15R 253:13 0 596.2G 0 mpath
Hitachi_HTS547564A9E384_J21B0053G8E15R1 253:15 0 596.2G 0 part
So I think I can do a force because they are on different VGs. Can anyone confirm? Is there an unseen scenario (IE- bad practices) where forcing could cause issues?