
So, it's up and running now. Next step would be to configure ovirt to use the new gluster volume. And for backups to be moved from linode to this area instead. Please note I only allocated 1 TB to glusterlv, and 100 GB to backuplv, This leave ~700 GB untouched for allocation where we need it. "It's better to grow then to show" ... I just thought of that one, I know its bad. Sorry :-D --- ## The below commands are run on each machine, until otherwise is stated # find wwid of sdb1 multipath -l # add wwid of sdb1 to multipath.conf to blacklist it cat >> /etc/multipath.conf <<EOF blacklist { wwid 36848f690e6d9480019ac01600496584f } EOF # reload multipath multipath -F multipath -v2 # Create a volume group so that we can administrate storage more efficiently vgcreate datavg /dev/sdb1 # One for gluster, one for backup lvcreate -L 1T -n glusterlv datavg lvcreate -L 100G -n backuplv datavg # Create filesystems on both mkfs.ext4 /dev/disk/by-id/dm-name-datavg-backuplv mkfs.ext4 /dev/disk/by-id/dm-name-datavg-glusterlv # Create directories to mount at mkdir /srv/gluster /srv/backup # Find UUID of new filesystems blkid # Fix so that they mount on boot. cat >> /etc/fstab <<EOF UUID="69b0c4e5-ded7-4fc9-aa6f-03f6cc4f60c2" /srv/gluster ext4 defaults 1 2 UUID="4bae10e7-0d8e-477c-aa08-15f885bc52bd" /srv/backup ext4 defaults 1 2 EOF # Mount it! mount -a # Start gluster service glusterd start # Enable on boot chkconfig glusterd on # Add other node gluster peer probe rackspace02.ovirt org # Verify gluster peer status ## Only execute on one node gluster volume create vmstorage replica 2 transport tcp rackspace01.ovirt.org:/srv/gluster rackspace02.ovirt.org:/srv/gluster gluster volume start vmstorage -- /Alexander Rydekull