On 03/29/2013 07:19 PM, Tony Feldmann wrote:
Aren't there concerns with xfs and large files in cases of
failures? I
was under the impression that if xfs was writing to a file and the
system died it would zero out the entire file. Just hesitant to put
large vm files on a fs like that. Is this still an issue with xfs?
There are no known problems with recent kernels. There are quite a few
enterprise storage solutions that run on xfs.
Thanks,
Vijay
On Fri, Mar 29, 2013 at 1:08 AM, Vijay Bellur <vbellur(a)redhat.com
<mailto:vbellur@redhat.com>> wrote:
On 03/28/2013 08:19 PM, Tony Feldmann wrote:
I have been trying for a month or so to get a 2 node cluster up and
running. I have engine installed on the first node, then add
each each
system as a host to a posix dc. Both boxes have 4 data disks.
After
adding the hosts I create a distributed replicate volume using 3
disk
from each host with ext4 filesystems. I click the 'optimize for
virt'
option on the volume. There is a message in events that says
that it
can't set a volume option, then it sets 2 volume options.
Checking the
options tab I see that it added the gid/uid options. I was
unable to
find in the logs what option was not set, I just see a message about
usage for volume set <volname> <option>. The volume starts fine
and I
am able to create a data domain on the volume. Once the domain is
created I try to create a vm and it fails creating the disk. Error
messages are along the lines of task file exists and can't
remove task
files. There are directories under tasks and when trying to
manually
remove them I get the "directory not empty" error. Can someone
please
shed some light on what I am doing wrong to get this 2 node
cluster with
local disk as shared storage up and running?
There are known problems with ext4 and gluster at the moment. Can
you please confirm if you see similar behaviour with xfs and gluster?
Thanks,
Vijay