[Users] Issues using local storage for gluster shared volume
Tony Feldmann
trfeldmann at gmail.com
Fri Mar 29 13:49:32 UTC 2013
Aren't there concerns with xfs and large files in cases of failures? I was
under the impression that if xfs was writing to a file and the system died
it would zero out the entire file. Just hesitant to put large vm files on
a fs like that. Is this still an issue with xfs?
On Fri, Mar 29, 2013 at 1:08 AM, Vijay Bellur <vbellur at redhat.com> wrote:
> On 03/28/2013 08:19 PM, Tony Feldmann wrote:
>
>> I have been trying for a month or so to get a 2 node cluster up and
>> running. I have engine installed on the first node, then add each each
>> system as a host to a posix dc. Both boxes have 4 data disks. After
>> adding the hosts I create a distributed replicate volume using 3 disk
>> from each host with ext4 filesystems. I click the 'optimize for virt'
>> option on the volume. There is a message in events that says that it
>> can't set a volume option, then it sets 2 volume options. Checking the
>> options tab I see that it added the gid/uid options. I was unable to
>> find in the logs what option was not set, I just see a message about
>> usage for volume set <volname> <option>. The volume starts fine and I
>> am able to create a data domain on the volume. Once the domain is
>> created I try to create a vm and it fails creating the disk. Error
>> messages are along the lines of task file exists and can't remove task
>> files. There are directories under tasks and when trying to manually
>> remove them I get the "directory not empty" error. Can someone please
>> shed some light on what I am doing wrong to get this 2 node cluster with
>> local disk as shared storage up and running?
>>
>>
> There are known problems with ext4 and gluster at the moment. Can you
> please confirm if you see similar behaviour with xfs and gluster?
>
> Thanks,
> Vijay
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20130329/a39768c6/attachment-0001.html>
More information about the Users
mailing list