[ovirt-users] gluster replica volumes with different data usage
Ravishankar N
ravishankar at redhat.com
Mon Sep 18 06:31:07 UTC 2017
Possibly. I don't think changing shard size on the fly is supported,
especially when there are files on the volume that are sharded with a
different size.
-Ravi
On 09/18/2017 11:40 AM, Alex K wrote:
> The heal status is showing that no pending files need healing (also
> shown at GUI).
> When checking the bricks on the file system I see that what is
> different between the server is the .shard folder of the volume. One
> server reports 835GB while the other 1.1 TB.
> I recall to have changed the shard size at some point from 4 MB to 64MB.
> Could this be the cause?
>
> Thanx,
> Alex
>
> On Mon, Sep 18, 2017 at 8:14 AM, Ravishankar N <ravishankar at redhat.com
> <mailto:ravishankar at redhat.com>> wrote:
>
>
> On 09/18/2017 10:08 AM, Alex K wrote:
>> Hi Ravishankar,
>>
>> I am not referring to the arbiter volume(which is showing 0%
>> usage). I am referring to the other 2 volumes which are replicas
>> and should have the exact same data. Checking the status of other
>> bricks in ovirt (bricks used from iso and export domain) I see
>> that they all report same usage of data on the data volumes,
>> except the "vms" volume used for storing vms.
>
> Ah, okay. Some of the things that can cause a variation in disk
> usage:
> - Pending self-heals in gluster (check if `gluster volume heal
> <volname> info` doesn't show any entries. Also if there is
> anything under `.glusterfs/landfill` folder of the bricks).
> - XFS speculative preallocation
> - Possibly some bug in self-healing of sparse files by gluster
> (although we fixed known bugs a long time back in this area).
>
> Regards
> Ravi
>
>>
>> Thanx,
>> Alex
>>
>> On Sep 18, 2017 07:00, "Ravishankar N" <ravishankar at redhat.com
>> <mailto:ravishankar at redhat.com>> wrote:
>>
>>
>>
>> On 09/17/2017 08:41 PM, Alex K wrote:
>>> Hi all,
>>>
>>> I have replica 3 with 1 arbiter.
>>> When checking the gluster volume bricks they are reported as
>>> using different space, as per attached. How come they use
>>> different space? One would expect to use exactly the same
>>> space since they are replica.
>>>
>> The 3rd brick (arbiter ) only holds meta data, so it would
>> not consume as much space as the other 2 data bricks. So what
>> you are seeing is expected behaviour.
>> Regards,
>> Ravi
>>> Thanx,
>>> Alex
>>>
>>>
>>> _______________________________________________
>>> Users mailing list
>>> Users at ovirt.org <mailto:Users at ovirt.org>
>>> http://lists.ovirt.org/mailman/listinfo/users
>>> <http://lists.ovirt.org/mailman/listinfo/users>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170918/939013ab/attachment.html>
More information about the Users
mailing list