Hi

I have a hosted engine cluster running 3.5.2 on f20.  I have 6 nodes running centos 6.6 and three storage nodes also running centos 6.6 with gluster 3.6.3,

My primary data store is a replica 3 gluster volume.  I noticed that the size of some image files differs wildly on one server's brick.  Disks are all thin provisioned.  The bricks are thin provisioned lvm volume with xfs file systems.   The only difference between the systems is that the problem node is newer, a Dell R530 with MD1400 where as the other two are Dell R510's each with MD1200s.  The storage arrays all have the same 4TB disks.

e.g. for a disk that the ovirt console repost as having virtual size 500G and actual size 103G I see:


[root@gluster0 479d2197-de09-4012-8183-43c6baa7e65b]# cd ../d0d58fb9-ecaa-446f-bc42-dd681a16aee2/
[root@gluster0 d0d58fb9-ecaa-446f-bc42-dd681a16aee2]# du -sh *
106G c1b70bf0-c750-4177-8485-7b981e1f21a3
1.0M c1b70bf0-c750-4177-8485-7b981e1f21a3.lease
4.0K c1b70bf0-c750-4177-8485-7b981e1f21a3.meta
[root@gluster1 d0d58fb9-ecaa-446f-bc42-dd681a16aee2]# pwd
/export/brick5/ovirt-data/54d9ee82-0974-4a72-98a5-328d2e4007f1/images/d0d58fb9-ecaa-446f-bc42-dd681a16aee2
[root@gluster1 d0d58fb9-ecaa-446f-bc42-dd681a16aee2]# du -sh *
103G c1b70bf0-c750-4177-8485-7b981e1f21a3
1.0M c1b70bf0-c750-4177-8485-7b981e1f21a3.lease
4.0K c1b70bf0-c750-4177-8485-7b981e1f21a3.meta
[root@gluster-2 d0d58fb9-ecaa-446f-bc42-dd681a16aee2]# pwd
/export/brick5/ovirt-data/54d9ee82-0974-4a72-98a5-328d2e4007f1/images/d0d58fb9-ecaa-446f-bc42-dd681a16aee2
[root@gluster-2 d0d58fb9-ecaa-446f-bc42-dd681a16aee2]# du -sh *
501G c1b70bf0-c750-4177-8485-7b981e1f21a3
1.0M c1b70bf0-c750-4177-8485-7b981e1f21a3.lease
4.0K c1b70bf0-c750-4177-8485-7b981e1f21a3.meta

I'd appreciate any suggestions about troubleshooting and resolving his. Here is the volume info:

Volume Name: data
Type: Replicate
Volume ID: 5c6ff46d-1159-4c7e-8b16-5ffeb15cbaf9
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: gluster-2:/export/brick5/ovirt-data
Brick2: gluster1:/export/brick5/ovirt-data
Brick3: gluster0:/export/brick5/ovirt-data
Options Reconfigured:
performance.least-prio-threads: 4
performance.low-prio-threads: 16
performance.normal-prio-threads: 24
performance.high-prio-threads: 24
performance.io-thread-count: 32
diagnostics.count-fop-hits: off
diagnostics.latency-measurement: off
auth.allow: *
nfs.rpc-auth-allow: *
network.remote-dio: on
performance.stat-prefetch: off
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
storage.owner-gid: 36
storage.owner-uid: 36
cluster.eager-lock: enable
cluster.min-free-disk: 5%
cluster.rebalance-stats: on
cluster.background-self-heal-count: 16
cluster.readdir-optimize: on
cluster.metadata-self-heal: on
cluster.data-self-heal: on
cluster.entry-self-heal: on
cluster.self-heal-daemon: on
cluster.heal-timeout: 500
cluster.self-heal-window-size: 8
cluster.data-self-heal-algorithm: diff
cluster.quorum-type: auto
cluster.self-heal-readdir-size: 64KB
network.ping-timeout: 20
performance.open-behind: disable
cluster.server-quorum-ratio: 51%