--Apple-Mail=_33A5E632-5864-4B33-9FED-1901AA57DEBD
Content-Transfer-Encoding: quoted-printable
Content-Type: text/plain;
charset=utf-8
Why are you using an arbitrator if all your HW configs are identical? =
I=E2=80=99d use a true replica 3 in this case.
Also in my experience with gluster and vm hosting, the ZIL/slog degrades =
write performance unless it=E2=80=99s a truly dedicated disk. But I have =
8 spinners backing my ZFS volumes, so trying to share a sata disk =
wasn=E2=80=99t a good zil. If yours is dedicated SAS, keep it, if it=E2=80=
=99s SATA, try testing without it.
You don=E2=80=99t have compression enabled on your zfs volume, and I=E2=80=
=99d recommend enabling relatime on it. Depending on the amount of RAM =
in these boxes, you probably want to limit your zfs arc size to 8G or so =
(1/4 total ram or less). Gluster just works volumes hard during a =
rebuild, what=E2=80=99s the problem you=E2=80=99re seeing? If it=E2=80=99s=
affecting your VMs, using shading and tuning client & server threads =
can help avoid interruptions to your VMs while repairs are running. If =
you really need to limit it, you can use cgroups to keep it from hogging =
all the CPU, but it takes longer to heal, of course. There are a couple =
older posts and blogs about it, if you go back a while.
On Mar 3, 2017, at 9:02 AM, Arman Khalatyan <arm2arm(a)gmail.com>
wrote:
=20
The problem itself is not the streaming data performance., and also dd =
zero does
not help much in the production zfs running with compression.
the main problem comes when the gluster is starting to do something =
with that, it is using xattrs, probably accessing extended attributes =
inside the zfs is slower than XFS.
Also primitive find file or ls -l in the (dot)gluster folders takes =
ages:=20
=20
now I can see that arbiter host has almost 100% cache miss during the =
rebuild,
which is actually natural while he is reading always the new =
datasets:
[root@clei26 ~]# arcstat.py 1
time read miss miss% dmis dm% pmis pm% mmis mm% arcsz =
c =20
15:57:31 29 29 100 29 100 0 0 29 100 685M
=
31G =20
15:57:32 530 476 89 476 89 0 0 457 89 685M
=
31G =20
15:57:33 480 467 97 467 97 0 0 463 97 685M
=
31G =20
15:57:34 452 443 98 443 98 0 0 435 97 685M
=
31G =20
15:57:35 582 547 93 547 93 0 0 536 94 685M
=
31G =20
15:57:36 439 417 94 417 94 0 0 393 94 685M
=
31G =20
15:57:38 435 392 90 392 90 0 0 374 89 685M
=
31G =20
15:57:39 364 352 96 352 96 0 0 352 96 685M
=
31G =20
15:57:40 408 375 91 375 91 0 0 360 91 685M
=
31G =20
15:57:41 552 539 97 539 97 0 0 539 97 685M
=
31G =20
=20
It looks like we cannot have in the same system performance and =
reliability :(
Simply final conclusion is with the single disk+ssd even zfs doesnot
=
help to speedup the glusterfs healing.
I will stop here:)
=20
=20
=20
=20
On Fri, Mar 3, 2017 at 3:35 PM, Juan Pablo <pablo.localhost(a)gmail.com =
<mailto:pablo.localhost@gmail.com>> wrote:
cd to inside the pool path
then dd if=3D/dev/zero of=3Dtest.tt <
http://test.tt/> bs=3D1M=20
leave it runing 5/10 minutes.
do ctrl+c paste result here.
etc.
=20
2017-03-03 11:30 GMT-03:00 Arman Khalatyan <arm2arm(a)gmail.com =
<mailto:arm2arm@gmail.com>>:
No, I have one pool made of the one disk and ssd as a cache and log =
device.
I have 3 Glusterfs bricks- separate 3 hosts:Volume type Replicate =
(Arbiter)=3D replica 2+1!
That how much you can push into compute nodes(they have only 3 disk =
slots).
=20
=20
On Fri, Mar 3, 2017 at 3:19 PM, Juan Pablo <pablo.localhost(a)gmail.com =
<mailto:pablo.localhost@gmail.com>> wrote:
ok, you have 3 pools, zclei22, logs and cache, thats wrong. you
should =
have 1 pool, with zlog+cache if you are looking for performance.
also, dont mix drives.=20
whats the performance issue you are facing?=20
=20
=20
regards,
=20
2017-03-03 11:00 GMT-03:00 Arman Khalatyan <arm2arm(a)gmail.com =
<mailto:arm2arm@gmail.com>>:
This is CentOS 7.3 ZoL version 0.6.5.9-1
=20
[root@clei22 ~]# lsscsi
=20
[2:0:0:0] disk ATA INTEL SSDSC2CW24 400i /dev/sda
=20
[3:0:0:0] disk ATA HGST HUS724040AL AA70 /dev/sdb
=20
[4:0:0:0] disk ATA WDC WD2002FYPS-0 1G01 /dev/sdc
=20
=20
=20
[root@clei22 ~]# pvs ;vgs;lvs
=20
PV VG Fmt =
Attr
PSize PFree
=20
/dev/mapper/INTEL_SSDSC2CW240A3_CVCV306302RP240CGN vg_cache =
lvm2 a--
223.57g 0
=20
/dev/sdc2 centos_clei22 =
lvm2 a--
1.82t 64.00m
=20
VG #PV #LV #SN Attr VSize VFree
=20
centos_clei22 1 3 0 wz--n- 1.82t 64.00m
=20
vg_cache 1 2 0 wz--n- 223.57g 0
=20
LV VG Attr LSize Pool Origin Data% Meta% =
Move Log
Cpy%Sync Convert
=20
home centos_clei22 -wi-ao---- 1.74t =
=20
=20
root centos_clei22 -wi-ao---- 50.00g =
=20
=20
swap centos_clei22 -wi-ao---- 31.44g =
=20
=20
lv_cache vg_cache -wi-ao---- 213.57g =
=20
=20
lv_slog vg_cache -wi-ao---- 10.00g =20
=20
=20
=20
[root@clei22 ~]# zpool status -v
=20
pool: zclei22
=20
state: ONLINE
=20
scan: scrub repaired 0 in 0h0m with 0 errors on Tue Feb 28 14:16:07 =
2017
=20
config:
=20
=20
=20
NAME STATE READ WRITE CKSUM
=20
zclei22 ONLINE 0 0 0
=20
HGST_HUS724040ALA640_PN2334PBJ4SV6T1 ONLINE 0 0 0
=20
logs
=20
lv_slog ONLINE 0 0 0
=20
cache
=20
lv_cache ONLINE 0 0 0
=20
=20
=20
errors: No known data errors
=20
=20
ZFS config:
=20
[root@clei22 ~]# zfs get all zclei22/01
=20
NAME PROPERTY VALUE SOURCE
=20
zclei22/01 type filesystem -
=20
zclei22/01 creation Tue Feb 28 14:06 2017 -
=20
zclei22/01 used 389G -
=20
zclei22/01 available 3.13T -
=20
zclei22/01 referenced 389G -
=20
zclei22/01 compressratio 1.01x -
=20
zclei22/01 mounted yes -
=20
zclei22/01 quota none default
=20
zclei22/01 reservation none default
=20
zclei22/01 recordsize 128K local
=20
zclei22/01 mountpoint /zclei22/01 default
=20
zclei22/01 sharenfs off default
=20
zclei22/01 checksum on default
=20
zclei22/01 compression off local
=20
zclei22/01 atime on default
=20
zclei22/01 devices on default
=20
zclei22/01 exec on default
=20
zclei22/01 setuid on default
=20
zclei22/01 readonly off default
=20
zclei22/01 zoned off default
=20
zclei22/01 snapdir hidden default
=20
zclei22/01 aclinherit restricted default
=20
zclei22/01 canmount on default
=20
zclei22/01 xattr sa local
=20
zclei22/01 copies 1 default
=20
zclei22/01 version 5 -
=20
zclei22/01 utf8only off -
=20
zclei22/01 normalization none -
=20
zclei22/01 casesensitivity sensitive -
=20
zclei22/01 vscan off default
=20
zclei22/01 nbmand off default
=20
zclei22/01 sharesmb off default
=20
zclei22/01 refquota none default
=20
zclei22/01 refreservation none default
=20
zclei22/01 primarycache metadata local
=20
zclei22/01 secondarycache metadata local
=20
zclei22/01 usedbysnapshots 0 -
=20
zclei22/01 usedbydataset 389G -
=20
zclei22/01 usedbychildren 0 -
=20
zclei22/01 usedbyrefreservation 0 -
=20
zclei22/01 logbias latency default
=20
zclei22/01 dedup off default
=20
zclei22/01 mlslabel none default
=20
zclei22/01 sync disabled local
=20
zclei22/01 refcompressratio 1.01x -
=20
zclei22/01 written 389G -
=20
zclei22/01 logicalused 396G -
=20
zclei22/01 logicalreferenced 396G -
=20
zclei22/01 filesystem_limit none default
=20
zclei22/01 snapshot_limit none default
=20
zclei22/01 filesystem_count none default
=20
zclei22/01 snapshot_count none default
=20
zclei22/01 snapdev hidden default
=20
zclei22/01 acltype off default
=20
zclei22/01 context none default
=20
zclei22/01 fscontext none default
=20
zclei22/01 defcontext none default
=20
zclei22/01 rootcontext none default
=20
zclei22/01 relatime off default
=20
zclei22/01 redundant_metadata all default
=20
zclei22/01 overlay off default
=20
=20
=20
=20
=20
=20
On Fri, Mar 3, 2017 at 2:52 PM, Juan Pablo <pablo.localhost(a)gmail.com =
<mailto:pablo.localhost@gmail.com>> wrote:
Which operating system version are you using for your zfs
storage?=20
do:
zfs get all your-pool-name
use arc_summary.py from freenas git repo if you wish.
=20
=20
2017-03-03 10:33 GMT-03:00 Arman Khalatyan <arm2arm(a)gmail.com =
<mailto:arm2arm@gmail.com>>:
Pool load:
[root@clei21 ~]# zpool iostat -v 1=20
capacity operations =
bandwidth
pool alloc free read write
=
read write
-------------------------------------- ----- ----- ----- ----- =
----- -----
zclei21 10.1G 3.62T 0 112
=
823 8.82M
HGST_HUS724040ALA640_PN2334PBJ52XWT1 10.1G 3.62T 0 46
=
626 4.40M
logs - - - -
=
- -
lv_slog 225M 9.72G 0 66
=
198 4.45M
cache - - - -
=
- -
lv_cache 9.81G 204G 0 46
=
56 4.13M
-------------------------------------- ----- ----- ----- ----- =
----- -----
=20
capacity operations =
bandwidth
pool alloc free read write
=
read write
-------------------------------------- ----- ----- ----- ----- =
----- -----
zclei21 10.1G 3.62T 0 191
=
0 12.8M
HGST_HUS724040ALA640_PN2334PBJ52XWT1 10.1G 3.62T 0 0
=
0 0
logs - - - -
=
- -
lv_slog 225M 9.72G 0 191
=
0 12.8M
cache - - - -
=
- -
lv_cache 9.83G 204G 0 218
=
0 20.0M
-------------------------------------- ----- ----- ----- ----- =
----- -----
=20
capacity operations =
bandwidth
pool alloc free read write
=
read write
-------------------------------------- ----- ----- ----- ----- =
----- -----
zclei21 10.1G 3.62T 0 191
=
0 12.7M
HGST_HUS724040ALA640_PN2334PBJ52XWT1 10.1G 3.62T 0 0
=
0 0
logs - - - -
=
- -
lv_slog 225M 9.72G 0 191
=
0 12.7M
cache - - - -
=
- -
lv_cache 9.83G 204G 0 72
=
0 7.68M
-------------------------------------- ----- ----- ----- ----- =
----- -----
=20
=20
On Fri, Mar 3, 2017 at 2:32 PM, Arman Khalatyan <arm2arm(a)gmail.com =
<mailto:arm2arm@gmail.com>> wrote:
Glusterfs now in healing mode:
Receiver:
[root@clei21 ~]# arcstat.py 1
time read miss miss% dmis dm% pmis pm% mmis mm% arcsz =
c =20
13:24:49 0 0 0 0 0 0 0 0 0 4.6G
=
31G =20
13:24:50 154 80 51 80 51 0 0 80 51 4.6G
=
31G =20
13:24:51 179 62 34 62 34 0 0 62 42 4.6G
=
31G =20
13:24:52 148 68 45 68 45 0 0 68 45 4.6G
=
31G =20
13:24:53 140 64 45 64 45 0 0 64 45 4.6G
=
31G =20
13:24:54 124 48 38 48 38 0 0 48 38 4.6G
=
31G =20
13:24:55 157 80 50 80 50 0 0 80 50 4.7G
=
31G =20
13:24:56 202 68 33 68 33 0 0 68 41 4.7G
=
31G =20
13:24:57 127 54 42 54 42 0 0 54 42 4.7G
=
31G =20
13:24:58 126 50 39 50 39 0 0 50 39 4.7G
=
31G =20
13:24:59 116 40 34 40 34 0 0 40 34 4.7G
=
31G =20
=20
=20
Sender
[root@clei22 ~]# arcstat.py 1
time read miss miss% dmis dm% pmis pm% mmis mm% arcsz =
c =20
13:28:37 8 2 25 2 25 0 0 2 25 468M
=
31G =20
13:28:38 1.2K 727 62 727 62 0 0 525 54 469M
=
31G =20
13:28:39 815 508 62 508 62 0 0 376 55 469M
=
31G =20
13:28:40 994 624 62 624 62 0 0 450 54 469M
=
31G =20
13:28:41 783 456 58 456 58 0 0 338 50 470M
=
31G =20
13:28:42 916 541 59 541 59 0 0 390 50 470M
=
31G =20
13:28:43 768 437 56 437 57 0 0 313 48 471M
=
31G =20
13:28:44 877 534 60 534 60 0 0 393 53 470M
=
31G =20
13:28:45 957 630 65 630 65 0 0 450 57 470M
=
31G =20
13:28:46 819 479 58 479 58 0 0 357 51 471M
=
31G =20
=20
=20
On Thu, Mar 2, 2017 at 7:18 PM, Juan Pablo <pablo.localhost(a)gmail.com =
<mailto:pablo.localhost@gmail.com>> wrote:
hey,
what are you using for zfs? get an arc status and show please
=20
=20
2017-03-02 9:57 GMT-03:00 Arman Khalatyan <arm2arm(a)gmail.com =
<mailto:arm2arm@gmail.com>>:
no,=20
ZFS itself is not on top of lvm. only ssd was spitted by lvm for =
slog(10G) and
cache (the rest)
but in any-case the ssd does not help much on glusterfs/ovirt load
it =
has almost 100% cache misses....:( (terrible performance compare with =
nfs)
=20
=20
=20
=20
=20
On Thu, Mar 2, 2017 at 1:47 PM, FERNANDO FREDIANI =
<fernando.frediani(a)upx.com
<mailto:fernando.frediani@upx.com>> wrote:
Am I understanding correctly, but you have Gluster on the top of ZFS
=
which is on the top of LVM ? If so, why the usage of LVM was necessary ? =
I have ZFS with any need of LVM.
=20
Fernando
=20
On 02/03/2017 06:19, Arman Khalatyan wrote:
> Hi,=20
> I use 3 nodes with zfs and glusterfs.
> Are there any suggestions to optimize it?
>=20
> host zfs config 4TB-HDD+250GB-SSD:
> [root@clei22 ~]# zpool status=20
> pool: zclei22
> state: ONLINE
> scan: scrub repaired 0 in 0h0m with 0 errors on Tue Feb 28 14:16:07 =
2017
> config:
>=20
> NAME STATE READ WRITE =
CKSUM
> zclei22 ONLINE 0 0
=
0
> HGST_HUS724040ALA640_PN2334PBJ4SV6T1 ONLINE 0 0
=
0
> logs
> lv_slog ONLINE 0 0 =
0
> cache
> lv_cache ONLINE 0 0 =
0
>=20
> errors: No known data errors
>=20
> Name:
> GluReplica
> Volume ID:
> ee686dfe-203a-4caa-a691-26353460cc48
> Volume Type:
> Replicate (Arbiter)
> Replica Count:
> 2 + 1
> Number of Bricks:
> 3
> Transport Types:
> TCP, RDMA
> Maximum no of snapshots:
> 256
> Capacity:
> 3.51 TiB total, 190.56 GiB used, 3.33 TiB free
>=20
>=20
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org <mailto:Users@ovirt.org>
>
http://lists.ovirt.org/mailman/listinfo/users =
<
http://lists.ovirt.org/mailman/listinfo/users>
=20
=20
_______________________________________________
Users mailing list
Users(a)ovirt.org <mailto:Users@ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users =
<
http://lists.ovirt.org/mailman/listinfo/users>
>=20
=20
=20
_______________________________________________
Users mailing list
Users(a)ovirt.org <mailto:Users@ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users =
<
http://lists.ovirt.org/mailman/listinfo/users>
=20
=20
=20
=20
=20
=20
=20
=20
=20
=20
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
--Apple-Mail=_33A5E632-5864-4B33-9FED-1901AA57DEBD
Content-Transfer-Encoding: quoted-printable
Content-Type: text/html;
charset=utf-8
<html><head><meta http-equiv=3D"Content-Type"
content=3D"text/html =
charset=3Dutf-8"></head><body style=3D"word-wrap: break-word; =
-webkit-nbsp-mode: space; -webkit-line-break: after-white-space;" =
class=3D"">Why are you using an arbitrator if all your HW configs are =
identical? I=E2=80=99d use a true replica 3 in this case.<div =
class=3D""><br class=3D""></div><div
class=3D"">Also in my experience =
with gluster and vm hosting, the ZIL/slog degrades write performance =
unless it=E2=80=99s a truly dedicated disk. But I have 8 spinners =
backing my ZFS volumes, so trying to share a sata disk wasn=E2=80=99t a =
good zil. If yours is dedicated SAS, keep it, if it=E2=80=99s SATA, try =
testing without it.</div><div class=3D""><br
class=3D""></div><div =
class=3D"">You don=E2=80=99t have compression enabled on your zfs =
volume, and I=E2=80=99d recommend enabling relatime on it. Depending on =
the amount of RAM in these boxes, you probably want to limit your zfs =
arc size to 8G or so (1/4 total ram or less). Gluster just works volumes =
hard during a rebuild, what=E2=80=99s the problem you=E2=80=99re seeing? =
If it=E2=80=99s affecting your VMs, using shading and tuning client =
& server threads can help avoid interruptions to your VMs while =
repairs are running. If you really need to limit it, you can use cgroups =
to keep it from hogging all the CPU, but it takes longer to heal, of =
course. There are a couple older posts and blogs about it, if you go =
back a while.</div><div class=3D""><br
class=3D""></div><div =
class=3D""><br class=3D""><div><blockquote
type=3D"cite" class=3D""><div =
class=3D"">On Mar 3, 2017, at 9:02 AM, Arman Khalatyan <<a =
href=3D"mailto:arm2arm@gmail.com"
class=3D"">arm2arm(a)gmail.com</a>&gt; =
wrote:</div><br class=3D"Apple-interchange-newline"><div
class=3D""><div =
dir=3D"ltr" class=3D""><div class=3D""><div
class=3D""><div =
class=3D""><div class=3D""><div
class=3D""><div class=3D"">The problem =
itself is not the streaming data performance., and also dd zero does not =
help much in the production zfs running with compression.<br =
class=3D""></div>the main problem comes when the gluster is starting to
=
do something with that, it is using xattrs, probably accessing extended =
attributes inside the zfs is slower than XFS.<br
class=3D""></div>Also =
primitive find file or ls -l in the (dot)gluster folders takes ages: <br =
class=3D""><br class=3D""></div>now I can see that
arbiter host has =
almost 100% cache miss during the rebuild, which is actually natural =
while he is reading always the new datasets:<br class=3D"">[root@clei26 =
~]# arcstat.py 1<br class=3D"">
time read =
miss miss% dmis dm% pmis pm%
=
mmis mm% arcsz c
<br =
class=3D"">15:57:31
29 =
29 100 29 =
100 0
0 =
29 100 685M 31G <br =
class=3D"">15:57:32 530 =
476 89 476
=
89 0
0 =
457 89 685M 31G
<br =
class=3D"">15:57:33 480 =
467 97 467
=
97 0
0 =
463 97 685M 31G
<br =
class=3D"">15:57:34 452 =
443 98 443
=
98 0
0 =
435 97 685M 31G
<br =
class=3D"">15:57:35 582 =
547 93 547
=
93 0
0 =
536 94 685M 31G
<br =
class=3D"">15:57:36 439 =
417 94 417
=
94 0
0 =
393 94 685M 31G
<br =
class=3D"">15:57:38 435 =
392 90 392
=
90 0
0 =
374 89 685M 31G
<br =
class=3D"">15:57:39 364 =
352 96 352
=
96 0
0 =
352 96 685M 31G
<br =
class=3D"">15:57:40 408 =
375 91 375
=
91 0
0 =
360 91 685M 31G
<br =
class=3D"">15:57:41 552 =
539 97 539
=
97 0
0 =
539 97 685M 31G
<br =
class=3D""><br class=3D""></div>It looks like we cannot
have in the same =
system performance and reliability :(<br class=3D""></div>Simply
final =
conclusion is with the single disk+ssd even zfs doesnot help to speedup =
the glusterfs healing.<br class=3D""></div>I will stop here:)<br
=
class=3D""><br class=3D""><div
class=3D""><div class=3D""><div =
class=3D""><br class=3D""><div
class=3D""><br class=3D""><br =
class=3D""><div class=3D"gmail_extra"><div
class=3D"gmail_quote">On Fri, =
Mar 3, 2017 at 3:35 PM, Juan Pablo <span dir=3D"ltr"
class=3D""><<a =
href=3D"mailto:pablo.localhost@gmail.com" target=3D"_blank" =
class=3D"">pablo.localhost(a)gmail.com</a>&gt;</span>
wrote:<br =
class=3D""><blockquote class=3D"gmail_quote"
style=3D"margin:0px 0px 0px =
0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div =
dir=3D"ltr" class=3D""><div class=3D""><div
class=3D""><div =
class=3D""><div class=3D"">cd to inside the pool path<br
=
class=3D""></div>then dd if=3D/dev/zero of=3D<a
href=3D"http://test.tt/" =
target=3D"_blank" class=3D"">test.tt</a> bs=3D1M <br =
class=3D""></div>leave it runing 5/10 minutes.<br
class=3D""></div>do =
ctrl+c paste result here.<br class=3D""></div>etc.<br =
class=3D""></div><div class=3D"gmail-HOEnZb"><div
class=3D"gmail-h5"><div =
class=3D"gmail_extra"><br class=3D""><div
class=3D"gmail_quote">2017-03-03=
11:30 GMT-03:00 Arman Khalatyan <span dir=3D"ltr"
class=3D""><<a =
href=3D"mailto:arm2arm@gmail.com" target=3D"_blank" =
class=3D"">arm2arm@gmail.com</a>></span>:<br
class=3D""><blockquote =
class=3D"gmail_quote" style=3D"margin:0px 0px 0px 0.8ex;border-left:1px =
solid rgb(204,204,204);padding-left:1ex"><div dir=3D"ltr"
class=3D""><div =
class=3D""><div class=3D"">No, I have one pool made of the
one disk and =
ssd as a cache and log device.<br class=3D""></div>I have 3
Glusterfs =
bricks- separate 3 hosts:Volume type Replicate (Arbiter)=3D replica =
2+1!<br class=3D""></div>That how much you can push into compute =
nodes(they have only 3 disk slots).<br class=3D""><br =
class=3D""></div><div
class=3D"gmail-m_6281861324822600694HOEnZb"><div =
class=3D"gmail-m_6281861324822600694h5"><div
class=3D"gmail_extra"><br =
class=3D""><div class=3D"gmail_quote">On Fri, Mar 3, 2017 at
3:19 PM, =
Juan Pablo <span dir=3D"ltr" class=3D""><<a =
href=3D"mailto:pablo.localhost@gmail.com" target=3D"_blank" =
class=3D"">pablo.localhost(a)gmail.com</a>&gt;</span>
wrote:<br =
class=3D""><blockquote class=3D"gmail_quote"
style=3D"margin:0px 0px 0px =
0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div =
dir=3D"ltr" class=3D""><div class=3D""><div
class=3D"">ok, you have 3 =
pools, zclei22, logs and cache, thats wrong. you should have 1 pool, =
with zlog+cache if you are looking for performance.<br =
class=3D""></div>also, dont mix drives. <br
class=3D""></div>whats the =
performance issue you are facing? <br class=3D""><div
class=3D""><div =
class=3D""><br class=3D""><br
class=3D""></div><div =
class=3D"">regards,</div></div></div><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465HOEnZb"><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465h5"><div =
class=3D"gmail_extra"><br class=3D""><div
class=3D"gmail_quote">2017-03-03=
11:00 GMT-03:00 Arman Khalatyan <span dir=3D"ltr"
class=3D""><<a =
href=3D"mailto:arm2arm@gmail.com" target=3D"_blank" =
class=3D"">arm2arm@gmail.com</a>></span>:<br
class=3D""><blockquote =
class=3D"gmail_quote" style=3D"margin:0px 0px 0px 0.8ex;border-left:1px =
solid rgb(204,204,204);padding-left:1ex"><div dir=3D"ltr"
class=3D"">This =
is CentOS 7.3 ZoL version 0.6.5.9-1<br class=3D""><br =
class=3D""><blockquote class=3D""><p
class=3D"">[root@clei22 ~]# lsscsi =
</p><p class=3D"">[2:0:0:0]
disk =
ATA INTEL SSDSC2CW24 400i
/dev/sda =
</p><p class=3D"">[3:0:0:0]
disk =
ATA HGST HUS724040AL AA70
/dev/sdb =
</p><p class=3D"">[4:0:0:0]
disk =
ATA WDC WD2002FYPS-0 1G01
/dev/sdc =
</p><p class=3D""><br class=3D""></p><p
class=3D"">[root@clei22 ~]# pvs =
;vgs;lvs</p><p class=3D""> =
PV =
&n=
bsp; <wbr =
class=3D""> &nb=
sp;
=
VG
=
Fmt Attr PSize PFree </p><p
class=3D""> =
/dev/mapper/INTEL_SSDSC2CW240A<wbr class=3D"">3_CVCV306302RP240CGN =
vg_cache lvm2 a-- =
223.57g 0 </p><p
class=3D""> =
/dev/sdc2  =
; <wbr
=
class=3D""> &nb=
sp;
centos_clei22 =
lvm2 a-- 1.82t 64.00m</p><p
class=3D""> =
VG
#PV =
#LV #SN Attr VSize VFree </p><p
class=3D""> =
centos_clei22 1 3 0 =
wz--n- 1.82t 64.00m</p><p class=3D"">
=
vg_cache
1 =
2 0 wz--n- 223.57g 0
</p><p =
class=3D"">
LV =
VG
=
Attr LSize
Pool Origin =
Data% Meta% Move Log Cpy%Sync Convert</p><p
class=3D""> =
home centos_clei22 -wi-ao----
=
1.74t &nb=
sp;  =
; <wbr =
class=3D""> &nb=
sp;  =
; </p><p class=3D""> =
root centos_clei22 -wi-ao---- =
50.00g &n=
bsp; &nbs=
p;<wbr =
class=3D""> &nb=
sp;  =
; </p><p
class=3D""> =
swap centos_clei22 -wi-ao---- =
31.44g &n=
bsp; &nbs=
p;<wbr =
class=3D""> &nb=
sp;  =
; </p><p
class=3D""> lv_cache =
vg_cache -wi-ao---- =
213.57g &=
nbsp; <wb=
r =
class=3D""> &nb=
sp;  =
; </p><p
class=3D""> lv_slog =
vg_cache -wi-ao---- =
10.00g </p><p class=3D""><br
class=3D""></p><p =
class=3D"">[root@clei22 ~]# zpool status -v</p><span
class=3D""><p =
class=3D""> pool: zclei22</p><p
class=3D""> state: =
ONLINE</p><p class=3D""> scan: scrub repaired 0 in 0h0m
with 0 =
errors on Tue Feb 28 14:16:07 2017</p><p
class=3D"">config:</p><p =
class=3D""><br class=3D""></p><p
class=3D""> =
NAME &nbs=
p; =
<wbr =
class=3D"">
=
STATE READ WRITE CKSUM</p><p =
class=3D""> =
zclei22 &=
nbsp; <wb=
r
class=3D"">
=
ONLINE
0 =
0 0</p><p
class=3D""> =
HGST_HUS724040ALA640_PN2334PBJ<wbr class=3D"">4SV6T1 =
ONLINE
0 =
0 0</p><p
class=3D""> =
logs</p><p class=3D""> =
lv_slog &=
nbsp; <wb=
r
class=3D"">
=
ONLINE
0 =
0 0</p><p
class=3D""> =
cache</p><p class=3D""> =
lv_cache =
<wbr
=
class=3D"">
=
ONLINE
0 =
0 0</p><p class=3D""><br
class=3D""></p><p =
class=3D"">errors: No known data
errors</p></span></blockquote><p =
class=3D""><font size=3D"3" class=3D""><b
class=3D""><br class=3D"">ZFS =
config:</b></font></p><blockquote =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294gmail-tr_bq"><p class=3D"">[root@clei22
~]# zfs =
get all zclei22/01</p><p =
class=3D"">NAME
=
PROPERTY =
=
VALUE &nb=
sp; SOURCE</p><p
class=3D"">zclei22/01 =
=
type &nbs=
p; =
filesystem &nbs=
p; -</p><p class=3D"">zclei22/01 =
creation =
Tue Feb 28 14:06 2017 -</p><p =
class=3D"">zclei22/01 =
used &nbs=
p; =
389G &nbs=
p; -</p><p
class=3D"">zclei22/01 =
=
available  =
; =
3.13T &nb=
sp; -</p><p
class=3D"">zclei22/01 =
referenced &nbs=
p; =
389G &nbs=
p; -</p><p
class=3D"">zclei22/01 =
compressratio
=
1.01x &nb=
sp; -</p><p
class=3D"">zclei22/01 =
mounted &=
nbsp; =
yes  =
; -</p><p
=
class=3D"">zclei22/01 =
quota &nb=
sp; =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
reservation
=
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
recordsize &nbs=
p; =
128K &nbs=
p; local</p><p =
class=3D"">zclei22/01 =
mountpoint &nbs=
p; =
/zclei22/01 &nb=
sp; default</p><p class=3D"">zclei22/01 =
sharenfs =
=
off  =
;
default</p><p =
class=3D"">zclei22/01 =
checksum =
=
on =
default</p><p =
class=3D"">zclei22/01 =
compression
=
off  =
;
local</p><p =
class=3D"">zclei22/01 =
atime &nb=
sp; =
on =
default</p><p =
class=3D"">zclei22/01 =
devices &=
nbsp; =
on =
default</p><p =
class=3D"">zclei22/01 =
exec &nbs=
p; =
on =
default</p><p =
class=3D"">zclei22/01 =
setuid &n=
bsp; =
on =
default</p><p =
class=3D"">zclei22/01 =
readonly =
=
off  =
;
default</p><p =
class=3D"">zclei22/01 =
zoned &nb=
sp; =
off  =
;
default</p><p =
class=3D"">zclei22/01 =
snapdir &=
nbsp; =
hidden &n=
bsp; default</p><p
class=3D"">zclei22/01 =
aclinherit &nbs=
p; =
restricted &nbs=
p; default</p><p class=3D"">zclei22/01 =
canmount =
=
on =
default</p><p =
class=3D"">zclei22/01 =
xattr &nb=
sp; =
sa =
local</p><p =
class=3D"">zclei22/01 =
copies &n=
bsp; =
1 &=
nbsp;
default</p><p =
class=3D"">zclei22/01 =
version &=
nbsp; =
5 &=
nbsp;
-</p><p =
class=3D"">zclei22/01 =
utf8only =
=
off  =
; -</p><p
=
class=3D"">zclei22/01 =
normalization
=
none &nbs=
p; -</p><p
class=3D"">zclei22/01 =
casesensitivity =
sensitive  =
; -</p><p class=3D"">zclei22/01 =
vscan &nb=
sp; =
off  =
;
default</p><p =
class=3D"">zclei22/01 =
nbmand &n=
bsp; =
off  =
;
default</p><p =
class=3D"">zclei22/01 =
sharesmb =
=
off  =
;
default</p><p =
class=3D"">zclei22/01 =
refquota =
=
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
refreservation =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
primarycache
=
metadata =
local</p><p
class=3D"">zclei22/01 =
secondarycache =
metadata =
local</p><p
class=3D"">zclei22/01 =
usedbysnapshots =
0 &=
nbsp;
-</p><p =
class=3D"">zclei22/01 =
usedbydataset
=
389G &nbs=
p; -</p><p
class=3D"">zclei22/01 =
usedbychildren =
0 &=
nbsp;
-</p><p =
class=3D"">zclei22/01 usedbyrefreservation =
0 &=
nbsp;
-</p><p =
class=3D"">zclei22/01 =
logbias &=
nbsp; =
latency &=
nbsp; default</p><p
class=3D"">zclei22/01 =
dedup &nb=
sp; =
off  =
;
default</p><p =
class=3D"">zclei22/01 =
mlslabel =
=
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
sync &nbs=
p; =
disabled =
local</p><p
class=3D"">zclei22/01 =
refcompressratio =
1.01x &nb=
sp; -</p><p
class=3D"">zclei22/01 =
written &=
nbsp; =
389G &nbs=
p; -</p><p
class=3D"">zclei22/01 =
logicalused
=
396G &nbs=
p; -</p><p
class=3D"">zclei22/01 =
logicalreferenced =
396G &nbs=
p; -</p><p
class=3D"">zclei22/01 =
filesystem_limit =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
snapshot_limit =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
filesystem_count =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
snapshot_count =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
snapdev &=
nbsp; =
hidden &n=
bsp; default</p><p
class=3D"">zclei22/01 =
acltype &=
nbsp; =
off  =
;
default</p><p =
class=3D"">zclei22/01 =
context &=
nbsp; =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
fscontext  =
; =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
defcontext &nbs=
p; =
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
rootcontext
=
none &nbs=
p; default</p><p =
class=3D"">zclei22/01 =
relatime =
=
off  =
;
default</p><p =
class=3D"">zclei22/01
redundant_metadata =
all  =
;
default</p><p =
class=3D"">zclei22/01 =
overlay &=
nbsp; =
off  =
;
default</p></blockquote><p =
class=3D""><br class=3D""></p><br
class=3D""><br class=3D""></div><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878HOEnZb"><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878h5"><div class=3D"gmail_extra"><br
class=3D""><div =
class=3D"gmail_quote">On Fri, Mar 3, 2017 at 2:52 PM, Juan Pablo <span =
dir=3D"ltr" class=3D""><<a
href=3D"mailto:pablo.localhost@gmail.com" =
target=3D"_blank"
class=3D"">pablo.localhost(a)gmail.com</a>&gt;</span> =
wrote:<br class=3D""><blockquote class=3D"gmail_quote"
style=3D"margin:0px=
0px 0px 0.8ex;border-left:1px solid =
rgb(204,204,204);padding-left:1ex"><div dir=3D"ltr"
class=3D""><div =
class=3D""><div class=3D""><div
class=3D"">Which operating system =
version are you using for your zfs storage? <br
class=3D""></div>do:<br =
class=3D""></div>zfs get all your-pool-name<br
class=3D""></div>use =
arc_summary.py from freenas git repo if you wish.<br class=3D""><br =
class=3D""></div><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294HOEnZb"><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294h5"><div class=3D"gmail_extra"><br =
class=3D""><div class=3D"gmail_quote">2017-03-03 10:33
GMT-03:00 Arman =
Khalatyan <span dir=3D"ltr" class=3D""><<a =
href=3D"mailto:arm2arm@gmail.com" target=3D"_blank" =
class=3D"">arm2arm@gmail.com</a>></span>:<br
class=3D""><blockquote =
class=3D"gmail_quote" style=3D"margin:0px 0px 0px 0.8ex;border-left:1px =
solid rgb(204,204,204);padding-left:1ex"><div dir=3D"ltr"
class=3D"">Pool =
load:<br class=3D"">[root@clei21 ~]# zpool iostat -v 1 <br =
class=3D""> &nb=
sp;  =
; <wbr =
class=3D""> &nb=
sp; capacity
operations =
bandwidth<br =
class=3D"">pool  =
; &=
nbsp; <wbr =
class=3D"">
=
alloc free read
write =
read write<br class=3D"">------------------------------<wbr
=
class=3D"">-------- ----- ----- -----
=
----- ----- -----<br =
class=3D"">zclei21 &n=
bsp; &nbs=
p; <wbr =
class=3D"">
=
10.1G 3.62T
0 =
112 823 8.82M<br
class=3D""> =
HGST_HUS724040ALA640_PN2334PBJ<wbr class=3D"">52XWT1
10.1G =
3.62T
0 =
46 626 4.40M<br =
class=3D"">logs  =
; &=
nbsp; <wbr =
class=3D""> &nb=
sp; - =
-
- =
-
- -<br =
class=3D""> =
lv_slog &=
nbsp; <wb=
r
class=3D"">
225M =
9.72G
0 =
66 198 4.45M<br =
class=3D"">cache &nbs=
p; =
<wbr =
class=3D""> &nb=
sp; - =
-
- =
-
- -<br =
class=3D""> =
lv_cache =
<wbr
=
class=3D"">
9.81G =
204G
0 =
46 56 4.13M<br =
class=3D"">------------------------------<wbr
class=3D"">-------- =
----- ----- ----- ----- -----
-----<br =
class=3D""><br =
class=3D""> &nb=
sp;  =
; <wbr =
class=3D""> &nb=
sp; capacity
operations =
bandwidth<br =
class=3D"">pool  =
; &=
nbsp; <wbr =
class=3D"">
=
alloc free read
write =
read write<br class=3D"">------------------------------<wbr
=
class=3D"">-------- ----- ----- -----
=
----- ----- -----<br =
class=3D"">zclei21 &n=
bsp; &nbs=
p; <wbr =
class=3D"">
=
10.1G 3.62T
0 =
191 0 12.8M<br
class=3D""> =
HGST_HUS724040ALA640_PN2334PBJ<wbr class=3D"">52XWT1
10.1G =
3.62T
0 =
0
0 0<br =
class=3D"">logs  =
; &=
nbsp; <wbr =
class=3D""> &nb=
sp; - =
-
- =
-
- -<br =
class=3D""> =
lv_slog &=
nbsp; <wb=
r
class=3D"">
225M =
9.72G 0 =
191 0 12.8M<br =
class=3D"">cache &nbs=
p; =
<wbr =
class=3D""> &nb=
sp; - =
-
- =
-
- -<br =
class=3D""> =
lv_cache =
<wbr
=
class=3D"">
9.83G =
204G 0 =
218 0 20.0M<br =
class=3D"">------------------------------<wbr
class=3D"">-------- =
----- ----- ----- ----- -----
-----<br =
class=3D""><br =
class=3D""> &nb=
sp;  =
; <wbr =
class=3D""> &nb=
sp; capacity
operations =
bandwidth<br =
class=3D"">pool  =
; &=
nbsp; <wbr =
class=3D"">
=
alloc free read
write =
read write<br class=3D"">------------------------------<wbr
=
class=3D"">-------- ----- ----- -----
=
----- ----- -----<br =
class=3D"">zclei21 &n=
bsp; &nbs=
p; <wbr =
class=3D"">
=
10.1G 3.62T
0 =
191 0 12.7M<br
class=3D""> =
HGST_HUS724040ALA640_PN2334PBJ<wbr class=3D"">52XWT1
10.1G =
3.62T
0 =
0
0 0<br =
class=3D"">logs  =
; &=
nbsp; <wbr =
class=3D""> &nb=
sp; - =
-
- =
-
- -<br =
class=3D""> =
lv_slog &=
nbsp; <wb=
r
class=3D"">
225M =
9.72G 0 =
191 0 12.7M<br =
class=3D"">cache &nbs=
p; =
<wbr =
class=3D""> &nb=
sp; - =
-
- =
-
- -<br =
class=3D""> =
lv_cache =
<wbr
=
class=3D"">
9.83G =
204G
0 =
72 0 7.68M<br =
class=3D"">------------------------------<wbr
class=3D"">-------- =
----- ----- ----- ----- -----
-----<br =
class=3D""><br class=3D""></div><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647HOEnZb"><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647h5"><div =
class=3D"gmail_extra"><br class=3D""><div
class=3D"gmail_quote">On Fri, =
Mar 3, 2017 at 2:32 PM, Arman Khalatyan <span dir=3D"ltr" =
class=3D""><<a href=3D"mailto:arm2arm@gmail.com"
target=3D"_blank" =
class=3D"">arm2arm(a)gmail.com</a>&gt;</span> wrote:<br =
class=3D""><blockquote class=3D"gmail_quote"
style=3D"margin:0px 0px 0px =
0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div =
dir=3D"ltr" class=3D""><div class=3D"">Glusterfs
now in healing mode:<br =
class=3D""></div>Receiver:<br class=3D"">[root@clei21
~]# arcstat.py =
1<br class=3D""> time
read miss =
miss% dmis dm% pmis pm% mmis
=
mm% arcsz c <br =
class=3D"">13:24:49
0 =
0
0 =
0 0
0 =
0 0
0 =
4.6G 31G <br
class=3D"">13:24:50 =
154 80
51 =
80 51
0 =
0 80 51
4.6G =
31G <br class=3D"">13:24:51
179 =
62 34
62 =
34 0
0 =
62 42 4.6G 31G
<br =
class=3D"">13:24:52 148 =
68 45
68 =
45 0
0 =
68 45 4.6G 31G
<br =
class=3D"">13:24:53 140 =
64 45
64 =
45 0
0 =
64 45 4.6G 31G
<br =
class=3D"">13:24:54 124 =
48 38
48 =
38 0
0 =
48 38 4.6G 31G
<br =
class=3D"">13:24:55 157 =
80 50
80 =
50 0
0 =
80 50 4.7G 31G
<br =
class=3D"">13:24:56 202 =
68 33
68 =
33 0
0 =
68 41 4.7G 31G
<br =
class=3D"">13:24:57 127 =
54 42
54 =
42 0
0 =
54 42 4.7G 31G
<br =
class=3D"">13:24:58 126 =
50 39
50 =
39 0
0 =
50 39 4.7G 31G
<br =
class=3D"">13:24:59 116 =
40 34
40 =
34 0
0 =
40 34 4.7G 31G
<br =
class=3D""><br class=3D""><div
class=3D""><br class=3D"">Sender<br =
class=3D"">[root@clei22 ~]# arcstat.py 1<br
class=3D""> =
time read miss miss% dmis dm%
=
pmis pm% mmis mm%
arcsz =
c <br
class=3D"">13:28:37 =
8 2 =
25 2
25 =
0 0
2 =
25 468M 31G <br
class=3D"">13:28:38 =
1.2K 727
62 =
727 62
0 =
0 525 54
469M 31G =
<br class=3D"">13:28:39 815 =
508 62 508
=
62 0
0 =
376 55 469M 31G
<br =
class=3D"">13:28:40 994 =
624 62 624
=
62 0
0 =
450 54 469M 31G
<br =
class=3D"">13:28:41 783 =
456 58 456
=
58 0
0 =
338 50 470M 31G
<br =
class=3D"">13:28:42 916 =
541 59 541
=
59 0
0 =
390 50 470M 31G
<br =
class=3D"">13:28:43 768 =
437 56 437
=
57 0
0 =
313 48 471M 31G
<br =
class=3D"">13:28:44 877 =
534 60 534
=
60 0
0 =
393 53 470M 31G
<br =
class=3D"">13:28:45 957 =
630 65 630
=
65 0
0 =
450 57 470M 31G
<br =
class=3D"">13:28:46 819 =
479 58 479
=
58 0
0 =
357 51 471M 31G
<br =
class=3D""><br class=3D""></div></div><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713HOEnZ=
b"><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713h5"><=
div class=3D"gmail_extra"><br class=3D""><div
class=3D"gmail_quote">On =
Thu, Mar 2, 2017 at 7:18 PM, Juan Pablo <span dir=3D"ltr" =
class=3D""><<a href=3D"mailto:pablo.localhost@gmail.com"
=
target=3D"_blank"
class=3D"">pablo.localhost(a)gmail.com</a>&gt;</span> =
wrote:<br class=3D""><blockquote class=3D"gmail_quote"
style=3D"margin:0px=
0px 0px 0.8ex;border-left:1px solid =
rgb(204,204,204);padding-left:1ex"><div dir=3D"ltr"
class=3D"">hey,<br =
class=3D"">what are you using for zfs? get an arc status and show =
please<br class=3D""><br class=3D""></div><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781HOEnZb"><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781h5"><div class=3D"gmail_extra"><br
class=3D""><div =
class=3D"gmail_quote">2017-03-02 9:57 GMT-03:00 Arman Khalatyan <span =
dir=3D"ltr" class=3D""><<a
href=3D"mailto:arm2arm@gmail.com" =
target=3D"_blank"
class=3D"">arm2arm@gmail.com</a>></span>:<br =
class=3D""><blockquote class=3D"gmail_quote"
style=3D"margin:0px 0px 0px =
0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><div =
dir=3D"ltr" class=3D""><div class=3D""><div
class=3D"">no, <br =
class=3D""></div>ZFS itself is not on top of lvm. only ssd was spitted
=
by lvm for slog(10G) and cache (the rest)<br
class=3D""></div><div =
class=3D"">but in any-case the ssd does not help much on =
glusterfs/ovirt load it has almost 100% cache misses....:( =
(terrible performance compare with nfs)<br class=3D""><br =
class=3D""></div><div class=3D""><br
class=3D""></div><br class=3D""><br =
class=3D""></div><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570HOEnZb"><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570h5"><div
class=3D"gmail_extra"><br =
class=3D""><div class=3D"gmail_quote">On Thu, Mar 2, 2017 at
1:47 PM, =
FERNANDO FREDIANI <span dir=3D"ltr" class=3D""><<a =
href=3D"mailto:fernando.frediani@upx.com" target=3D"_blank" =
class=3D"">fernando.frediani(a)upx.com</a>&gt;</span>
wrote:<br =
class=3D""><blockquote class=3D"gmail_quote"
style=3D"margin:0px 0px 0px =
0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">
=20
=20
=20
<div bgcolor=3D"#FFFFFF" class=3D""><p
class=3D"">Am I understanding =
correctly, but you have Gluster on the top of
ZFS which is on the top of LVM ? If so, why the usage of LVM was
necessary ? I have ZFS with any need of LVM.</p><span =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570m_-6203522006917276901HOEnZb"><font =
color=3D"#888888" class=3D""><p
class=3D"">Fernando<br class=3D"">
</p></font></span><div class=3D""><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570m_-6203522006917276901h5">
<br class=3D"">
<div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570m_-6203522006917276901m_35422971570659266=
81moz-cite-prefix">On 02/03/2017 06:19, Arman Khalatyan
wrote:<br class=3D"">
</div>
</div></div><blockquote type=3D"cite"
class=3D""><div class=3D""><div =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570m_-6203522006917276901h5">
<div dir=3D"ltr" class=3D"">
<div class=3D"">
<div class=3D"">Hi, <br class=3D"">
</div>
I use 3 nodes with zfs and glusterfs.<br class=3D"">
</div>
Are there any suggestions to optimize it?<br class=3D"">
<div class=3D""><br class=3D"">
host zfs config 4TB-HDD+250GB-SSD:<br class=3D"">
[root@clei22 ~]# zpool status <br class=3D"">
pool: zclei22<br class=3D"">
state: ONLINE<br class=3D"">
scan: scrub repaired 0 in 0h0m with 0 errors on Tue Feb =
28
14:16:07 2017<br class=3D"">
config:<br class=3D"">
<br class=3D"">
=
NAME &nbs=
p; =
<wbr =
class=3D"">
=
STATE READ
WRITE CKSUM<br class=3D"">
=
zclei22 &=
nbsp; <wb=
r
class=3D"">
=
ONLINE
0
0 0<br class=3D"">
HGST_HUS724040ALA640_PN2334PBJ<wbr
=
class=3D"">4SV6T1
ONLINE =
0
0 0<br class=3D"">
logs<br class=3D"">
=
lv_slog &=
nbsp; <wb=
r
class=3D"">
=
ONLINE
0
0 0<br class=3D"">
cache<br class=3D"">
=
lv_cache =
<wbr
=
class=3D"">
=
ONLINE
0
0 0<br class=3D"">
<br class=3D"">
errors: No known data errors<br class=3D"">
<br class=3D"">
Name:<br class=3D"">
GluReplica<br class=3D"">
Volume ID:<br class=3D"">
ee686dfe-203a-4caa-a691-263534<wbr class=3D"">60cc48<br =
class=3D"">
Volume Type:<br class=3D"">
Replicate (Arbiter)<br class=3D"">
Replica Count:<br class=3D"">
2 + 1<br class=3D"">
Number of Bricks:<br class=3D"">
3<br class=3D"">
Transport Types:<br class=3D"">
TCP, RDMA<br class=3D"">
Maximum no of snapshots:<br class=3D"">
256<br class=3D"">
Capacity:<br class=3D"">
3.51 TiB total, 190.56 GiB used, 3.33 TiB free<br class=3D"">
</div>
</div>
<br class=3D"">
<fieldset =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570m_-6203522006917276901m_35422971570659266=
81mimeAttachmentHeader"></fieldset>
<br class=3D"">
</div></div><span class=3D""><pre =
class=3D"">______________________________<wbr
class=3D"">_________________=
Users mailing list
<a =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570m_-6203522006917276901m_35422971570659266=
81moz-txt-link-abbreviated" href=3D"mailto:Users@ovirt.org" =
target=3D"_blank">Users(a)ovirt.org</a>
<a =
class=3D"gmail-m_6281861324822600694m_9102919535904979465m_-77235589057693=
4878m_8401411119881083294m_-2168826882466388647m_-4808390406438333713m_690=
6981569320781m_349397128160904570m_-6203522006917276901m_35422971570659266=
81moz-txt-link-freetext" =
href=3D"http://lists.ovirt.org/mailman/listinfo/users" =
target=3D"_blank">http://lists.ovirt.org/mailman<wbr =
class=3D"">/listinfo/users</a>
</pre>
</span></blockquote>
<br class=3D"">
</div>
<br class=3D"">______________________________<wbr =
class=3D"">_________________<br class=3D"">
Users mailing list<br class=3D"">
<a href=3D"mailto:Users@ovirt.org" target=3D"_blank" =
class=3D"">Users(a)ovirt.org</a><br class=3D"">
<a
href=3D"http://lists.ovirt.org/mailman/listinfo/users" =
rel=3D"noreferrer" target=3D"_blank" =
class=3D"">http://lists.ovirt.org/mailman<wbr =
class=3D"">/listinfo/users</a><br class=3D"">
<br class=3D""></blockquote></div><br
class=3D""></div>
</div></div><br
class=3D"">______________________________<wbr =
class=3D"">_________________<br class=3D"">
Users mailing list<br class=3D"">
<a href=3D"mailto:Users@ovirt.org" target=3D"_blank" =
class=3D"">Users(a)ovirt.org</a><br class=3D"">
<a
href=3D"http://lists.ovirt.org/mailman/listinfo/users" =
rel=3D"noreferrer" target=3D"_blank" =
class=3D"">http://lists.ovirt.org/mailman<wbr =
class=3D"">/listinfo/users</a><br class=3D"">
<br class=3D""></blockquote></div><br
class=3D""></div>
</div></div></blockquote></div><br
class=3D""></div>
</div></div></blockquote></div><br
class=3D""></div>
</div></div></blockquote></div><br
class=3D""></div>
</div></div></blockquote></div><br
class=3D""></div>
</div></div></blockquote></div><br
class=3D""></div>
</div></div></blockquote></div><br
class=3D""></div>
</div></div></blockquote></div><br
class=3D""></div>
</div></div></blockquote></div><br =
class=3D""></div></div></div></div></div></div>
_______________________________________________<br class=3D"">Users =
mailing list<br class=3D""><a href=3D"mailto:Users@ovirt.org"
=
class=3D"">Users(a)ovirt.org</a><br =
class=3D"">http://lists.ovirt.org/mailman/listinfo/users<br =
class=3D""></div></blockquote></div><br
class=3D""></div></body></html>=
--Apple-Mail=_33A5E632-5864-4B33-9FED-1901AA57DEBD--