--_000_DB6P190MB0280F066762EF81DE4270280C88A0DB6P190MB0280EURP_
Content-Type: text/plain; charset="Windows-1252"
Content-Transfer-Encoding: quoted-printable
Thanks once again Johan!
What would be your approach: JBOD straight or JBOD made of RAIDed bricks?
Moacir
________________________________
From: Johan Bernhardsson <johan(a)kafit.se
Sent:
Tuesday, August 8, 2017 11:24 AM
To: Moacir Ferreira; Devin Acosta; users(a)ovirt.org
Subject: Re: [ovirt-users] Good practices
On ovirt gluster uses sharding. So all large files are broken up in small p=
ieces on the gluster bricks.
/Johan
On August 8, 2017 12:19:39 Moacir Ferreira <moacirferreira(a)hotmail.com> wro=
te:
Thanks Johan, you brought "light" into my darkness! I went looking for the =
GlusterFS tiering how-to and it looks like quite simple to attach a SSD as =
hot tier. For those willing to read about it, go here:
http://blog.gluster.=
org/2016/03/automated-tiering-in-gluster/
Now, I still have a question: VMs are made of very large .qcow2 files. My u=
nderstanding is that files in Gluster are kept all together in a single bri=
ck. If so, I will not benefit from tiering as a single SSD will not be big =
enough to fit all my large VM .qcow2 files. This would not be true if Glust=
er can store "blocks" of data that compose a large file spread on several b=
ricks. But if I am not wrong, this is one of key differences in between Glu=
sterFS and Ceph. Can you comment?
Moacir
________________________________
From: Johan Bernhardsson <johan(a)kafit.se
Sent:
Tuesday, August 8, 2017 7:03 AM
To: Moacir Ferreira; Devin Acosta; users(a)ovirt.org
Subject: Re: [ovirt-users] Good practices
You attach the ssd as a hot tier with a gluster command. I don't think that=
gdeploy or ovirt gui can do it.
The gluster docs and redhat docs explains tiering quite good.
/Johan
On August 8, 2017 07:06:42 Moacir Ferreira <moacirferreira(a)hotmail.com> wro=
te:
Hi Devin,
Please consider that for the OS I have a RAID 1. Now, lets say I use RAID 5=
to assemble a single disk on each server. In this case, the SSD will not m=
ake any difference, right? I guess that to be possible to use it, the SSD s=
hould not be part of the RAID 5. In this case I could create a logical volu=
me made of the RAIDed brick and then extend it using the SSD. I.e.: Using g=
deploy:
[disktype]
jbod
....
[pv1]
action=3Dcreate
devices=3Dsdb, sdc
wipefs=3Dyes
ignore_vg_erros=3Dno
[vg1]
action=3Dcreate
vgname=3Dgluster_vg_jbod
pvname=3Dsdb
ignore_vg_erros=3Dno
[vg2]
action=3Dextend
vgname=3Dgluster_vg_jbod
pvname=3Dsdc
ignore_vg_erros=3Dno
But will Gluster be able to auto-detect and use this SSD brick for tiering?=
Do I have to do some other configurations? Also, as the VM files (.qcow2) =
are quite big will I benefit from tiering? This is wrong and my approach sh=
ould be other?
Thanks,
Moacir
________________________________
From: Devin Acosta <devin(a)pabstatencio.com
Sent:
Monday, August 7, 2017 7:46 AM
To: Moacir Ferreira; users(a)ovirt.org
Subject: Re: [ovirt-users] Good practices
Moacir,
I have recently installed multiple Red Hat Virtualization hosts for several=
different companies, and have dealt with the Red Hat Support Team in depth=
about optimal configuration in regards to setting up GlusterFS most effici=
ently and I wanted to share with you what I learned.
In general Red Hat Virtualization team frowns upon using each DISK of the s=
ystem as just a JBOD, sure there is some protection by having the data repl=
icated, however, the recommendation is to use RAID 6 (preferred) or RAID-5,=
or at least RAID-1 at the very least.
Here is the direct quote from Red Hat when I asked about RAID and Bricks:
"A typical Gluster configuration would use RAID underneath the bricks. RAID=
6 is most typical as it gives you 2 disk failure protection, but RAID 5 co=
uld be used too. Once you have the RAIDed bricks, you'd then apply the desi=
red replication on top of that. The most popular way of doing this would be=
distributed replicated with 2x replication. In general you'll get better p=
erformance with larger bricks. 12 drives is often a sweet spot. Another opt=
ion would be to create a separate tier using all SSD=92s.=94
In order to SSD tiering from my understanding you would need 1 x NVMe drive=
in each server, or 4 x SSD hot tier (it needs to be distributed, replicate=
d for the hot tier if not using NVME). So with you only having 1 SSD drive =
in each server, I=92d suggest maybe looking into the NVME option.
Since your using only 3-servers, what I=92d probably suggest is to do (2 Re=
plicas + Arbiter Node), this setup actually doesn=92t require the 3rd serve=
r to have big drives at all as it only stores meta-data about the files and=
not actually a full copy.
Please see the attached document that was given to me by Red Hat to get mor=
e information on this. Hope this information helps you.
--
Devin Acosta, RHCA, RHVCA
Red Hat Certified Architect
On August 6, 2017 at 7:29:29 PM, Moacir Ferreira (moacirferreira(a)hotmail.co=
m<mailto:moacirferreira@hotmail.com>) wrote:
I am willing to assemble a oVirt "pod", made of 3 servers, each with 2 CPU =
sockets of 12 cores, 256GB RAM, 7 HDD 10K, 1 SSD. The idea is to use Gluste=
rFS to provide HA for the VMs. The 3 servers have a dual 40Gb NIC and a dua=
l 10Gb NIC. So my intention is to create a loop like a server triangle usin=
g the 40Gb NICs for virtualization files (VMs .qcow2) access and to move VM=
s around the pod (east /west traffic) while using the 10Gb interfaces for g=
iving services to the outside world (north/south traffic).
This said, my first question is: How should I deploy GlusterFS in such oVir=
t scenario? My questions are:
1 - Should I create 3 RAID (i.e.: RAID 5), one on each oVirt node, and then=
create a GlusterFS using them?
2 - Instead, should I create a JBOD array made of all server's disks?
3 - What is the best Gluster configuration to provide for HA while not cons=
uming too much disk space?
4 - Does a oVirt hypervisor pod like I am planning to build, and the virtua=
lization environment, benefits from tiering when using a SSD disk? And yes,=
will Gluster do it by default or I have to configure it to do so?
At the bottom line, what is the good practice for using GlusterFS in small =
pods for enterprises?
You opinion/feedback will be really appreciated!
Moacir
_______________________________________________
Users mailing list
Users@ovirt.org<mailto:Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________
Users mailing list
Users@ovirt.org<mailto:Users%40ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
--_000_DB6P190MB0280F066762EF81DE4270280C88A0DB6P190MB0280EURP_
Content-Type: text/html; charset="Windows-1252"
Content-Transfer-Encoding: quoted-printable
<html
<head
<meta http-equiv=3D"Content-Type"
content=3D"text/html; charset=3DWindows-1=
252"
<style
type=3D"text/css" style=3D"display:none;"><!-- P
{margin-top:0;margi=
n-bottom:0;} --></style
</head
<body dir=3D"ltr"
<div
id=3D"divtagdefaultwrapper" style=3D"font-size:12pt;color:#000000;font=
-family:Calibri,Helvetica,sans-serif;" dir=3D"ltr"
<p>Thanks once again Johan!<br
</p
<br
<p>What would be your approach: JBOD straight or JBOD made of RAIDed bricks=
?</p
<p><br
</p
<p>Moacir<br
</p
<br
<div
style=3D"color: rgb(49, 55, 57);"
<hr
tabindex=3D"-1" style=3D"display:inline-block; width:98%"
<div id=3D"divRplyFwdMsg"
dir=3D"ltr"><font style=3D"font-size:11pt" face=
=3D"Calibri, sans-serif" color=3D"#000000"><b>From:</b>
Johan Bernhardsson =
&lt;johan(a)kafit.se&gt;<br
<b>Sent:</b> Tuesday, August 8, 2017 11:24 AM<br
<b>To:</b> Moacir Ferreira; Devin Acosta;
users(a)ovirt.org<br
<b>Subject:</b> Re:
[ovirt-users] Good practices</font
<div> </div
</div
<div
<div
style=3D"color:black"
<div
style=3D"color:black"
<p style=3D"margin:0 0 1em
0; color:black">On ovirt gluster uses sharding. =
So all large files are broken up in small pieces on the gluster bricks.</p
<p style=3D"margin:0 0 1em 0;
color:black">/Johan</p
</div
<div style=3D"color:black"
<p style=3D"color:black; font-size:10pt;
font-family:Arial,sans-serif; marg=
in:10pt 0"
On August 8, 2017 12:19:39 Moacir
Ferreira &lt;moacirferreira(a)hotmail.com&g=
t; wrote:</p
<blockquote
type=3D"cite" class=3D"gmail_quote" style=3D"margin:0 0 0 0.75e=
x; border-left:1px solid #808080; padding-left:0.75ex"
<div
id=3D"divtagdefaultwrapper" dir=3D"ltr" style=3D"font-size:12pt;
color=
:#000000; font-family:Calibri,Helvetica,sans-serif"
<p>Thanks Johan, you brought "light" into my darkness! I
went loo=
king for the GlusterFS tiering how-to and it looks like quite simple to att=
ach a SSD as hot tier. For those willing to read about it, go here:
<a
href=3D"http://blog.gluster.org/2016/03/automated-tiering-in-gluster... c=
lass=3D"OWAAutoLink" id=3D"LPlnk939343"
previewremoved=3D"true"
http://blog.gluster.org/2016/03/automated-tiering-in-gluster/</a>&l...
<p><br
</p
<p>Now, I still have a
question: VMs are made of very large .qcow2 files. M=
y understanding is that files in Gluster are kept all together in a single =
brick. If so, I will not benefit from tiering as a single SSD will not be b=
ig enough to fit all my large VM
.qcow2 files. This would not be true if Gluster can store "blocks&quo=
t; of data that compose a large file spread on several bricks. But if I am =
not wrong, this is one of key differences in between GlusterFS and Ceph. Ca=
n you comment?</p
<p><br
</p
<p>Moacir<br
</p
<br
<br
<div
style=3D"color:rgb(49,55,57)"
<hr
tabindex=3D"-1" style=3D"display:inline-block; width:98%"
<div id=3D"divRplyFwdMsg"
dir=3D"ltr"><font style=3D"font-size:11pt" face=
=3D"Calibri, sans-serif" color=3D"#000000"><b>From:</b>
Johan Bernhardsson =
&lt;johan(a)kafit.se&gt;<br
<b>Sent:</b> Tuesday, August 8, 2017 7:03 AM<br
<b>To:</b> Moacir Ferreira; Devin Acosta;
users(a)ovirt.org<br
<b>Subject:</b> Re:
[ovirt-users] Good practices</font
<div> </div
</div
<div
<div
style=3D"color:black"
<div
style=3D"color:black"
<p style=3D"margin:0 0 1em
0; color:black">You attach the ssd as a hot tier=
with a gluster command. I don't think that gdeploy or ovirt gui can do it.=
</p
<p style=3D"margin:0 0 1em 0; color:black">The
gluster docs and redhat docs=
explains tiering quite good.</p
<p
style=3D"margin:0 0 1em 0; color:black">/Johan</p
</div
<div
style=3D"color:black"
<p style=3D"color:black;
font-size:10pt; font-family:Arial,sans-serif; marg=
in:10pt 0"
On August 8, 2017 07:06:42 Moacir
Ferreira &lt;moacirferreira(a)hotmail.com&g=
t; wrote:</p
<blockquote
type=3D"cite" class=3D"gmail_quote" style=3D"margin:0 0 0 0.75e=
x; border-left:1px solid #808080; padding-left:0.75ex"
<div
id=3D"divtagdefaultwrapper" dir=3D"ltr" style=3D"font-size:12pt;
color=
:#000000; font-family:Calibri,Helvetica,sans-serif"
<p>Hi Devin,</p
<p><br
</p
<p>Please consider that for
the OS I have a RAID 1. Now, lets say I use RAI=
D 5 to assemble a single disk on each server. In this case, the SSD will no=
t make any difference, right? I guess that to be possible to use it, the SS=
D should not be part of the RAID
5. In this case I could create a logical volume made of the RAIDed brick a=
nd then extend it using the SSD. I.e.: Using gdeploy:</p
<p><br
</p
<p>[disktype]</p
<p>jbod</p
<p>....</p
<p>[pv1]</p
<p>action=3Dcreate</p
<p>devices=3Dsdb, sdc</p
<p>wipefs=3Dyes</p
<p></p
<p>ignore_vg_erros=3Dno<br
</p
<p><br
</p
<p></p
<p>[vg1]</p
<p><b>action=3Dcreate</b></p
<p>vgname=3Dgluster_vg_jbod</p
<p>pvname=3Dsdb</p
<p>ignore_vg_erros=3Dno<br
</p
<p><br
</p
<p>[vg2]</p
<p><b>action=3Dextend</b></p
<p>vgname=3Dgluster_vg_jbod</p
<p>pvname=3Dsdc</p
<p>ignore_vg_erros=3Dno<br
</p
<br
<p>But will Gluster be able to auto-detect and use this SSD brick for tieri=
ng? Do I have to do some other configurations? Also, as the VM files (.qcow=
2) are quite big will I benefit from tiering? This is wrong and my approach=
should be other?</p
<p><br
</p
<p>Thanks,</p
<p>Moacir<br
</p
<p><br
</p
<br
<div
style=3D"color:rgb(49,55,57)"
<hr
tabindex=3D"-1" style=3D"display:inline-block; width:98%"
<div id=3D"divRplyFwdMsg"
dir=3D"ltr"><font style=3D"font-size:11pt" face=
=3D"Calibri, sans-serif" color=3D"#000000"><b>From:</b>
Devin Acosta <de=
vin(a)pabstatencio.com&gt;<br
<b>Sent:</b> Monday,
August 7, 2017 7:46 AM<br
<b>To:</b> Moacir
Ferreira; users(a)ovirt.org<br
<b>Subject:</b> Re:
[ovirt-users] Good practices</font
<div> </div
</div
<div
<div
id=3D"bloop_customfont" style=3D"color:rgb(0,0,0);
margin:0px"><font f=
ace=3D"Input Mono"><br
</font></div
<div
id=3D"bloop_customfont" style=3D"color:rgb(0,0,0);
margin:0px"><font f=
ace=3D"Input Mono">Moacir,</font></div
<div
id=3D"bloop_customfont" style=3D"color:rgb(0,0,0);
margin:0px"><font f=
ace=3D"Input Mono"><br
</font></div
<div
id=3D"bloop_customfont" style=3D"color:rgb(0,0,0);
margin:0px"><font f=
ace=3D"Input Mono">I have recently installed multiple Red Hat Virtualizatio=
n hosts for several different companies, and have dealt with the Red Hat Su=
pport Team in depth about optimal configuration
in regards to setting up GlusterFS most efficiently and I wanted to share =
with you what I learned.</font></div
<div
id=3D"bloop_customfont" style=3D"color:rgb(0,0,0);
margin:0px"><font f=
ace=3D"Input Mono"><br
</font></div
<div
id=3D"bloop_customfont" style=3D"color:rgb(0,0,0);
margin:0px"><font f=
ace=3D"Input Mono">In general Red Hat Virtualization team frowns upon using=
each DISK of the system as just a JBOD, sure there is some protection by h=
aving the data replicated, however, the
recommendation is to use RAID 6 (preferred) or RAID-5, or at least RAID-1 =
at the very least.</font></div
<div
id=3D"bloop_customfont" style=3D"color:rgb(0,0,0);
margin:0px"><font f=
ace=3D"Input Mono"><br
</font></div
<div
id=3D"bloop_customfont" style=3D"margin:0px"><font
face=3D"Input Mono"=
Here is the direct quote from Red Hat when I asked about RAID
and Bri=
cks:</font></div
<div
id=3D"bloop_customfont" style=3D"margin:0px"><font
face=3D"Input Mono"=
<i><br>
</i></font></div
<div id=3D"bloop_customfont"
style=3D"margin:0px"><font face=3D"Input Mono"=
<i>"A typical Gluster configuration would use RAID
underneath the bri=
cks. RAID 6 is most typical as it gives you 2 disk failure
protection, but =
RAID 5 could be used too. Once you have the
RAIDed bricks, you'd then apply the desired replication on top of that. Th=
e most popular way of doing this would be distributed replicated with 2x re=
plication. In general you'll get better performance with larger bricks=
. 12 drives is often a sweet spot. Another
option would be to create a separate tier using all SSD=92s.=94 </i><=
/font></div
<div
id=3D"bloop_customfont" style=3D"margin:0px"><br
</div
<div
id=3D"bloop_customfont" style=3D"margin:0px"><font
face=3D"Input Mono"=
<i>In order to SSD tiering from my understanding you
would need 1 x N=
VMe drive in each server, or 4 x SSD hot tier (it needs to be
distributed, =
replicated for the hot tier if not using NVME).
So with you only having 1 SSD drive in each server, I=92d suggest may=
be looking into the NVME option. </i></font></div
<div id=3D"bloop_customfont"
style=3D"margin:0px"><font face=3D"Input Mono"=
<i><br>
</i></font></div
<div id=3D"bloop_customfont"
style=3D"margin:0px"><font face=3D"Input Mono"=
<i>Since your using only 3-servers, what I=92d probably suggest
is to do (=
2 Replicas + Arbiter Node), this setup actually doesn=92t
require the 3=
rd server to have big drives at all as it only
stores meta-data about the files and not actually a full copy. </i></=
font></div
<div
id=3D"bloop_customfont" style=3D"margin:0px"><font
face=3D"Input Mono"=
<i><br>
</i></font></div
<div id=3D"bloop_customfont"
style=3D"margin:0px"><font face=3D"Input Mono"=
<i>Please see the attached document that was given to me by Red
Hat to get=
more information on this. Hope this information helps
you.</i></font></div=
<div id=3D"bloop_customfont"
style=3D"margin:0px"><font face=3D"Input Mono"=
<i><br>
</i></font></div
<br
<div
id=3D"bloop_sign_1502087376725469184" class=3D"bloop_sign"><span
style=
=3D"font-family:'helvetica Neue',helvetica;
font-size:14px">--</span><br st=
yle=3D"font-family:'helvetica Neue',helvetica; font-size:14px"
<div class=3D"gmail_signature"
style=3D"font-family:'helvetica Neue',helvet=
ica; font-size:14px"
<div dir=3D"ltr"
<div><br
</div
<div>Devin Acosta, RHCA,
RHVCA</div
<div>Red Hat Certified
Architect</div
<div></div
</div
</div
</div
<br
<p
class=3D"airmail_on">On August 6, 2017 at 7:29:29 PM, Moacir Ferreira (<=
a
href=3D"mailto:moacirferreira@hotmail.com">moacirferreira@hotmail.com</a>=
) wrote:</p
<blockquote
type=3D"cite" class=3D"clean_bq"><span
<div dir=3D"ltr"
<div></div
<div
<div
id=3D"divtagdefaultwrapper" dir=3D"ltr" style=3D"font-size:12pt;
color=
:#000000; font-family:Calibri,Helvetica,sans-serif"
<p><span>I am willing to assemble a oVirt "pod", made
of 3 server=
s, each with 2 CPU sockets of 12 cores, 256GB RAM, 7 HDD 10K, 1 SSD. The id=
ea is to use GlusterFS to provide HA for the VMs. The 3 servers have a dual=
40Gb NIC and a dual 10Gb NIC. So my intention
is to create a loop like a server triangle using the 40Gb NICs for virtual=
ization files (VMs .qcow2) access and to move VMs around the pod (east /wes=
t traffic) while using the 10Gb interfaces for giving services to the outsi=
de world (north/south traffic).</span></p
<p><br
</p
<p>This said, my first question is: How should I deploy GlusterFS in such o=
Virt scenario? My questions are:</p
<p><br
</p
<p>1 - Should I create 3 RAID (i.e.: RAID 5), one on each oVirt node, and t=
hen create a GlusterFS using them?</p
<p>2 - Instead, should I create a JBOD array made of all server's
disks?</p=
<p>3 - What is the best Gluster configuration to provide
for HA while not c=
onsuming too much disk space?<br
</p
<p>4 - Does a oVirt hypervisor
pod like I am planning to build, and the vir=
tualization environment, benefits from tiering when using a SSD disk? And y=
es, will Gluster do it by default or I have to configure it to do so?</p
<p><br
</p
<p>At the bottom line, what is
the good practice for using GlusterFS in sma=
ll pods for enterprises?<br
</p
<p><br
</p
<p>You opinion/feedback will be really appreciated!</p
<p>Moacir<br
</p
</div
_______________________________________________ <br
Users mailing list <br
<a
href=3D"mailto:Users@ovirt.org">Users@ovirt.org</a> <br
<a
href=3D"http://lists.ovirt.org/mailman/listinfo/users">http:...
t.org/mailman/listinfo/users</a
<br
</div
</div
</span></blockquote
</div
</div
</div
_______________________________________________<br
Users
mailing list<br
<a class=3D"aqm-autolink
aqm-autowrap" href=3D"mailto:Users%40ovirt.org">Us=
ers(a)ovirt.org</a><br
<a class=3D"aqm-autolink
aqm-autowrap"
href=3D"http://lists.ovirt.org/mailm=
an/listinfo/users">http://lists.ovirt.org/mailman/listinfo/users&...
<br
</blockquote
</div
</div
</div
</div
</div
</blockquote
</div
</div
</div
</div
</div
</body
</html
--_000_DB6P190MB0280F066762EF81DE4270280C88A0DB6P190MB0280EURP_--