On Tue, Mar 15, 2022 at 7:43 AM Jonas Liechti <jonas(a)rabe.ch> wrote:
Thank you for the confirmation Strahil!
As our current environment is more or less the same (except the Hardware
RAID, which is not possible with NVMe disks), we planned to use Gluster. I
guess we will proceed as we originally planned as we are satisfied with the
performance.
Sorry for being late to the party.
Two items to consider:
A. If you have more than two drives, consider using MDRAID 10 on NVME drives
We have been using it for a while now and getting *very* good performance
(more or less in-line with a single NVME drive).
As an added bonus, we managed to survive a single drive failure without
fault.
B .If possible use a 40GbE network between the nodes instead of using
10GbE. It makes a huge impact. (SSDs or even a large HDD array can
easily saturate a single 10GbE link).
- Gilboa
On 3/11/22 07:08, Strahil Nikolov via Users wrote:
Red Hat Gluster Storage is discontinued, but the Gluster (upstream) is
pretty active and as Sandro Bonazzola (RH) confirmed -> there are no plans
to remove support for Gluster.
I think it's still a good choice, especially if you don't have SAN/
Higly-Available NFS.
Also, storage migration is transparent for the VMs, so you can add SAN on
a later stage and move all VMs from Gluster to SAN without disruption* .
Keep in mind that Gluster is a tier2 storage and if you really need a lot
of IOPS, CEPH might be suitable.
Best Regards,
Strahil Nikolov
*: Note that this is valid if the FUSE client is used. Other oVirt users
report huge performance increase when using libgfapi interface, but it has
drawbacks like storage migration can happen only when you switch off
libgfapi, power off the VM (on a scheduled basis), power on the VM, live
migrate the VM to other storage type, enable libgfapi again for the rest of
the VMs.
Thanks to Nikolov and Strahil for the valuable input! I was off for a few
weeks, so I would like to apologize if I'm potentially reviving a zombie
thread.
I am a bit confused about where to go with this environment after the
discontinuation of the hyperconverged setup. What alternative options are
there for us? Or do you think going the Gluster way would still be
advisable, even though it seems as it is being discontinued over time?
Thanks for any input on this!
Best regards,
Jonas
On 1/22/22 14:31, Strahil Nikolov via Users wrote:
Using the wizzard is utilizing the Gluster Andible roles.
I would highly recommend using it, unless you know what you are doing (for
example storage alignment when using Hardware raid).
Keep in mind that the DHT xlator (the logic in distributed volumes) is
shard aware, so your shards are spread between subvolumes and additional
performance can be gained.So using replicated-distributed volumes have
their benefits.
If you decide to avoid the software raid, use only replica3 volumes as
with SSDs/NVMEs usually the failures are not physical, but logical (maximum
writes reached -> predictive failure -> total failure).
Also, consider mounting via noatime/relatime and
context="system_u:object_r:glusterd_brick_t:s0" for your gluster bricks.
Best Regards,
Strahil Nikolov
On Fri, Jan 21, 2022 at 11:00, Gilboa Davara
<gilboad(a)gmail.com> <gilboad(a)gmail.com> wrote:
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/U2ZEWLRF5D6...
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/26AHNDSOJSI...
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org
Privacy Statement:
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JFFS2T25TII...