The VMs are basically as stock CentOS 7x as you can get. There are so many
layers to deal with in HCI it's difficult to know where to begin with
tuning. I was focusing mainly on gluster. Is it recommended to do tuning
directly on oVirt host nodes as well such as I/O scheduler and tuned-adm
profiles etc?
On Fri, Oct 18, 2019 at 6:55 AM Strahil <hunter86_bg(a)yahoo.com> wrote:
What is your I/O scheduler and tuned-adm profile in the VM.
RedHat based VMs use deadline which prioritizes reads before writes ->
you can use 'noop' or 'none'.
For profile, you can use high-performance.
Best Regards,
Strahil Nikolov
On Oct 18, 2019 06:45, Jayme <jaymef(a)gmail.com> wrote:
I'm wondering if anyone has any tips to improve file/directory operations
in HCI replica 3 (no arbtr) configuration with SSDs and 10Gbe storage
network.
I am running stock optimize for virt store volume settings currently and
am wondering what if any improvements I can make for VM write speed and
more specifically anything I can tune to increase performance of small file
operations such as copying, untar, npm installs etc.
For some context, I'm seeing ~50MB/s write speeds inner VM with: dd
if=/dev/zero of=./test bs=512k count=2048 oflag=direct -- I am not sure how
this compares to other HCI setups, I feel like it should be higher with SSD
backed storage. Same command from gluster mount is over 400MB/s
I've read some things about meta data caching, read ahead and other
options. There are so many and I'm not sure where to start, I'm also not
sure which could potentially have a negative impact on VM
stability/reliability.
Here are options for one of my volumes:
Volume Name: prod_b
Type: Replicate
Volume ID: c3e7447e-8514-4e4a-9ff5-a648fe6aa537
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: gluster0.example.com:/gluster_bricks/prod_b/prod_b
Brick2: gluster1.example.com:/gluster_bricks/prod_b/prod_b
Brick3: gluster2.example.com:/gluster_bricks/prod_b/prod_b
Options Reconfigured:
server.event-threads: 4
client.event-threads: 4
performance.client-io-threads: on
nfs.disable: on
transport.address-family: inet
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.low-prio-threads: 32
network.remote-dio: off
cluster.eager-lock: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
cluster.data-self-heal-algorithm: full
cluster.locking-scheme: granular
cluster.shd-max-threads: 8
cluster.shd-wait-qlength: 10000
features.shard: on
user.cifs: off
storage.owner-uid: 36
storage.owner-gid: 36
network.ping-timeout: 30
performance.strict-o-direct: on
cluster.granular-entry-heal: enable
server.allow-insecure: on
cluster.choose-local: off