+Krutika Dhananjay +Sachidananda URS 

Adrian, can you provide more details on the performance issue you're seeing?
We do have some scripts to collect data to analyse. Perhaps you can run this and provide us the details in a bug. The ansible scripts to do this is still under review - https://github.com/gluster/gluster-ansible-maintenance/pull/4
But would be great if you can provide us some feedback on this as well.

On Thu, Aug 22, 2019 at 7:50 AM <adrianquintero@gmail.com> wrote:
Hello,
I have a hyperconverged setup using ovirt 4.3.5 and the "optimize for ovirt store" seems to fail on gluster volumes.
I am seeing poor performance and trying to see how should I tune gluster to give better performance.
Can you provide any suggestions on the following volume settings(parameters)?

option     Value
------     -----
cluster.lookup-unhashed    on
cluster.lookup-optimize    on
cluster.min-free-disk      10%
cluster.min-free-inodes    5%
cluster.rebalance-stats    off
cluster.subvols-per-directory      (null)
cluster.readdir-optimize           off
cluster.rsync-hash-regex           (null)
cluster.extra-hash-regex           (null)
cluster.dht-xattr-name     trusted.glusterfs.dht
cluster.randomize-hash-range-by-gfid       off
cluster.rebal-throttle     normal
cluster.lock-migration     off
cluster.force-migration    off
cluster.local-volume-name          (null)
cluster.weighted-rebalance         on
cluster.switch-pattern     (null)
cluster.entry-change-log           on
cluster.read-subvolume     (null)
cluster.read-subvolume-index    -1
cluster.read-hash-mode  1
cluster.background-self-heal-count      8
cluster.metadata-self-heal         off
cluster.data-self-heal     off
cluster.entry-self-heal    off
cluster.self-heal-daemon           on
cluster.heal-timeout    600
cluster.self-heal-window-size   1
cluster.data-change-log    on
cluster.metadata-change-log        on
cluster.data-self-heal-algorithm           full
cluster.eager-lock         enable
disperse.eager-lock        on
disperse.other-eager-lock          on
disperse.eager-lock-timeout     1
disperse.other-eager-lock-timeout       1
cluster.quorum-type        auto
cluster.quorum-count       (null)
cluster.choose-local       off
cluster.self-heal-readdir-size     1KB
cluster.post-op-delay-secs      1
cluster.ensure-durability          on
cluster.consistent-metadata        no
cluster.heal-wait-queue-length  128
cluster.favorite-child-policy      none
cluster.full-lock          yes
diagnostics.latency-measurement    off
diagnostics.dump-fd-stats          off
diagnostics.count-fop-hits         off
diagnostics.brick-log-level        INFO
diagnostics.client-log-level       INFO
diagnostics.brick-sys-log-level    CRITICAL
diagnostics.client-sys-log-level           CRITICAL
diagnostics.brick-logger           (null)
diagnostics.client-logger          (null)
diagnostics.brick-log-format       (null)
diagnostics.client-log-format      (null)
diagnostics.brick-log-buf-size  5
diagnostics.client-log-buf-size 5
diagnostics.brick-log-flush-timeout     120
diagnostics.client-log-flush-timeout    120
diagnostics.stats-dump-interval 0
diagnostics.fop-sample-interval 0
diagnostics.stats-dump-format      json
diagnostics.fop-sample-buf-size 65535
diagnostics.stats-dnscache-ttl-sec      86400
performance.cache-max-file-size 0
performance.cache-min-file-size 0
performance.cache-refresh-timeout       1
performance.cache-priority     
performance.cache-size     32MB
performance.io-thread-count     16
performance.high-prio-threads   16
performance.normal-prio-threads 16
performance.low-prio-threads    32
performance.least-prio-threads  1
performance.enable-least-priority          on
performance.iot-watchdog-secs      (null)
performance.iot-cleanup-disconnected-   reqsoff
performance.iot-pass-through       false
performance.io-cache-pass-through          false
performance.cache-size     128MB
performance.qr-cache-timeout    1
performance.cache-invalidation     false
performance.ctime-invalidation     false
performance.flush-behind           on
performance.nfs.flush-behind       on
performance.write-behind-window-size       1MB
performance.resync-failed-syncs-after   -fsyncoff
performance.nfs.write-behind-window-s   ize1MB
performance.strict-o-direct        on
performance.nfs.strict-o-direct    off
performance.strict-write-ordering          off
performance.nfs.strict-write-ordering      off
performance.write-behind-trickling-wr   iteson
performance.aggregate-size         128KB
performance.nfs.write-behind-tricklin   g-writeson
performance.lazy-open      yes
performance.read-after-open        yes
performance.open-behind-pass-through       false
performance.read-ahead-page-count       4
performance.read-ahead-pass-through        false
performance.readdir-ahead-pass-throug   h  false
performance.md-cache-pass-through          false
performance.md-cache-timeout    1
performance.cache-swift-metadata           true
performance.cache-samba-metadata           false
performance.cache-capability-xattrs        true
performance.cache-ima-xattrs       true
performance.md-cache-statfs        off
performance.xattr-cache-list   
performance.nl-cache-pass-through          false
features.encryption        off
network.frame-timeout   1800
network.ping-timeout    30
network.tcp-window-size    (null)
client.ssl         off
network.remote-dio         off
client.event-threads    4
client.tcp-user-timeout 0
client.keepalive-time   20
client.keepalive-interval       2
client.keepalive-count  9
network.tcp-window-size    (null)
network.inode-lru-limit 16384
auth.allow         *
auth.reject        (null)
transport.keepalive     1
server.allow-insecure      on
server.root-squash         off
server.all-squash          off
server.anonuid  65534
server.anongid  65534
server.statedump-path      /var/run/gluster
server.outstanding-rpc-limit    64
server.ssl         off
auth.ssl-allow     *
server.manage-gids         off
server.dynamic-auth        on
client.send-gids           on
server.gid-timeout      300
server.own-thread          (null)
server.event-threads    4
server.tcp-user-timeout 42
server.keepalive-time   20
server.keepalive-interval       2
server.keepalive-count  9
transport.listen-backlog        1024
transport.address-family           inet
performance.write-behind           on
performance.read-ahead     off
performance.readdir-ahead          on
performance.io-cache       off
performance.open-behind    on
performance.quick-read     off
performance.nl-cache       off
performance.stat-prefetch          on
performance.client-io-threads      on
performance.nfs.write-behind       on
performance.nfs.read-ahead         off
performance.nfs.io-cache           off
performance.nfs.quick-read         off
performance.nfs.stat-prefetch      off
performance.nfs.io-threads         off
performance.force-readdirp         true
performance.cache-invalidation     false
performance.global-cache-invalidation      true
features.uss       off
features.snapshot-directory        .snaps
features.show-snapshot-directory           off
features.tag-namespaces    off
network.compression        off
network.compression.window-size -15
network.compression.mem-level   8
network.compression.min-size    0
network.compression.compression-level   -1
network.compression.debug          false
features.default-soft-limit        80%
features.soft-timeout   60
features.hard-timeout   5
features.alert-time     86400
features.quota-deem-statfs         off
geo-replication.indexing           off
geo-replication.indexing           off
geo-replication.ignore-pid-check           off
geo-replication.ignore-pid-check           off
features.quota     off
features.inode-quota       off
features.bitrot    disable
debug.trace        off
debug.log-history          no
debug.log-file     no
debug.exclude-ops          (null)
debug.include-ops          (null)
debug.error-gen    off
debug.error-failure        (null)
debug.error-number         (null)
debug.random-failure       off
debug.error-fops           (null)
nfs.disable        on
features.read-only         off
features.worm      off
features.worm-file-level           off
features.worm-files-deletable      on
features.default-retention-period       120
features.retention-mode    relax
features.auto-commit-period     180
storage.linux-aio          off
storage.batch-fsync-mode           reverse-fsync
storage.batch-fsync-delay-usec  0
storage.owner-uid       36
storage.owner-gid       36
storage.node-uuid-pathinfo         off
storage.health-check-interval   30
storage.build-pgfid        off
storage.gfid2path          on
storage.gfid2path-separator        :
storage.reserve 1
storage.health-check-timeout    10
storage.fips-mode-rchecksum        off
storage.force-create-mode       0
storage.force-directory-mode    0
storage.create-mask     777
storage.create-directory-mask   777
storage.max-hardlinks   100
features.ctime     on
config.gfproxyd    off
cluster.server-quorum-type         server
cluster.server-quorum-ratio     0
changelog.changelog        off
changelog.changelog-dir    {{ brick.path }}/.glusterfs/changelogs
changelog.encoding         ascii
changelog.rollover-time 15
changelog.fsync-interval        5
changelog.changelog-barrier-timeout     120
changelog.capture-del-path         off
features.barrier           disable
features.barrier-timeout        120
features.trash     off
features.trash-dir         .trashcan
features.trash-eliminate-path      (null)
features.trash-max-filesize        5MB
features.trash-internal-op         off
cluster.enable-shared-storage      disable
locks.trace        off
locks.mandatory-locking    off
cluster.disperse-self-heal-daemon          enable
cluster.quorum-reads       no
client.bind-insecure       (null)
features.shard     on
features.shard-block-size          64MB
features.shard-lru-limit        16384
features.shard-deletion-rate    100
features.scrub-throttle    lazy
features.scrub-freq        biweekly
features.scrub     false
features.expiry-time    120
features.cache-invalidation        off
features.cache-invalidation-timeout     60
features.leases    off
features.lease-lock-recall-timeout      60
disperse.background-heals       8
disperse.heal-wait-qlength      128
cluster.heal-timeout    600
dht.force-readdirp         on
disperse.read-policy       gfid-hash
cluster.shd-max-threads 8
cluster.shd-wait-qlength        10000
cluster.locking-scheme     granular
cluster.granular-entry-heal        enable
features.locks-revocation-secs  0
features.locks-revocation-clear-all        false
features.locks-revocation-max-blocked   0
features.locks-monkey-unlocking    false
features.locks-notify-contention           no
features.locks-notify-contention-dela   y  5
disperse.shd-max-threads        1
disperse.shd-wait-qlength       1024
disperse.cpu-extensions    auto
disperse.self-heal-window-size  1
cluster.use-compound-fops          off
performance.parallel-readdir       off
performance.rda-request-size    131072
performance.rda-low-wmark       4096
performance.rda-high-wmark         128KB
performance.rda-cache-limit        10MB
performance.nl-cache-positive-entry        false
performance.nl-cache-limit         10MB
performance.nl-cache-timeout    60
cluster.brick-multiplex    off
cluster.max-bricks-per-process  250
disperse.optimistic-change-log     on
disperse.stripe-cache   4
cluster.halo-enabled       False
cluster.halo-shd-max-latency    99999
cluster.halo-nfsd-max-latency   5
cluster.halo-max-latency        5
cluster.halo-max-replicas       99999
cluster.halo-min-replicas       2
features.selinux           on
cluster.daemon-log-level           INFO
debug.delay-gen    off
delay-gen.delay-percentage         10%
delay-gen.delay-duration        100000
delay-gen.enable       
disperse.parallel-writes           on
features.sdfs      off
features.cloudsync         off
features.ctime     on
ctime.noatime      on
feature.cloudsync-storetype        (null)
features.enforce-mandatory-lock    off


thank you!
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/EU4TYVYEDB4GRQJD5GXSBBCUR3ZHPYTS/