It is my impression also that there are several moving parts that are not
thoroughly tested before every release cycle and forces one to be hesitant
to use in production. Overall is a great product but seems that the actual
testing is left to the end users with all its implicit difficulties and
pain. I understand the bargain though.
Alex
On Wed, May 30, 2018, 22:28 Jayme <jaymef(a)gmail.com> wrote:
as someone with hardware on the way to build a 3 node HCI cluster
with
only 2 2tb SSDs per node this problem concerns me, I sometimes worry about
the complexity glusterfs adds to the mix and how how difficult it can be to
manage when things go wrong.
Ive been following this thread and feel for you. You must be getting
tired, been there done that and don't wish it upon anyone. I hope you get
your environment back in working order soon and at some point discover what
went wrong to cause this problem.
On Wed, May 30, 2018, 8:55 AM Jim Kusznir, <jim(a)palousetech.com> wrote:
> So, I appear to be at the point where gluster appears to be functional,
> but engine still won't start. Here's from agent.log on one machine after
> giving the hosted-engine --vm-start command:
>
> MainThread::INFO::2018-05-29
>
23:52:09,953::hosted_engine::948::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_start_engine_vm)
> Engine VM started on localhost
> MainThread::INFO::2018-05-29
>
23:52:10,230::brokerlink::68::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
> Success, was notification of state_transition (EngineStart-EngineStarting)
> sent? sent
> MainThread::INFO::2018-05-29
>
23:56:16,061::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_monitoring_loop)
> Current state EngineStarting (score: 3400)
> MainThread::INFO::2018-05-29
>
23:56:16,075::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(consume)
> VM is unexpectedly down.
> MainThread::INFO::2018-05-29
>
23:56:16,076::state_decorators::92::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
> Timeout cleared while transitioning <class
> 'ovirt_hosted_engine_ha.agent.states.EngineStarting'> -> <class
> 'ovirt_hosted_engine_ha.agent.states.EngineMaybeAway'>
> MainThread::INFO::2018-05-29
>
23:56:16,359::brokerlink::68::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
> Success, was notification of state_transition
> (EngineStarting-EngineMaybeAway) sent? sent
> MainThread::INFO::2018-05-29
>
23:56:16,554::ovf_store::151::ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(getEngineVMOVF)
> OVF_STORE volume path:
>
/var/run/vdsm/storage/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/f22829ab-9fd5-415a-9a8f-809d3f7887d4/9f4760ee-119c-412a-a1e8-49e73e6ba929
> ---------------
> here's an excerpt from vdsm.log:
> 2018-05-30 00:00:23,114-0700 INFO (jsonrpc/2) [jsonrpc.JsonRpcServer]
> RPC call VM.create succeeded in 0.02 seconds (__init__:573)
> 2018-05-30 00:00:23,115-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') VM wrapper has started
> (vm:2764)
> 2018-05-30 00:00:23,125-0700 INFO (vm/50392390) [vdsm.api] START
> getVolumeSize(sdUUID='c0acdefb-7d16-48ec-9d76-659b8fe33e2a',
> spUUID='00000000-0000-0000-0000-000000000000',
> imgUUID='3de00d8c-d8b0-4ae0-9363-38a504f5d2b2',
> volUUID='d0fcd8de-6105-4f33-a674-727e3a11e89f', options=None)
> from=internal, task_id=c6d74ce9-1944-457a-9deb-d0a6929f39ce (api:46)
> 2018-05-30 00:00:23,130-0700 INFO (vm/50392390) [vdsm.api] FINISH
> getVolumeSize return={'truesize': '11298571776',
'apparentsize':
> '10737418240'} from=internal, task_id=c6d74ce9-1944-457a-9deb-d0a6929f39ce
> (api:52)
> 2018-05-30 00:00:23,131-0700 INFO (vm/50392390) [vds] prepared volume
> path: (clientIF:497)
> 2018-05-30 00:00:23,131-0700 INFO (vm/50392390) [vdsm.api] START
> prepareImage(sdUUID='c0acdefb-7d16-48ec-9d76-659b8fe33e2a',
> spUUID='00000000-0000-0000-0000-000000000000',
> imgUUID='3de00d8c-d8b0-4ae0-9363-38a504f5d2b2',
> leafUUID='d0fcd8de-6105-4f33-a674-727e3a11e89f', allowIllegal=False)
> from=internal, task_id=d2797364-270f-4800-92b0-d114441194a3 (api:46)
> 2018-05-30 00:00:23,155-0700 INFO (vm/50392390) [storage.StorageDomain]
> Fixing permissions on
/rhev/data-center/mnt/glusterSD/192.168.8.11:_engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f
> (fileSD:622)
> 2018-05-30 00:00:23,156-0700 INFO (vm/50392390) [storage.StorageDomain]
> Creating domain run directory
> u'/var/run/vdsm/storage/c0acdefb-7d16-48ec-9d76-659b8fe33e2a' (fileSD:576)
> 2018-05-30 00:00:23,156-0700 INFO (vm/50392390) [storage.fileUtils]
> Creating directory:
> /var/run/vdsm/storage/c0acdefb-7d16-48ec-9d76-659b8fe33e2a mode: None
> (fileUtils:197)
> 2018-05-30 00:00:23,157-0700 INFO (vm/50392390) [storage.StorageDomain]
> Creating symlink from
/rhev/data-center/mnt/glusterSD/192.168.8.11:_engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2
> to
>
/var/run/vdsm/storage/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2
> (fileSD:579)
> 2018-05-30 00:00:23,228-0700 INFO (jsonrpc/0) [root]
> /usr/libexec/vdsm/hooks/after_get_caps/openstacknet_utils.py: rc=0 err=
> (hooks:110)
> 2018-05-30 00:00:23,418-0700 INFO (vm/50392390) [vdsm.api] FINISH
> prepareImage return={'info': {'path':
>
u'engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f',
> 'type': 'network', 'hosts': [{'port': '0',
'transport': 'tcp', 'name':
> '192.168.8.11'}, {'port': '0', 'transport':
'tcp', 'name': '
> ovirt1.nwfiber.com'}, {'port': '0', 'transport':
'tcp', 'name': '
> ovirt2.nwfiber.com'}, {'port': '0', 'transport':
'tcp', 'name': '
> ovirt3.nwfiber.com'}], 'protocol': 'gluster'}, 'path':
>
u'/var/run/vdsm/storage/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f',
> 'imgVolumesInfo': [{'domainID':
'c0acdefb-7d16-48ec-9d76-659b8fe33e2a',
> 'leaseOffset': 0, 'path':
u'/rhev/data-center/mnt/glusterSD/192.168.8.11:_engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f',
> 'volumeID': u'd0fcd8de-6105-4f33-a674-727e3a11e89f',
'leasePath':
>
u'/rhev/data-center/mnt/glusterSD/192.168.8.11:_engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f.lease',
> 'imageID': '3de00d8c-d8b0-4ae0-9363-38a504f5d2b2'}]} from=internal,
> task_id=d2797364-270f-4800-92b0-d114441194a3 (api:52)
> 2018-05-30 00:00:23,419-0700 INFO (vm/50392390) [vds] prepared volume
> path:
>
/var/run/vdsm/storage/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f
> (clientIF:497)
> 2018-05-30 00:00:23,420-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') Enabling drive monitoring
> (drivemonitor:54)
> 2018-05-30 00:00:23,434-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') drive 'hdc' path:
'file=' ->
> '*file=' (storagexml:323)
> 2018-05-30 00:00:23,435-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') drive 'vda' path:
>
'file=/rhev/data-center/00000000-0000-0000-0000-000000000000/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f'
> ->
>
u'*file=/var/run/vdsm/storage/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f'
> (storagexml:323)
> 2018-05-30 00:00:23,435-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') Using drive lease:
> {'domainID': 'c0acdefb-7d16-48ec-9d76-659b8fe33e2a',
'leaseOffset': 0,
> 'path':
u'/rhev/data-center/mnt/glusterSD/192.168.8.11:_engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f',
> 'volumeID': u'd0fcd8de-6105-4f33-a674-727e3a11e89f',
'leasePath':
>
u'/rhev/data-center/mnt/glusterSD/192.168.8.11:_engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f.lease',
> 'imageID': '3de00d8c-d8b0-4ae0-9363-38a504f5d2b2'} (lease:327)
> 2018-05-30 00:00:23,436-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') lease
> c0acdefb-7d16-48ec-9d76-659b8fe33e2a:d0fcd8de-6105-4f33-a674-727e3a11e89f
> path:
>
'LEASE-PATH:d0fcd8de-6105-4f33-a674-727e3a11e89f:c0acdefb-7d16-48ec-9d76-659b8fe33e2a'
> ->
u'/rhev/data-center/mnt/glusterSD/192.168.8.11:_engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f.lease'
> (lease:316)
> 2018-05-30 00:00:23,436-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') lease
> c0acdefb-7d16-48ec-9d76-659b8fe33e2a:d0fcd8de-6105-4f33-a674-727e3a11e89f
> offset:
>
'LEASE-OFFSET:d0fcd8de-6105-4f33-a674-727e3a11e89f:c0acdefb-7d16-48ec-9d76-659b8fe33e2a'
> -> 0 (lease:316)
> 2018-05-30 00:00:23,580-0700 INFO (jsonrpc/0) [root]
> /usr/libexec/vdsm/hooks/after_get_caps/ovirt_provider_ovn_hook: rc=0
> err=Failed to get Open VSwitch system-id . err = ['ovs-vsctl:
> unix:/var/run/openvswitch/db.sock: database connection failed (No such file
> or directory)']
> (hooks:110)
> 2018-05-30 00:00:23,582-0700 INFO (jsonrpc/0) [api.host] FINISH
> getCapabilities return={'status': {'message': 'Done',
'code': 0}, 'info':
> {u'HBAInventory': {u'iSCSI': [{u'InitiatorName':
> u'iqn.1994-05.com.redhat:1283ce7a3171'}], u'FC': []},
u'packages2':
> {u'kernel': {u'release': u'862.3.2.el7.x86_64',
u'version': u'3.10.0'},
> u'glusterfs-rdma': {u'release': u'1.el7', u'version':
u'3.12.9'},
> u'glusterfs-fuse': {u'release': u'1.el7', u'version':
u'3.12.9'},
> u'spice-server': {u'release': u'2.el7_5.3',
u'version': u'0.14.0'},
> u'librbd1': {u'release': u'2.el7', u'version':
u'0.94.5'}, u'vdsm':
> {u'release': u'1.el7.centos', u'version':
u'4.20.27.1'}, u'qemu-kvm':
> {u'release': u'21.el7_5.3.1', u'version': u'2.10.0'},
u'openvswitch':
> {u'release': u'3.el7', u'version': u'2.9.0'},
u'libvirt': {u'release':
> u'14.el7_5.5', u'version': u'3.9.0'},
u'ovirt-hosted-engine-ha':
> {u'release': u'1.el7.centos', u'version': u'2.2.11'},
u'qemu-img':
> {u'release': u'21.el7_5.3.1', u'version': u'2.10.0'},
u'mom': {u'release':
> u'1.el7.centos', u'version': u'0.5.12'},
u'glusterfs': {u'release':
> u'1.el7', u'version': u'3.12.9'}, u'glusterfs-cli':
{u'release': u'1.el7',
> u'version': u'3.12.9'}, u'glusterfs-server':
{u'release': u'1.el7',
> u'version': u'3.12.9'}, u'glusterfs-geo-replication':
{u'release':
> u'1.el7', u'version': u'3.12.9'}},
u'numaNodeDistance': {u'1': [20, 10],
> u'0': [10, 20]}, u'cpuModel': u'Intel(R) Xeon(R) CPU
E5540 @
> 2.53GHz', u'nestedVirtualization': False, u'liveMerge':
u'true', u'hooks':
> {u'before_vm_start': {u'50_hostedengine': {u'md5':
> u'95c810cdcfe4195302a59574a5148289'}, u'50_vhostmd':
{u'md5':
> u'9206bc390bcbf208b06a8e899581be2d'}, u'50_vfio_mdev':
{u'md5':
> u'b77eaad7b8f5ca3825dca1b38b0dd446'}}, u'after_network_setup':
> {u'30_ethtool_options': {u'md5':
u'f04c2ca5dce40663e2ed69806eea917c'}},
> u'after_vm_destroy': {u'50_vhostmd': {u'md5':
> u'd70f7ee0453f632e87c09a157ff8ff66'}, u'50_vfio_mdev':
{u'md5':
> u'35364c6e08ecc58410e31af6c73b17f6'}}, u'after_vm_start':
> {u'openstacknet_utils.py': {u'md5':
u'640ef8b473cbc16b102c5ebf57ea533f'},
> u'50_openstacknet': {u'md5':
u'ea0a5a715da8c1badbcda28e8b8fa00e'}},
> u'after_device_migrate_destination': {u'openstacknet_utils.py':
{u'md5':
> u'640ef8b473cbc16b102c5ebf57ea533f'}, u'50_openstacknet':
{u'md5':
> u'f93c12af7454bbba4c1ef445c2bc9860'}}, u'before_network_setup':
> {u'50_fcoe': {u'md5': u'28c352339c8beef1e1b05c67d106d062'}},
> u'after_get_caps': {u'openstacknet_utils.py': {u'md5':
> u'640ef8b473cbc16b102c5ebf57ea533f'}, u'50_openstacknet':
{u'md5':
> u'0e00c63ab44a952e722209ea31fd7a71'}, u'ovirt_provider_ovn_hook':
{u'md5':
> u'257990644ee6c7824b522e2ab3077ae7'}}, u'after_nic_hotplug':
> {u'openstacknet_utils.py': {u'md5':
u'640ef8b473cbc16b102c5ebf57ea533f'},
> u'50_openstacknet': {u'md5':
u'f93c12af7454bbba4c1ef445c2bc9860'}},
> u'before_vm_migrate_destination': {u'50_vhostmd': {u'md5':
> u'9206bc390bcbf208b06a8e899581be2d'}}, u'after_device_create':
> {u'openstacknet_utils.py': {u'md5':
u'640ef8b473cbc16b102c5ebf57ea533f'},
> u'50_openstacknet': {u'md5':
u'f93c12af7454bbba4c1ef445c2bc9860'}},
> u'before_vm_dehibernate': {u'50_vhostmd': {u'md5':
> u'9206bc390bcbf208b06a8e899581be2d'}}, u'before_nic_hotplug':
{u'50_vmfex':
> {u'md5': u'49caba1a5faadd8efacef966f79bc30a'},
u'openstacknet_utils.py':
> {u'md5': u'640ef8b473cbc16b102c5ebf57ea533f'},
u'50_openstacknet': {u'md5':
> u'0438de5ff9b6bf8d3160804ff71bf827'}, u'ovirt_provider_ovn_hook':
{u'md5':
> u'ada7e4e757fd0241b682b5cc6a545fc8'}},
> u'before_device_migrate_destination': {u'50_vmfex': {u'md5':
> u'49caba1a5faadd8efacef966f79bc30a'}}, u'before_device_create':
> {u'50_vmfex': {u'md5': u'49caba1a5faadd8efacef966f79bc30a'},
> u'openstacknet_utils.py': {u'md5':
u'640ef8b473cbc16b102c5ebf57ea533f'},
> u'50_openstacknet': {u'md5':
u'0438de5ff9b6bf8d3160804ff71bf827'},
> u'ovirt_provider_ovn_hook': {u'md5':
> u'ada7e4e757fd0241b682b5cc6a545fc8'}}}, u'supportsIPv6': True,
> u'realtimeKernel': False, u'vmTypes': [u'kvm'],
u'liveSnapshot': u'true',
> u'cpuThreads': u'16', u'kdumpStatus': 0, u'networks':
{u'ovirtmgmt':
> {u'iface': u'ovirtmgmt', u'ipv6autoconf': False,
u'addr': u'192.168.8.12',
> u'dhcpv6': False, u'ipv6addrs': [], u'switch':
u'legacy', u'bridged': True,
> u'mtu': u'1500', u'dhcpv4': True, u'netmask':
u'255.255.255.0',
> u'ipv4defaultroute': True, u'stp': u'off',
u'ipv4addrs': [u'
> 192.168.8.12/24'], u'ipv6gateway': u'::', u'gateway':
u'192.168.8.1',
> u'ports': [u'em1']}, u'Public_Fiber': {u'iface':
u'Public_Fiber',
> u'ipv6autoconf': False, u'addr': u'', u'dhcpv6':
False, u'ipv6addrs': [],
> u'switch': u'legacy', u'bridged': True, u'mtu':
u'1500', u'dhcpv4': False,
> u'netmask': u'', u'ipv4defaultroute': False, u'stp':
u'off', u'ipv4addrs':
> [], u'ipv6gateway': u'::', u'gateway': u'',
u'ports': [u'em1.3']},
> u'Gluster': {u'iface': u'em4', u'ipv6autoconf':
False, u'addr':
> u'172.172.1.12', u'dhcpv6': False, u'ipv6addrs': [],
u'switch': u'legacy',
> u'bridged': False, u'mtu': u'1500', u'dhcpv4': False,
u'netmask':
> u'255.255.255.0', u'ipv4defaultroute': False, u'ipv4addrs':
[u'
> 172.172.1.12/24'], u'interface': u'em4', u'ipv6gateway':
u'::',
> u'gateway': u''}}, u'kernelArgs':
> u'BOOT_IMAGE=/vmlinuz-3.10.0-862.3.2.el7.x86_64
> root=/dev/mapper/centos_ovirt-root ro crashkernel=auto rd.lvm.lv=centos_ovirt/root
> rd.lvm.lv=centos_ovirt/swap rhgb quiet LANG=en_US.UTF-8', u'bridges':
> {u'Public_Fiber': {u'ipv6autoconf': False, u'addr':
u'', u'dhcpv6': False,
> u'ipv6addrs': [], u'mtu': u'1500', u'dhcpv4': False,
u'netmask': u'',
> u'ipv4defaultroute': False, u'stp': u'off',
u'ipv4addrs': [],
> u'ipv6gateway': u'::', u'gateway': u'',
u'opts':
> {u'multicast_last_member_count': u'2', u'vlan_protocol':
u'0x8100',
> u'hash_elasticity': u'4',
u'multicast_query_response_interval': u'1000',
> u'group_fwd_mask': u'0x0', u'multicast_snooping':
u'1',
> u'multicast_startup_query_interval': u'3125', u'hello_timer':
u'0',
> u'multicast_querier_interval': u'25500', u'max_age':
u'2000', u'hash_max':
> u'512', u'stp_state': u'0',
u'topology_change_detected': u'0', u'priority':
> u'32768', u'multicast_igmp_version': u'2',
> u'multicast_membership_interval': u'26000',
u'root_path_cost': u'0',
> u'root_port': u'0', u'multicast_stats_enabled':
u'0',
> u'multicast_startup_query_count': u'2', u'nf_call_iptables':
u'0',
> u'vlan_stats_enabled': u'0', u'hello_time': u'200',
u'topology_change':
> u'0', u'bridge_id': u'8000.00219b900710',
u'topology_change_timer': u'0',
> u'ageing_time': u'30000', u'nf_call_ip6tables':
u'0',
> u'multicast_mld_version': u'1', u'gc_timer':
u'25509', u'root_id':
> u'8000.00219b900710', u'nf_call_arptables': u'0',
u'group_addr':
> u'1:80:c2:0:0:0', u'multicast_last_member_interval': u'100',
> u'default_pvid': u'1', u'multicast_query_interval':
u'12500',
> u'multicast_query_use_ifaddr': u'0', u'tcn_timer':
u'0',
> u'multicast_router': u'1', u'vlan_filtering': u'0',
u'multicast_querier':
> u'0', u'forward_delay': u'0'}, u'ports':
[u'em1.3']}, u'ovirtmgmt':
> {u'ipv6autoconf': False, u'addr': u'192.168.8.12',
u'dhcpv6': False,
> u'ipv6addrs': [], u'mtu': u'1500', u'dhcpv4': True,
u'netmask':
> u'255.255.255.0', u'ipv4defaultroute': True, u'stp':
u'off', u'ipv4addrs':
> [u'192.168.8.12/24'], u'ipv6gateway': u'::',
u'gateway': u'192.168.8.1',
> u'opts': {u'multicast_last_member_count': u'2',
u'vlan_protocol':
> u'0x8100', u'hash_elasticity': u'4',
u'multicast_query_response_interval':
> u'1000', u'group_fwd_mask': u'0x0',
u'multicast_snooping': u'1',
> u'multicast_startup_query_interval': u'3125', u'hello_timer':
u'0',
> u'multicast_querier_interval': u'25500', u'max_age':
u'2000', u'hash_max':
> u'512', u'stp_state': u'0',
u'topology_change_detected': u'0', u'priority':
> u'32768', u'multicast_igmp_version': u'2',
> u'multicast_membership_interval': u'26000',
u'root_path_cost': u'0',
> u'root_port': u'0', u'multicast_stats_enabled':
u'0',
> u'multicast_startup_query_count': u'2', u'nf_call_iptables':
u'0',
> u'vlan_stats_enabled': u'0', u'hello_time': u'200',
u'topology_change':
> u'0', u'bridge_id': u'8000.00219b900710',
u'topology_change_timer': u'0',
> u'ageing_time': u'30000', u'nf_call_ip6tables':
u'0',
> u'multicast_mld_version': u'1', u'gc_timer':
u'13221', u'root_id':
> u'8000.00219b900710', u'nf_call_arptables': u'0',
u'group_addr':
> u'1:80:c2:0:0:0', u'multicast_last_member_interval': u'100',
> u'default_pvid': u'1', u'multicast_query_interval':
u'12500',
> u'multicast_query_use_ifaddr': u'0', u'tcn_timer':
u'0',
> u'multicast_router': u'1', u'vlan_filtering': u'0',
u'multicast_querier':
> u'0', u'forward_delay': u'0'}, u'ports':
[u'em1']}}, u'uuid':
> u'4C4C4544-005A-4710-8034-B2C04F4C4B31', u'onlineCpus':
> u'0,2,4,6,8,10,12,14,1,3,5,7,9,11,13,15', u'nameservers':
[u'192.168.8.1',
> u'8.8.8.8'], u'nics': {u'em4': {u'ipv6autoconf':
False, u'addr':
> u'172.172.1.12', u'speed': 1000, u'dhcpv6': False,
u'ipv6addrs': [],
> u'mtu': u'1500', u'dhcpv4': False, u'netmask':
u'255.255.255.0',
> u'ipv4defaultroute': False, u'ipv4addrs':
[u'172.172.1.12/24'],
> u'hwaddr': u'00:21:9b:90:07:16', u'ipv6gateway':
u'::', u'gateway': u''},
> u'em1': {u'ipv6autoconf': False, u'addr': u'',
u'speed': 1000, u'dhcpv6':
> False, u'ipv6addrs': [], u'mtu': u'1500', u'dhcpv4':
False, u'netmask':
> u'', u'ipv4defaultroute': False, u'ipv4addrs': [],
u'hwaddr':
> u'00:21:9b:90:07:10', u'ipv6gateway': u'::',
u'gateway': u''}, u'em3':
> {u'ipv6autoconf': True, u'addr': u'', u'speed': 0,
u'dhcpv6': False,
> u'ipv6addrs': [], u'mtu': u'1500', u'dhcpv4': False,
u'netmask': u'',
> u'ipv4defaultroute': False, u'ipv4addrs': [], u'hwaddr':
> u'00:21:9b:90:07:14', u'ipv6gateway': u'::',
u'gateway': u''}, u'em2':
> {u'ipv6autoconf': True, u'addr': u'', u'speed': 0,
u'dhcpv6': False,
> u'ipv6addrs': [], u'mtu': u'1500', u'dhcpv4': False,
u'netmask': u'',
> u'ipv4defaultroute': False, u'ipv4addrs': [], u'hwaddr':
> u'00:21:9b:90:07:12', u'ipv6gateway': u'::',
u'gateway': u''}},
> u'software_revision': u'1', u'hostdevPassthrough':
u'false',
> u'clusterLevels': [u'3.6', u'4.0', u'4.1',
u'4.2'], u'cpuFlags':
>
u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,nopl,xtopology,nonstop_tsc,aperfmperf,pni,dtes64,monitor,ds_cpl,vmx,est,tm2,ssse3,cx16,xtpr,pdcm,dca,sse4_1,sse4_2,popcnt,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,dtherm,ida,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_n270',
> u'kernelFeatures': {u'RETP': 1, u'IBRS': 0, u'PTI':
1},
> u'ISCSIInitiatorName': u'iqn.1994-05.com.redhat:1283ce7a3171',
> u'netConfigDirty': u'False', u'selinux': {u'mode':
u'1'},
> u'autoNumaBalancing': 1, u'reservedMem': u'321',
u'containers': False,
> u'bondings': {}, u'software_version': u'4.20',
u'supportedENGINEs':
> [u'3.6', u'4.0', u'4.1', u'4.2'],
u'cpuSpeed': u'2527.017', u'numaNodes':
> {u'1': {u'totalMemory': u'16371', u'cpus': [1, 3, 5,
7, 9, 11, 13, 15]},
> u'0': {u'totalMemory': u'16384', u'cpus': [0, 2, 4,
6, 8, 10, 12, 14]}},
> u'cpuSockets': u'2', u'vlans': {u'em1.3':
{u'iface': u'em1',
> u'ipv6autoconf': False, u'addr': u'', u'dhcpv6':
False, u'ipv6addrs': [],
> u'vlanid': 3, u'mtu': u'1500', u'dhcpv4': False,
u'netmask': u'',
> u'ipv4defaultroute': False, u'ipv4addrs': [], u'ipv6gateway':
u'::',
> u'gateway': u''}}, u'version_name': u'Snow Man',
'lastClientIface': 'lo',
> u'cpuCores': u'8', u'hostedEngineDeployed': True,
u'hugepages': [2048],
> u'guestOverhead': u'65', u'additionalFeatures':
[u'libgfapi_supported',
> u'GLUSTER_SNAPSHOT', u'GLUSTER_GEO_REPLICATION',
> u'GLUSTER_BRICK_MANAGEMENT'], u'kvmEnabled': u'true',
u'memSize': u'31996',
> u'emulatedMachines': [u'pc-i440fx-rhel7.1.0',
u'pc-q35-rhel7.3.0',
> u'rhel6.3.0', u'pc-i440fx-rhel7.5.0', u'pc-i440fx-rhel7.0.0',
u'rhel6.1.0',
> u'pc-i440fx-rhel7.4.0', u'rhel6.6.0', u'pc-q35-rhel7.5.0',
u'rhel6.2.0',
> u'pc', u'pc-i440fx-rhel7.3.0', u'q35',
u'pc-i440fx-rhel7.2.0',
> u'rhel6.4.0', u'pc-q35-rhel7.4.0', u'rhel6.0.0',
u'rhel6.5.0'],
> u'rngSources': [u'hwrng', u'random'],
u'operatingSystem': {u'release':
> u'5.1804.el7.centos.2', u'pretty_name': u'CentOS Linux 7
(Core)',
> u'version': u'7', u'name': u'RHEL'}}} from=::1,35660
(api:52)
> 2018-05-30 00:00:23,593-0700 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer]
> RPC call Host.getCapabilities succeeded in 3.26 seconds (__init__:573)
> 2018-05-30 00:00:23,724-0700 INFO (vm/50392390) [root]
> /usr/libexec/vdsm/hooks/before_device_create/50_openstacknet: rc=0 err=
> (hooks:110)
> 2018-05-30 00:00:23,919-0700 INFO (vm/50392390) [root]
> /usr/libexec/vdsm/hooks/before_device_create/50_vmfex: rc=0 err= (hooks:110)
> 2018-05-30 00:00:24,213-0700 INFO (vm/50392390) [root]
> /usr/libexec/vdsm/hooks/before_device_create/openstacknet_utils.py: rc=0
> err= (hooks:110)
> 2018-05-30 00:00:24,392-0700 INFO (vm/50392390) [root]
> /usr/libexec/vdsm/hooks/before_device_create/ovirt_provider_ovn_hook: rc=0
> err= (hooks:110)
> 2018-05-30 00:00:24,401-0700 INFO (jsonrpc/1) [api.host] START
> getAllVmStats() from=::1,33766 (api:46)
> 2018-05-30 00:00:24,403-0700 INFO (jsonrpc/1) [api.host] FINISH
> getAllVmStats return={'status': {'message': 'Done',
'code': 0},
> 'statsList': (suppressed)} from=::1,33766 (api:52)
> 2018-05-30 00:00:24,403-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer]
> RPC call Host.getAllVmStats succeeded in 0.01 seconds (__init__:573)
> 2018-05-30 00:00:24,623-0700 INFO (vm/50392390) [root]
> /usr/libexec/vdsm/hooks/before_vm_start/50_hostedengine: rc=0 err=
> (hooks:110)
> 2018-05-30 00:00:24,836-0700 INFO (vm/50392390) [root]
> /usr/libexec/vdsm/hooks/before_vm_start/50_vfio_mdev: rc=0 err= (hooks:110)
> 2018-05-30 00:00:25,022-0700 INFO (vm/50392390) [root]
> /usr/libexec/vdsm/hooks/before_vm_start/50_vhostmd: rc=0 err= (hooks:110)
> 2018-05-30 00:00:25,023-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') <?xml
version="1.0"
> encoding="utf-8"?><domain type="kvm" xmlns:ns0="
>
http://ovirt.org/vm/tune/1.0"
xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
> <name>HostedEngine</name>
> <uuid>50392390-4f89-435c-bf3b-8254b58a4ef7</uuid>
> <memory>8388608</memory>
> <currentMemory>8388608</currentMemory>
> <maxMemory slots="16">33554432</maxMemory>
> <vcpu current="4">16</vcpu>
> <sysinfo type="smbios">
> <system>
> <entry name="manufacturer">oVirt</entry>
> <entry name="product">oVirt Node</entry>
> <entry
name="version">7-5.1804.el7.centos.2</entry>
> <entry
> name="serial">4C4C4544-005A-4710-8034-B2C04F4C4B31</entry>
> <entry
> name="uuid">50392390-4f89-435c-bf3b-8254b58a4ef7</entry>
> </system>
> </sysinfo>
> <clock adjustment="0" offset="variable">
> <timer name="rtc" tickpolicy="catchup"/>
> <timer name="pit" tickpolicy="delay"/>
> <timer name="hpet" present="no"/>
> </clock>
> <features>
> <acpi/>
> </features>
> <cpu match="exact">
> <model>Nehalem</model>
> <topology cores="1" sockets="16"
threads="1"/>
> <numa>
> <cell cpus="0,1,2,3" id="0"
memory="8388608"/>
> </numa>
> </cpu>
> <cputune/>
> <devices>
> <input bus="usb" type="tablet"/>
> <channel type="unix">
> <target name="ovirt-guest-agent.0"
type="virtio"/>
> <source mode="bind"
>
path="/var/lib/libvirt/qemu/channels/50392390-4f89-435c-bf3b-8254b58a4ef7.ovirt-guest-agent.0"/>
> </channel>
> <channel type="unix">
> <target name="org.qemu.guest_agent.0"
type="virtio"/>
> <source mode="bind"
>
path="/var/lib/libvirt/qemu/channels/50392390-4f89-435c-bf3b-8254b58a4ef7.org.qemu.guest_agent.0"/>
> </channel>
> <controller index="0" ports="16"
type="virtio-serial">
> <address bus="0x00" domain="0x0000"
function="0x0"
> slot="0x03" type="pci"/>
> </controller>
> <rng model="virtio">
> <backend model="random">/dev/urandom</backend>
> </rng>
> <graphics autoport="yes" keymap="en-us"
passwd="*****"
> passwdValidTo="1970-01-01T00:00:01" port="-1"
type="vnc">
> <listen network="vdsm-ovirtmgmt"
type="network"/>
> </graphics>
> <controller type="usb">
> <address bus="0x00" domain="0x0000"
function="0x2"
> slot="0x01" type="pci"/>
> </controller>
> <controller type="ide">
> <address bus="0x00" domain="0x0000"
function="0x1"
> slot="0x01" type="pci"/>
> </controller>
> <video>
> <model heads="1" type="cirrus"
vram="16384"/>
> <address bus="0x00" domain="0x0000"
function="0x0"
> slot="0x05" type="pci"/>
> </video>
> <memballoon model="none"/>
> <disk device="cdrom" snapshot="no"
type="file">
> <driver error_policy="report" name="qemu"
type="raw"/>
> <source file="" startupPolicy="optional"/>
> <target bus="ide" dev="hdc"/>
> <readonly/>
> <address bus="1" controller="0"
target="0" type="drive"
> unit="0"/>
> </disk>
> <disk device="disk" snapshot="no"
type="file">
> <target bus="virtio" dev="vda"/>
> <source
>
file="/var/run/vdsm/storage/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f"/>
> <driver cache="none" error_policy="stop"
io="threads"
> name="qemu" type="raw"/>
> <alias name="ua-3de00d8c-d8b0-4ae0-9363-38a504f5d2b2"/>
> <address bus="0x00" domain="0x0000"
function="0x0"
> slot="0x04" type="pci"/>
> <serial>3de00d8c-d8b0-4ae0-9363-38a504f5d2b2</serial>
> </disk>
> <lease>
> <key>d0fcd8de-6105-4f33-a674-727e3a11e89f</key>
> <lockspace>c0acdefb-7d16-48ec-9d76-659b8fe33e2a</lockspace>
> <target offset="0"
> path="/rhev/data-center/mnt/glusterSD/192.168.8.11:
>
_engine/c0acdefb-7d16-48ec-9d76-659b8fe33e2a/images/3de00d8c-d8b0-4ae0-9363-38a504f5d2b2/d0fcd8de-6105-4f33-a674-727e3a11e89f.lease"/>
> </lease>
> <interface type="bridge">
> <model type="virtio"/>
> <link state="up"/>
> <source bridge="ovirtmgmt"/>
> <alias name="ua-9f93c126-cbb3-4c5b-909e-41a53c2e31fb"/>
> <address bus="0x00" domain="0x0000"
function="0x0"
> slot="0x02" type="pci"/>
> <mac address="00:16:3e:2d:bd:b1"/>
> <filterref filter="vdsm-no-mac-spoofing"/>
> <bandwidth/>
> </interface>
> <channel type="unix"><target
name="org.ovirt.hosted-engine-setup.0"
> type="virtio"/><source mode="bind"
>
path="/var/lib/libvirt/qemu/channels/50392390-4f89-435c-bf3b-8254b58a4ef7.org.ovirt.hosted-engine-setup.0"/></channel></devices>
> <pm>
> <suspend-to-disk enabled="no"/>
> <suspend-to-mem enabled="no"/>
> </pm>
> <os>
> <type arch="x86_64"
machine="pc-i440fx-rhel7.3.0">hvm</type>
> <smbios mode="sysinfo"/>
> </os>
> <metadata>
> <ns0:qos/>
> <ovirt-vm:vm>
> <minGuaranteedMemoryMb
type="int">8192</minGuaranteedMemoryMb>
> <clusterVersion>4.2</clusterVersion>
> <ovirt-vm:custom/>
> <ovirt-vm:device mac_address="00:16:3e:2d:bd:b1">
> <ovirt-vm:custom/>
> </ovirt-vm:device>
> <ovirt-vm:device devtype="disk" name="vda">
>
> <ovirt-vm:poolID>00000000-0000-0000-0000-000000000000</ovirt-vm:poolID>
>
>
<ovirt-vm:volumeID>d0fcd8de-6105-4f33-a674-727e3a11e89f</ovirt-vm:volumeID>
> <ovirt-vm:shared>exclusive</ovirt-vm:shared>
>
>
<ovirt-vm:imageID>3de00d8c-d8b0-4ae0-9363-38a504f5d2b2</ovirt-vm:imageID>
>
>
<ovirt-vm:domainID>c0acdefb-7d16-48ec-9d76-659b8fe33e2a</ovirt-vm:domainID>
> </ovirt-vm:device>
> <launchPaused>false</launchPaused>
> <resumeBehavior>auto_resume</resumeBehavior>
> </ovirt-vm:vm>
> </metadata>
>
<on_poweroff>destroy</on_poweroff><on_reboot>destroy</on_reboot><on_crash>destroy</on_crash></domain>
> (vm:2867)
> 2018-05-30 00:00:25,252-0700 INFO (monitor/c0acdef) [storage.SANLock]
> Acquiring host id for domain c0acdefb-7d16-48ec-9d76-659b8fe33e2a (id=2,
> async=True) (clusterlock:284)
> 2018-05-30 00:00:26,808-0700 ERROR (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') The vm start process failed
> (vm:943)
> Traceback (most recent call last):
> File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 872, in
> _startUnderlyingVm
> self._run()
> File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 2872, in
> _run
> dom.createWithFlags(flags)
> File
> "/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py", line
> 130, in wrapper
> ret = f(*args, **kwargs)
> File "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line
> 92, in wrapper
> return func(inst, *args, **kwargs)
> File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1099, in
> createWithFlags
> if ret == -1: raise libvirtError ('virDomainCreateWithFlags()
> failed', dom=self)
> libvirtError: Failed to acquire lock: No space left on device
> 2018-05-30 00:00:26,809-0700 INFO (vm/50392390) [virt.vm]
> (vmId='50392390-4f89-435c-bf3b-8254b58a4ef7') Changed state to Down: Failed
> to acquire lock: No space left on device (code=1) (vm:1683)
>
> -------------
> The "failed to acquire lock: no space left on device" error has me
> conserned...Yet there is no device that is full on any servers (espciially
> the one I'm on):
>
> Filesystem Size Used Avail Use% Mounted on
> /dev/mapper/centos_ovirt-root 8.0G 6.6G 1.5G 82% /
> devtmpfs 16G 0 16G 0% /dev
> tmpfs 16G 4.0K 16G 1% /dev/shm
> tmpfs 16G 18M 16G 1% /run
> tmpfs 16G 0 16G 0% /sys/fs/cgroup
> /dev/mapper/gluster-data 177G 59G 119G 33% /gluster/brick2
> /dev/mapper/gluster-engine 25G 18G 7.9G 69% /gluster/brick1
> /dev/mapper/vg_thin-iso 25G 7.5G 18G 30% /gluster/brick4
> /dev/mapper/vg_thin-data--hdd 1.2T 308G 892G 26% /gluster/brick3
> /dev/sda1 497M 313M 185M 63% /boot
> 192.168.8.11:/engine 25G 20G 5.5G 79%
> /rhev/data-center/mnt/glusterSD/192.168.8.11:_engine
> tmpfs 3.2G 0 3.2G 0% /run/user/0
> 192.168.8.11:/data 136G 60G 77G 44%
> /rhev/data-center/mnt/glusterSD/192.168.8.11:_data
> 192.168.8.11:/iso 25G 7.5G 18G 30%
> /rhev/data-center/mnt/glusterSD/192.168.8.11:_iso
> 172.172.1.11:/data-hdd 1.2T 308G 892G 26%
> /rhev/data-center/mnt/glusterSD/172.172.1.11:_data-hdd
>
> --------
> --Jim
>
>
> On Tue, May 29, 2018 at 11:53 PM, Jim Kusznir <jim(a)palousetech.com>
> wrote:
>
>> I've just gotten back to where hsoted-engine --vm-status does
>> (eventually) return some output other than the ha-agent or storage is down
>> error. It does currently take up to two minutes, and the last run returned
>> stale data.
>>
>> Gluster-wise, gluster volume heal <volume> info shows the three volumes
>> with all bricks are all showing 0 entries (and return quickly). The 4th
>> volume missing one brick (but still has two replicas) does return entries,
>> and is currently taking a very long time to come back with them.
>>
>> Making progress toward getting it online again...Don't know why I get
>> stale data in hosted-engine --vm-status or how to overcome that...
>>
>> --Jim
>>
>>
>>
>> On Tue, May 29, 2018 at 11:38 PM, Jim Kusznir <jim(a)palousetech.com>
>> wrote:
>>
>>> I believe the gluster data store for the engine is up and working
>>> correctly. The rest are not mounted, as the engine hasn't managed to
start
>>> correctly yet. I did perform the copy-a-junk-file into the data store,
>>> then check to ensure its there on another host, then delete that and see
>>> that it disappeared on the first host; it passed that test. Here's the
>>> info and status. (I have NOT performed the steps that Krutika and
>>> Ravishankar suggested yet, as I don't have my data volumes working again
>>> yet.
>>>
>>> [root@ovirt2 images]# gluster volume info
>>>
>>> Volume Name: data
>>> Type: Replicate
>>> Volume ID: e670c488-ac16-4dd1-8bd3-e43b2e42cc59
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 1 x (2 + 1) = 3
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: ovirt1.nwfiber.com:/gluster/brick2/data
>>> Brick2: ovirt2.nwfiber.com:/gluster/brick2/data
>>> Brick3: ovirt3.nwfiber.com:/gluster/brick2/data (arbiter)
>>> Options Reconfigured:
>>> diagnostics.count-fop-hits: on
>>> diagnostics.latency-measurement: on
>>> changelog.changelog: on
>>> geo-replication.ignore-pid-check: on
>>> geo-replication.indexing: on
>>> server.allow-insecure: on
>>> performance.readdir-ahead: on
>>> performance.quick-read: off
>>> performance.read-ahead: off
>>> performance.io-cache: off
>>> performance.stat-prefetch: off
>>> cluster.eager-lock: enable
>>> network.remote-dio: enable
>>> cluster.quorum-type: auto
>>> cluster.server-quorum-type: server
>>> storage.owner-uid: 36
>>> storage.owner-gid: 36
>>> features.shard: on
>>> features.shard-block-size: 512MB
>>> performance.low-prio-threads: 32
>>> cluster.data-self-heal-algorithm: full
>>> cluster.locking-scheme: granular
>>> cluster.shd-wait-qlength: 10000
>>> cluster.shd-max-threads: 8
>>> network.ping-timeout: 30
>>> user.cifs: off
>>> nfs.disable: on
>>> performance.strict-o-direct: on
>>>
>>> Volume Name: data-hdd
>>> Type: Replicate
>>> Volume ID: d342a3ab-16f3-49f0-bbcf-f788be8ac5f1
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 1 x 3 = 3
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: 172.172.1.11:/gluster/brick3/data-hdd
>>> Brick2: 172.172.1.12:/gluster/brick3/data-hdd
>>> Brick3: 172.172.1.13:/gluster/brick3/data-hdd
>>> Options Reconfigured:
>>> diagnostics.count-fop-hits: on
>>> diagnostics.latency-measurement: on
>>> network.ping-timeout: 30
>>> server.allow-insecure: on
>>> storage.owner-gid: 36
>>> storage.owner-uid: 36
>>> user.cifs: off
>>> features.shard: on
>>> cluster.shd-wait-qlength: 10000
>>> cluster.shd-max-threads: 8
>>> cluster.locking-scheme: granular
>>> cluster.data-self-heal-algorithm: full
>>> cluster.server-quorum-type: server
>>> cluster.quorum-type: auto
>>> cluster.eager-lock: enable
>>> network.remote-dio: enable
>>> performance.low-prio-threads: 32
>>> performance.stat-prefetch: off
>>> performance.io-cache: off
>>> performance.read-ahead: off
>>> performance.quick-read: off
>>> changelog.changelog: on
>>> geo-replication.ignore-pid-check: on
>>> geo-replication.indexing: on
>>> transport.address-family: inet
>>> performance.readdir-ahead: on
>>>
>>> Volume Name: engine
>>> Type: Replicate
>>> Volume ID: 87ad86b9-d88b-457e-ba21-5d3173c612de
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 1 x (2 + 1) = 3
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: ovirt1.nwfiber.com:/gluster/brick1/engine
>>> Brick2: ovirt2.nwfiber.com:/gluster/brick1/engine
>>> Brick3: ovirt3.nwfiber.com:/gluster/brick1/engine (arbiter)
>>> Options Reconfigured:
>>> changelog.changelog: on
>>> geo-replication.ignore-pid-check: on
>>> geo-replication.indexing: on
>>> performance.readdir-ahead: on
>>> performance.quick-read: off
>>> performance.read-ahead: off
>>> performance.io-cache: off
>>> performance.stat-prefetch: off
>>> cluster.eager-lock: enable
>>> network.remote-dio: off
>>> cluster.quorum-type: auto
>>> cluster.server-quorum-type: server
>>> storage.owner-uid: 36
>>> storage.owner-gid: 36
>>> features.shard: on
>>> features.shard-block-size: 512MB
>>> performance.low-prio-threads: 32
>>> cluster.data-self-heal-algorithm: full
>>> cluster.locking-scheme: granular
>>> cluster.shd-wait-qlength: 10000
>>> cluster.shd-max-threads: 6
>>> network.ping-timeout: 30
>>> user.cifs: off
>>> nfs.disable: on
>>> performance.strict-o-direct: on
>>>
>>> Volume Name: iso
>>> Type: Replicate
>>> Volume ID: b1ba15f5-0f0f-4411-89d0-595179f02b92
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 1 x (2 + 1) = 3
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: ovirt1.nwfiber.com:/gluster/brick4/iso
>>> Brick2: ovirt2.nwfiber.com:/gluster/brick4/iso
>>> Brick3: ovirt3.nwfiber.com:/gluster/brick4/iso (arbiter)
>>> Options Reconfigured:
>>> performance.strict-o-direct: on
>>> nfs.disable: on
>>> user.cifs: off
>>> network.ping-timeout: 30
>>> cluster.shd-max-threads: 6
>>> cluster.shd-wait-qlength: 10000
>>> cluster.locking-scheme: granular
>>> cluster.data-self-heal-algorithm: full
>>> performance.low-prio-threads: 32
>>> features.shard-block-size: 512MB
>>> features.shard: on
>>> storage.owner-gid: 36
>>> storage.owner-uid: 36
>>> cluster.server-quorum-type: server
>>> cluster.quorum-type: auto
>>> network.remote-dio: off
>>> cluster.eager-lock: enable
>>> performance.stat-prefetch: off
>>> performance.io-cache: off
>>> performance.read-ahead: off
>>> performance.quick-read: off
>>> performance.readdir-ahead: on
>>>
>>>
>>> [root@ovirt2 images]# gluster volume status
>>> Status of volume: data
>>> Gluster process TCP Port RDMA Port
>>> Online Pid
>>>
>>>
------------------------------------------------------------------------------
>>> Brick ovirt1.nwfiber.com:/gluster/brick2/da
>>> ta 49152 0 Y
>>> 3226
>>> Brick ovirt2.nwfiber.com:/gluster/brick2/da
>>> ta 49152 0 Y
>>> 2967
>>> Brick ovirt3.nwfiber.com:/gluster/brick2/da
>>> ta 49152 0 Y
>>> 2554
>>> Self-heal Daemon on localhost N/A N/A Y
>>> 4818
>>> Self-heal Daemon on
ovirt3.nwfiber.com N/A N/A Y
>>> 17853
>>> Self-heal Daemon on
ovirt1.nwfiber.com N/A N/A Y
>>> 4771
>>>
>>> Task Status of Volume data
>>>
>>>
------------------------------------------------------------------------------
>>> There are no active volume tasks
>>>
>>> Status of volume: data-hdd
>>> Gluster process TCP Port RDMA Port
>>> Online Pid
>>>
>>>
------------------------------------------------------------------------------
>>> Brick 172.172.1.11:/gluster/brick3/data-hdd 49153 0 Y
>>> 3232
>>> Brick 172.172.1.12:/gluster/brick3/data-hdd 49153 0 Y
>>> 2976
>>> Brick 172.172.1.13:/gluster/brick3/data-hdd N/A N/A N
>>> N/A
>>> NFS Server on localhost N/A N/A N
>>> N/A
>>> Self-heal Daemon on localhost N/A N/A Y
>>> 4818
>>> NFS Server on
ovirt3.nwfiber.com N/A N/A N
>>> N/A
>>> Self-heal Daemon on
ovirt3.nwfiber.com N/A N/A Y
>>> 17853
>>> NFS Server on
ovirt1.nwfiber.com N/A N/A N
>>> N/A
>>> Self-heal Daemon on
ovirt1.nwfiber.com N/A N/A Y
>>> 4771
>>>
>>> Task Status of Volume data-hdd
>>>
>>>
------------------------------------------------------------------------------
>>> There are no active volume tasks
>>>
>>> Status of volume: engine
>>> Gluster process TCP Port RDMA Port
>>> Online Pid
>>>
>>>
------------------------------------------------------------------------------
>>> Brick ovirt1.nwfiber.com:/gluster/brick1/en
>>> gine 49154 0 Y
>>> 3239
>>> Brick ovirt2.nwfiber.com:/gluster/brick1/en
>>> gine 49154 0 Y
>>> 2982
>>> Brick ovirt3.nwfiber.com:/gluster/brick1/en
>>> gine 49154 0 Y
>>> 2578
>>> Self-heal Daemon on localhost N/A N/A Y
>>> 4818
>>> Self-heal Daemon on
ovirt3.nwfiber.com N/A N/A Y
>>> 17853
>>> Self-heal Daemon on
ovirt1.nwfiber.com N/A N/A Y
>>> 4771
>>>
>>> Task Status of Volume engine
>>>
>>>
------------------------------------------------------------------------------
>>> There are no active volume tasks
>>>
>>> Status of volume: iso
>>> Gluster process TCP Port RDMA Port
>>> Online Pid
>>>
>>>
------------------------------------------------------------------------------
>>> Brick ovirt1.nwfiber.com:/gluster/brick4/is
>>> o 49155 0 Y
>>> 3247
>>> Brick ovirt2.nwfiber.com:/gluster/brick4/is
>>> o 49155 0 Y
>>> 2990
>>> Brick ovirt3.nwfiber.com:/gluster/brick4/is
>>> o 49155 0 Y
>>> 2580
>>> Self-heal Daemon on localhost N/A N/A Y
>>> 4818
>>> Self-heal Daemon on
ovirt3.nwfiber.com N/A N/A Y
>>> 17853
>>> Self-heal Daemon on
ovirt1.nwfiber.com N/A N/A Y
>>> 4771
>>>
>>> Task Status of Volume iso
>>>
>>>
------------------------------------------------------------------------------
>>> There are no active volume tasks
>>>
>>>
>>> On Tue, May 29, 2018 at 11:30 PM, Sahina Bose <sabose(a)redhat.com>
>>> wrote:
>>>
>>>>
>>>>
>>>> On Wed, May 30, 2018 at 10:42 AM, Jim Kusznir
<jim(a)palousetech.com>
>>>> wrote:
>>>>
>>>>> hosted-engine --deploy failed (would not come up on my existing
>>>>> gluster storage). However, I realized no changes were written to my
>>>>> existing storage. So, I went back to trying to get my old engine
running.
>>>>>
>>>>> hosted-engine --vm-status is now taking a very long time (5+minutes)
>>>>> to return, and it returns stail information everywhere. I thought
perhaps
>>>>> the lockspace is corrupt, so tried to clean that and metadata, but
both are
>>>>> failing (--cleam-metadata has hung and I can't even ctrl-c out of
it).
>>>>>
>>>>> How can I reinitialize all the lockspace/metadata safely? There is
>>>>> no engine or VMs running currently....
>>>>>
>>>>
>>>> I think the first thing to make sure is that your storage is up and
>>>> running. So can you mount the gluster volumes and able to access the
>>>> contents there?
>>>> Please provide the gluster volume info and gluster volume status of
>>>> the volumes that you're using.
>>>>
>>>>
>>>>
>>>>> --Jim
>>>>>
>>>>> On Tue, May 29, 2018 at 9:33 PM, Jim Kusznir
<jim(a)palousetech.com>
>>>>> wrote:
>>>>>
>>>>>> Well, things went from bad to very, very bad....
>>>>>>
>>>>>> It appears that during one of the 2 minute lockups, the fencing
>>>>>> agents decided that another node in the cluster was down. As a
result, 2
>>>>>> of the 3 nodes were simultaneously reset with fencing agent
reboot. After
>>>>>> the nodes came back up, the engine would not start. All running
VMs
>>>>>> (including VMs on the 3rd node that was not rebooted) crashed.
>>>>>>
>>>>>> I've now been working for about 3 hours trying to get the
engine to
>>>>>> come up. I don't know why it won't start. hosted-engine
--vm-start says
>>>>>> its starting, but it doesn't start (virsh doesn't show
any VMs running).
>>>>>> I'm currently running --deploy, as I had run out of options
for anything
>>>>>> else I can come up with. I hope this will allow me to re-import
all my
>>>>>> existing VMs and allow me to start them back up after everything
comes back
>>>>>> up.
>>>>>>
>>>>>> I do have an unverified geo-rep backup; I don't know if it is
a good
>>>>>> backup (there were several prior messages to this list, but I
didn't get
>>>>>> replies to my questions. It was running in what I believe to be
"strange",
>>>>>> and the data directories are larger than their source).
>>>>>>
>>>>>> I'll see if my --deploy works, and if not, I'll be back
with another
>>>>>> message/help request.
>>>>>>
>>>>>> When the dust settles and I'm at least minimally functional
again, I
>>>>>> really want to understand why all these technologies designed to
offer
>>>>>> redundancy conspired to reduce uptime and create failures where
there
>>>>>> weren't any otherwise. I thought with hosted engine, 3 ovirt
servers and
>>>>>> glusterfs with minimum replica 2+arb or replica 3 should have
offered
>>>>>> strong resilience against server failure or disk failure, and
should have
>>>>>> prevented / recovered from data corruption. Instead, all of the
above
>>>>>> happened (once I get my cluster back up, I still have to try and
recover my
>>>>>> webserver VM, which won't boot due to XFS corrupt journal
issues created
>>>>>> during the gluster crashes). I think a lot of these issues were
rooted
>>>>>> from the upgrade from 4.1 to 4.2.
>>>>>>
>>>>>> --Jim
>>>>>>
>>>>>> On Tue, May 29, 2018 at 6:25 PM, Jim Kusznir
<jim(a)palousetech.com>
>>>>>> wrote:
>>>>>>
>>>>>>> I also finally found the following in my system log on one
server:
>>>>>>>
>>>>>>> [10679.524491] INFO: task glusterclogro:14933 blocked for
more than
>>>>>>> 120 seconds.
>>>>>>> [10679.525826] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [10679.527144] glusterclogro D ffff97209832bf40 0
14933
>>>>>>> 1 0x00000080
>>>>>>> [10679.527150] Call Trace:
>>>>>>> [10679.527161] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [10679.527218] [<ffffffffc060e388>]
_xfs_log_force_lsn+0x2e8/0x340
>>>>>>> [xfs]
>>>>>>> [10679.527225] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [10679.527254] [<ffffffffc05eeb97>]
xfs_file_fsync+0x107/0x1e0
>>>>>>> [xfs]
>>>>>>> [10679.527260] [<ffffffffb944f0e7>]
do_fsync+0x67/0xb0
>>>>>>> [10679.527268] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [10679.527271] [<ffffffffb944f3d0>]
SyS_fsync+0x10/0x20
>>>>>>> [10679.527275] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [10679.527279] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [10679.527283] INFO: task glusterposixfsy:14941 blocked for
more
>>>>>>> than 120 seconds.
>>>>>>> [10679.528608] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [10679.529956] glusterposixfsy D ffff972495f84f10 0
14941
>>>>>>> 1 0x00000080
>>>>>>> [10679.529961] Call Trace:
>>>>>>> [10679.529966] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [10679.530003] [<ffffffffc060e388>]
_xfs_log_force_lsn+0x2e8/0x340
>>>>>>> [xfs]
>>>>>>> [10679.530008] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [10679.530038] [<ffffffffc05eeb97>]
xfs_file_fsync+0x107/0x1e0
>>>>>>> [xfs]
>>>>>>> [10679.530042] [<ffffffffb944f0e7>]
do_fsync+0x67/0xb0
>>>>>>> [10679.530046] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [10679.530050] [<ffffffffb944f3f3>]
SyS_fdatasync+0x13/0x20
>>>>>>> [10679.530054] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [10679.530058] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [10679.530062] INFO: task glusteriotwr13:15486 blocked for
more
>>>>>>> than 120 seconds.
>>>>>>> [10679.531805] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [10679.533732] glusteriotwr13 D ffff9720a83f0000 0
15486
>>>>>>> 1 0x00000080
>>>>>>> [10679.533738] Call Trace:
>>>>>>> [10679.533747] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [10679.533799] [<ffffffffc060e388>]
_xfs_log_force_lsn+0x2e8/0x340
>>>>>>> [xfs]
>>>>>>> [10679.533806] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [10679.533846] [<ffffffffc05eeb97>]
xfs_file_fsync+0x107/0x1e0
>>>>>>> [xfs]
>>>>>>> [10679.533852] [<ffffffffb944f0e7>]
do_fsync+0x67/0xb0
>>>>>>> [10679.533858] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [10679.533863] [<ffffffffb944f3f3>]
SyS_fdatasync+0x13/0x20
>>>>>>> [10679.533868] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [10679.533873] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [10919.512757] INFO: task glusterclogro:14933 blocked for
more than
>>>>>>> 120 seconds.
>>>>>>> [10919.514714] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [10919.516663] glusterclogro D ffff97209832bf40 0
14933
>>>>>>> 1 0x00000080
>>>>>>> [10919.516677] Call Trace:
>>>>>>> [10919.516690] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [10919.516696] [<ffffffffb99118e9>]
schedule_timeout+0x239/0x2c0
>>>>>>> [10919.516703] [<ffffffffb951cc04>] ?
blk_finish_plug+0x14/0x40
>>>>>>> [10919.516768] [<ffffffffc05e9224>] ?
_xfs_buf_ioapply+0x334/0x460
>>>>>>> [xfs]
>>>>>>> [10919.516774] [<ffffffffb991432d>]
wait_for_completion+0xfd/0x140
>>>>>>> [10919.516782] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [10919.516821] [<ffffffffc05eb0a3>] ?
_xfs_buf_read+0x23/0x40 [xfs]
>>>>>>> [10919.516859] [<ffffffffc05eafa9>]
xfs_buf_submit_wait+0xf9/0x1d0
>>>>>>> [xfs]
>>>>>>> [10919.516902] [<ffffffffc061b279>] ?
>>>>>>> xfs_trans_read_buf_map+0x199/0x400 [xfs]
>>>>>>> [10919.516940] [<ffffffffc05eb0a3>]
_xfs_buf_read+0x23/0x40 [xfs]
>>>>>>> [10919.516977] [<ffffffffc05eb1b9>]
xfs_buf_read_map+0xf9/0x160
>>>>>>> [xfs]
>>>>>>> [10919.517022] [<ffffffffc061b279>]
>>>>>>> xfs_trans_read_buf_map+0x199/0x400 [xfs]
>>>>>>> [10919.517057] [<ffffffffc05c8d04>]
xfs_da_read_buf+0xd4/0x100
>>>>>>> [xfs]
>>>>>>> [10919.517091] [<ffffffffc05c8d53>]
xfs_da3_node_read+0x23/0xd0
>>>>>>> [xfs]
>>>>>>> [10919.517126] [<ffffffffc05c9fee>]
>>>>>>> xfs_da3_node_lookup_int+0x6e/0x2f0 [xfs]
>>>>>>> [10919.517160] [<ffffffffc05d5a1d>]
>>>>>>> xfs_dir2_node_lookup+0x4d/0x170 [xfs]
>>>>>>> [10919.517194] [<ffffffffc05ccf5d>]
xfs_dir_lookup+0x1bd/0x1e0
>>>>>>> [xfs]
>>>>>>> [10919.517233] [<ffffffffc05fd8d9>]
xfs_lookup+0x69/0x140 [xfs]
>>>>>>> [10919.517271] [<ffffffffc05fa018>]
xfs_vn_lookup+0x78/0xc0 [xfs]
>>>>>>> [10919.517278] [<ffffffffb9425cf3>]
lookup_real+0x23/0x60
>>>>>>> [10919.517283] [<ffffffffb9426702>]
__lookup_hash+0x42/0x60
>>>>>>> [10919.517288] [<ffffffffb942d519>]
SYSC_renameat2+0x3a9/0x5a0
>>>>>>> [10919.517296] [<ffffffffb94d3753>] ?
>>>>>>> selinux_file_free_security+0x23/0x30
>>>>>>> [10919.517304] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [10919.517309] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [10919.517313] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [10919.517318] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [10919.517323] [<ffffffffb942e58e>]
SyS_renameat2+0xe/0x10
>>>>>>> [10919.517328] [<ffffffffb942e5ce>]
SyS_rename+0x1e/0x20
>>>>>>> [10919.517333] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [10919.517339] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [11159.496095] INFO: task glusteriotwr9:15482 blocked for
more than
>>>>>>> 120 seconds.
>>>>>>> [11159.497546] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [11159.498978] glusteriotwr9 D ffff971fa0fa1fa0 0
15482
>>>>>>> 1 0x00000080
>>>>>>> [11159.498984] Call Trace:
>>>>>>> [11159.498995] [<ffffffffb9911f00>] ?
bit_wait+0x50/0x50
>>>>>>> [11159.498999] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [11159.499003] [<ffffffffb99118e9>]
schedule_timeout+0x239/0x2c0
>>>>>>> [11159.499056] [<ffffffffc05dd9b7>] ?
>>>>>>> xfs_iext_bno_to_ext+0xa7/0x1a0 [xfs]
>>>>>>> [11159.499082] [<ffffffffc05dd43e>] ?
>>>>>>> xfs_iext_bno_to_irec+0x8e/0xd0 [xfs]
>>>>>>> [11159.499090] [<ffffffffb92f7a12>] ?
ktime_get_ts64+0x52/0xf0
>>>>>>> [11159.499093] [<ffffffffb9911f00>] ?
bit_wait+0x50/0x50
>>>>>>> [11159.499097] [<ffffffffb991348d>]
io_schedule_timeout+0xad/0x130
>>>>>>> [11159.499101] [<ffffffffb9913528>]
io_schedule+0x18/0x20
>>>>>>> [11159.499104] [<ffffffffb9911f11>]
bit_wait_io+0x11/0x50
>>>>>>> [11159.499107] [<ffffffffb9911ac1>]
__wait_on_bit_lock+0x61/0xc0
>>>>>>> [11159.499113] [<ffffffffb9393634>]
__lock_page+0x74/0x90
>>>>>>> [11159.499118] [<ffffffffb92bc210>] ?
wake_bit_function+0x40/0x40
>>>>>>> [11159.499121] [<ffffffffb9394154>]
__find_lock_page+0x54/0x70
>>>>>>> [11159.499125] [<ffffffffb9394e85>]
>>>>>>> grab_cache_page_write_begin+0x55/0xc0
>>>>>>> [11159.499130] [<ffffffffb9484b76>]
iomap_write_begin+0x66/0x100
>>>>>>> [11159.499135] [<ffffffffb9484edf>]
iomap_write_actor+0xcf/0x1d0
>>>>>>> [11159.499140] [<ffffffffb9484e10>] ?
iomap_write_end+0x80/0x80
>>>>>>> [11159.499144] [<ffffffffb94854e7>]
iomap_apply+0xb7/0x150
>>>>>>> [11159.499149] [<ffffffffb9485621>]
>>>>>>> iomap_file_buffered_write+0xa1/0xe0
>>>>>>> [11159.499153] [<ffffffffb9484e10>] ?
iomap_write_end+0x80/0x80
>>>>>>> [11159.499182] [<ffffffffc05f025d>]
>>>>>>> xfs_file_buffered_aio_write+0x12d/0x2c0 [xfs]
>>>>>>> [11159.499213] [<ffffffffc05f057d>]
xfs_file_aio_write+0x18d/0x1b0
>>>>>>> [xfs]
>>>>>>> [11159.499217] [<ffffffffb941a533>]
do_sync_write+0x93/0xe0
>>>>>>> [11159.499222] [<ffffffffb941b010>]
vfs_write+0xc0/0x1f0
>>>>>>> [11159.499225] [<ffffffffb941c002>]
SyS_pwrite64+0x92/0xc0
>>>>>>> [11159.499230] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [11159.499234] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [11159.499238] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [11279.488720] INFO: task xfsaild/dm-10:1134 blocked for more
than
>>>>>>> 120 seconds.
>>>>>>> [11279.490197] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [11279.491665] xfsaild/dm-10 D ffff9720a8660fd0 0
1134
>>>>>>> 2 0x00000000
>>>>>>> [11279.491671] Call Trace:
>>>>>>> [11279.491682] [<ffffffffb92a3a2e>] ?
>>>>>>> try_to_del_timer_sync+0x5e/0x90
>>>>>>> [11279.491688] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [11279.491744] [<ffffffffc060de36>]
_xfs_log_force+0x1c6/0x2c0
>>>>>>> [xfs]
>>>>>>> [11279.491750] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [11279.491783] [<ffffffffc0619fec>] ?
xfsaild+0x16c/0x6f0 [xfs]
>>>>>>> [11279.491817] [<ffffffffc060df5c>]
xfs_log_force+0x2c/0x70 [xfs]
>>>>>>> [11279.491849] [<ffffffffc0619e80>] ?
>>>>>>> xfs_trans_ail_cursor_first+0x90/0x90 [xfs]
>>>>>>> [11279.491880] [<ffffffffc0619fec>]
xfsaild+0x16c/0x6f0 [xfs]
>>>>>>> [11279.491913] [<ffffffffc0619e80>] ?
>>>>>>> xfs_trans_ail_cursor_first+0x90/0x90 [xfs]
>>>>>>> [11279.491919] [<ffffffffb92bb161>] kthread+0xd1/0xe0
>>>>>>> [11279.491926] [<ffffffffb92bb090>] ?
insert_kthread_work+0x40/0x40
>>>>>>> [11279.491932] [<ffffffffb9920677>]
>>>>>>> ret_from_fork_nospec_begin+0x21/0x21
>>>>>>> [11279.491936] [<ffffffffb92bb090>] ?
insert_kthread_work+0x40/0x40
>>>>>>> [11279.491976] INFO: task glusterclogfsyn:14934 blocked for
more
>>>>>>> than 120 seconds.
>>>>>>> [11279.493466] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [11279.494952] glusterclogfsyn D ffff97209832af70 0
14934
>>>>>>> 1 0x00000080
>>>>>>> [11279.494957] Call Trace:
>>>>>>> [11279.494979] [<ffffffffc0309839>] ?
>>>>>>> __split_and_process_bio+0x2e9/0x520 [dm_mod]
>>>>>>> [11279.494983] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [11279.494987] [<ffffffffb99118e9>]
schedule_timeout+0x239/0x2c0
>>>>>>> [11279.494997] [<ffffffffc0309d98>] ?
dm_make_request+0x128/0x1a0
>>>>>>> [dm_mod]
>>>>>>> [11279.495001] [<ffffffffb991348d>]
io_schedule_timeout+0xad/0x130
>>>>>>> [11279.495005] [<ffffffffb99145ad>]
>>>>>>> wait_for_completion_io+0xfd/0x140
>>>>>>> [11279.495010] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [11279.495016] [<ffffffffb951e574>]
blkdev_issue_flush+0xb4/0x110
>>>>>>> [11279.495049] [<ffffffffc06064b9>]
>>>>>>> xfs_blkdev_issue_flush+0x19/0x20 [xfs]
>>>>>>> [11279.495079] [<ffffffffc05eec40>]
xfs_file_fsync+0x1b0/0x1e0
>>>>>>> [xfs]
>>>>>>> [11279.495086] [<ffffffffb944f0e7>]
do_fsync+0x67/0xb0
>>>>>>> [11279.495090] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [11279.495094] [<ffffffffb944f3d0>]
SyS_fsync+0x10/0x20
>>>>>>> [11279.495098] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [11279.495102] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [11279.495105] INFO: task glusterposixfsy:14941 blocked for
more
>>>>>>> than 120 seconds.
>>>>>>> [11279.496606] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [11279.498114] glusterposixfsy D ffff972495f84f10 0
14941
>>>>>>> 1 0x00000080
>>>>>>> [11279.498118] Call Trace:
>>>>>>> [11279.498134] [<ffffffffc0309839>] ?
>>>>>>> __split_and_process_bio+0x2e9/0x520 [dm_mod]
>>>>>>> [11279.498138] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [11279.498142] [<ffffffffb99118e9>]
schedule_timeout+0x239/0x2c0
>>>>>>> [11279.498152] [<ffffffffc0309d98>] ?
dm_make_request+0x128/0x1a0
>>>>>>> [dm_mod]
>>>>>>> [11279.498156] [<ffffffffb991348d>]
io_schedule_timeout+0xad/0x130
>>>>>>> [11279.498160] [<ffffffffb99145ad>]
>>>>>>> wait_for_completion_io+0xfd/0x140
>>>>>>> [11279.498165] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [11279.498169] [<ffffffffb951e574>]
blkdev_issue_flush+0xb4/0x110
>>>>>>> [11279.498202] [<ffffffffc06064b9>]
>>>>>>> xfs_blkdev_issue_flush+0x19/0x20 [xfs]
>>>>>>> [11279.498231] [<ffffffffc05eec40>]
xfs_file_fsync+0x1b0/0x1e0
>>>>>>> [xfs]
>>>>>>> [11279.498238] [<ffffffffb944f0e7>]
do_fsync+0x67/0xb0
>>>>>>> [11279.498242] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [11279.498246] [<ffffffffb944f3f3>]
SyS_fdatasync+0x13/0x20
>>>>>>> [11279.498250] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [11279.498254] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [11279.498257] INFO: task glusteriotwr1:14950 blocked for
more than
>>>>>>> 120 seconds.
>>>>>>> [11279.499789] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [11279.501343] glusteriotwr1 D ffff97208b6daf70 0
14950
>>>>>>> 1 0x00000080
>>>>>>> [11279.501348] Call Trace:
>>>>>>> [11279.501353] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [11279.501390] [<ffffffffc060e388>]
_xfs_log_force_lsn+0x2e8/0x340
>>>>>>> [xfs]
>>>>>>> [11279.501396] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [11279.501428] [<ffffffffc05eeb97>]
xfs_file_fsync+0x107/0x1e0
>>>>>>> [xfs]
>>>>>>> [11279.501432] [<ffffffffb944ef3f>]
generic_write_sync+0x4f/0x70
>>>>>>> [11279.501461] [<ffffffffc05f0545>]
xfs_file_aio_write+0x155/0x1b0
>>>>>>> [xfs]
>>>>>>> [11279.501466] [<ffffffffb941a533>]
do_sync_write+0x93/0xe0
>>>>>>> [11279.501471] [<ffffffffb941b010>]
vfs_write+0xc0/0x1f0
>>>>>>> [11279.501475] [<ffffffffb941c002>]
SyS_pwrite64+0x92/0xc0
>>>>>>> [11279.501479] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [11279.501483] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [11279.501489] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [11279.501493] INFO: task glusteriotwr4:14953 blocked for
more than
>>>>>>> 120 seconds.
>>>>>>> [11279.503047] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
>>>>>>> disables this message.
>>>>>>> [11279.504630] glusteriotwr4 D ffff972499f2bf40 0
14953
>>>>>>> 1 0x00000080
>>>>>>> [11279.504635] Call Trace:
>>>>>>> [11279.504640] [<ffffffffb9913f79>]
schedule+0x29/0x70
>>>>>>> [11279.504676] [<ffffffffc060e388>]
_xfs_log_force_lsn+0x2e8/0x340
>>>>>>> [xfs]
>>>>>>> [11279.504681] [<ffffffffb92cf1b0>] ?
wake_up_state+0x20/0x20
>>>>>>> [11279.504710] [<ffffffffc05eeb97>]
xfs_file_fsync+0x107/0x1e0
>>>>>>> [xfs]
>>>>>>> [11279.504714] [<ffffffffb944f0e7>]
do_fsync+0x67/0xb0
>>>>>>> [11279.504718] [<ffffffffb992076f>] ?
>>>>>>> system_call_after_swapgs+0xbc/0x160
>>>>>>> [11279.504722] [<ffffffffb944f3d0>]
SyS_fsync+0x10/0x20
>>>>>>> [11279.504725] [<ffffffffb992082f>]
system_call_fastpath+0x1c/0x21
>>>>>>> [11279.504730] [<ffffffffb992077b>] ?
>>>>>>> system_call_after_swapgs+0xc8/0x160
>>>>>>> [12127.466494] perf: interrupt took too long (8263 >
8150),
>>>>>>> lowering kernel.perf_event_max_sample_rate to 24000
>>>>>>>
>>>>>>> --------------------
>>>>>>> I think this is the cause of the massive ovirt performance
issues
>>>>>>> irrespective of gluster volume. At the time this happened, I
was also
>>>>>>> ssh'ed into the host, and was doing some rpm querry
commands. I had just
>>>>>>> run rpm -qa |grep glusterfs (to verify what version was
actually
>>>>>>> installed), and that command took almost 2 minutes to return!
Normally it
>>>>>>> takes less than 2 seconds. That is all pure local SSD IO,
too....
>>>>>>>
>>>>>>> I'm no expert, but its my understanding that anytime a
software
>>>>>>> causes these kinds of issues, its a serious bug in the
software, even if
>>>>>>> its mis-handled exceptions. Is this correct?
>>>>>>>
>>>>>>> --Jim
>>>>>>>
>>>>>>> On Tue, May 29, 2018 at 3:01 PM, Jim Kusznir
<jim(a)palousetech.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I think this is the profile information for one of the
volumes
>>>>>>>> that lives on the SSDs and is fully operational with no
down/problem disks:
>>>>>>>>
>>>>>>>> [root@ovirt2 yum.repos.d]# gluster volume profile data
info
>>>>>>>> Brick: ovirt2.nwfiber.com:/gluster/brick2/data
>>>>>>>> ----------------------------------------------
>>>>>>>> Cumulative Stats:
>>>>>>>> Block Size: 256b+
512b+
>>>>>>>> 1024b+
>>>>>>>> No. of Reads: 983
2696
>>>>>>>> 1059
>>>>>>>> No. of Writes: 0
1113
>>>>>>>> 302
>>>>>>>>
>>>>>>>> Block Size: 2048b+
4096b+
>>>>>>>> 8192b+
>>>>>>>> No. of Reads: 852
88608
>>>>>>>> 53526
>>>>>>>> No. of Writes: 522
812340
>>>>>>>> 76257
>>>>>>>>
>>>>>>>> Block Size: 16384b+
32768b+
>>>>>>>> 65536b+
>>>>>>>> No. of Reads: 54351
241901
>>>>>>>> 15024
>>>>>>>> No. of Writes: 21636
8656
>>>>>>>> 8976
>>>>>>>>
>>>>>>>> Block Size: 131072b+
>>>>>>>> No. of Reads: 524156
>>>>>>>> No. of Writes: 296071
>>>>>>>> %-latency Avg-latency Min-Latency Max-Latency
No. of
>>>>>>>> calls Fop
>>>>>>>> --------- ----------- ----------- -----------
>>>>>>>> ------------ ----
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 4189 RELEASE
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 1257 RELEASEDIR
>>>>>>>> 0.00 46.19 us 12.00 us 187.00 us
>>>>>>>> 69 FLUSH
>>>>>>>> 0.00 147.00 us 78.00 us 367.00 us
>>>>>>>> 86 REMOVEXATTR
>>>>>>>> 0.00 223.46 us 24.00 us 1166.00 us
>>>>>>>> 149 READDIR
>>>>>>>> 0.00 565.34 us 76.00 us 3639.00 us
>>>>>>>> 88 FTRUNCATE
>>>>>>>> 0.00 263.28 us 20.00 us 28385.00 us
>>>>>>>> 228 LK
>>>>>>>> 0.00 98.84 us 2.00 us 880.00 us
>>>>>>>> 1198 OPENDIR
>>>>>>>> 0.00 91.59 us 26.00 us 10371.00 us
>>>>>>>> 3853 STATFS
>>>>>>>> 0.00 494.14 us 17.00 us 193439.00 us
>>>>>>>> 1171 GETXATTR
>>>>>>>> 0.00 299.42 us 35.00 us 9799.00 us
>>>>>>>> 2044 READDIRP
>>>>>>>> 0.00 1965.31 us 110.00 us 382258.00 us
>>>>>>>> 321 XATTROP
>>>>>>>> 0.01 113.40 us 24.00 us 61061.00 us
>>>>>>>> 8134 STAT
>>>>>>>> 0.01 755.38 us 57.00 us 607603.00 us
>>>>>>>> 3196 DISCARD
>>>>>>>> 0.05 2690.09 us 58.00 us 2704761.00 us
>>>>>>>> 3206 OPEN
>>>>>>>> 0.10 119978.25 us 97.00 us 9406684.00 us
>>>>>>>> 154 SETATTR
>>>>>>>> 0.18 101.73 us 28.00 us 700477.00 us
>>>>>>>> 313379 FSTAT
>>>>>>>> 0.23 1059.84 us 25.00 us 2716124.00 us
>>>>>>>> 38255 LOOKUP
>>>>>>>> 0.47 1024.11 us 54.00 us 6197164.00 us
>>>>>>>> 81455 FXATTROP
>>>>>>>> 1.72 2984.00 us 15.00 us 37098954.00 us
>>>>>>>> 103020 FINODELK
>>>>>>>> 5.92 44315.32 us 51.00 us 24731536.00 us
>>>>>>>> 23957 FSYNC
>>>>>>>> 13.27 2399.78 us 25.00 us 22089540.00 us
>>>>>>>> 991005 READ
>>>>>>>> 37.00 5980.43 us 52.00 us 22099889.00 us
>>>>>>>> 1108976 WRITE
>>>>>>>> 41.04 5452.75 us 13.00 us 22102452.00 us
>>>>>>>> 1349053 INODELK
>>>>>>>>
>>>>>>>> Duration: 10026 seconds
>>>>>>>> Data Read: 80046027759 bytes
>>>>>>>> Data Written: 44496632320 bytes
>>>>>>>>
>>>>>>>> Interval 1 Stats:
>>>>>>>> Block Size: 256b+
512b+
>>>>>>>> 1024b+
>>>>>>>> No. of Reads: 983
2696
>>>>>>>> 1059
>>>>>>>> No. of Writes: 0
838
>>>>>>>> 185
>>>>>>>>
>>>>>>>> Block Size: 2048b+
4096b+
>>>>>>>> 8192b+
>>>>>>>> No. of Reads: 852
85856
>>>>>>>> 51575
>>>>>>>> No. of Writes: 382
705802
>>>>>>>> 57812
>>>>>>>>
>>>>>>>> Block Size: 16384b+
32768b+
>>>>>>>> 65536b+
>>>>>>>> No. of Reads: 52673
232093
>>>>>>>> 14984
>>>>>>>> No. of Writes: 13499
4908
>>>>>>>> 4242
>>>>>>>>
>>>>>>>> Block Size: 131072b+
>>>>>>>> No. of Reads: 460040
>>>>>>>> No. of Writes: 6411
>>>>>>>> %-latency Avg-latency Min-Latency Max-Latency
No. of
>>>>>>>> calls Fop
>>>>>>>> --------- ----------- ----------- -----------
>>>>>>>> ------------ ----
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 2093 RELEASE
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 1093 RELEASEDIR
>>>>>>>> 0.00 53.38 us 26.00 us 111.00 us
>>>>>>>> 16 FLUSH
>>>>>>>> 0.00 145.14 us 78.00 us 367.00 us
>>>>>>>> 71 REMOVEXATTR
>>>>>>>> 0.00 190.96 us 114.00 us 298.00 us
>>>>>>>> 71 SETATTR
>>>>>>>> 0.00 213.38 us 24.00 us 1145.00 us
>>>>>>>> 90 READDIR
>>>>>>>> 0.00 263.28 us 20.00 us 28385.00 us
>>>>>>>> 228 LK
>>>>>>>> 0.00 101.76 us 2.00 us 880.00 us
>>>>>>>> 1093 OPENDIR
>>>>>>>> 0.01 93.60 us 27.00 us 10371.00 us
>>>>>>>> 3090 STATFS
>>>>>>>> 0.02 537.47 us 17.00 us 193439.00 us
>>>>>>>> 1038 GETXATTR
>>>>>>>> 0.03 297.44 us 35.00 us 9799.00 us
>>>>>>>> 1990 READDIRP
>>>>>>>> 0.03 2357.28 us 110.00 us 382258.00 us
>>>>>>>> 253 XATTROP
>>>>>>>> 0.04 385.93 us 58.00 us 47593.00 us
>>>>>>>> 2091 OPEN
>>>>>>>> 0.04 114.86 us 24.00 us 61061.00 us
>>>>>>>> 7715 STAT
>>>>>>>> 0.06 444.59 us 57.00 us 333240.00 us
>>>>>>>> 3053 DISCARD
>>>>>>>> 0.42 316.24 us 25.00 us 290728.00 us
>>>>>>>> 29823 LOOKUP
>>>>>>>> 0.73 257.92 us 54.00 us 344812.00 us
>>>>>>>> 63296 FXATTROP
>>>>>>>> 1.37 98.30 us 28.00 us 67621.00 us
>>>>>>>> 313172 FSTAT
>>>>>>>> 1.58 2124.69 us 51.00 us 849200.00 us
>>>>>>>> 16717 FSYNC
>>>>>>>> 5.73 162.46 us 52.00 us 748492.00 us
>>>>>>>> 794079 WRITE
>>>>>>>> 7.19 2065.17 us 16.00 us 37098954.00 us
>>>>>>>> 78381 FINODELK
>>>>>>>> 36.44 886.32 us 25.00 us 2216436.00 us
>>>>>>>> 925421 READ
>>>>>>>> 46.30 1178.04 us 13.00 us 1700704.00 us
>>>>>>>> 884635 INODELK
>>>>>>>>
>>>>>>>> Duration: 7485 seconds
>>>>>>>> Data Read: 71250527215 bytes
>>>>>>>> Data Written: 5119903744 bytes
>>>>>>>>
>>>>>>>> Brick: ovirt3.nwfiber.com:/gluster/brick2/data
>>>>>>>> ----------------------------------------------
>>>>>>>> Cumulative Stats:
>>>>>>>> Block Size: 1b+
>>>>>>>> No. of Reads: 0
>>>>>>>> No. of Writes: 3264419
>>>>>>>> %-latency Avg-latency Min-Latency Max-Latency
No. of
>>>>>>>> calls Fop
>>>>>>>> --------- ----------- ----------- -----------
>>>>>>>> ------------ ----
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 90 FORGET
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 9462 RELEASE
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 4254 RELEASEDIR
>>>>>>>> 0.00 50.52 us 13.00 us 190.00 us
>>>>>>>> 71 FLUSH
>>>>>>>> 0.00 186.97 us 87.00 us 713.00 us
>>>>>>>> 86 REMOVEXATTR
>>>>>>>> 0.00 79.32 us 33.00 us 189.00 us
>>>>>>>> 228 LK
>>>>>>>> 0.00 220.98 us 129.00 us 513.00 us
>>>>>>>> 86 SETATTR
>>>>>>>> 0.01 259.30 us 26.00 us 2632.00 us
>>>>>>>> 137 READDIR
>>>>>>>> 0.02 322.76 us 145.00 us 2125.00 us
>>>>>>>> 321 XATTROP
>>>>>>>> 0.03 109.55 us 2.00 us 1258.00 us
>>>>>>>> 1193 OPENDIR
>>>>>>>> 0.05 70.21 us 21.00 us 431.00 us
>>>>>>>> 3196 DISCARD
>>>>>>>> 0.05 169.26 us 21.00 us 2315.00 us
>>>>>>>> 1545 GETXATTR
>>>>>>>> 0.12 176.85 us 63.00 us 2844.00 us
>>>>>>>> 3206 OPEN
>>>>>>>> 0.61 303.49 us 90.00 us 3085.00 us
>>>>>>>> 9633 FSTAT
>>>>>>>> 2.44 305.66 us 28.00 us 3716.00 us
>>>>>>>> 38230 LOOKUP
>>>>>>>> 4.52 266.22 us 55.00 us 53424.00 us
>>>>>>>> 81455 FXATTROP
>>>>>>>> 6.96 1397.99 us 51.00 us 64822.00 us
>>>>>>>> 23889 FSYNC
>>>>>>>> 16.48 84.74 us 25.00 us 6917.00 us
>>>>>>>> 932592 WRITE
>>>>>>>> 30.16 106.90 us 13.00 us 3920189.00 us
>>>>>>>> 1353046 INODELK
>>>>>>>> 38.55 1794.52 us 14.00 us 16210553.00 us
>>>>>>>> 103039 FINODELK
>>>>>>>>
>>>>>>>> Duration: 66562 seconds
>>>>>>>> Data Read: 0 bytes
>>>>>>>> Data Written: 3264419 bytes
>>>>>>>>
>>>>>>>> Interval 1 Stats:
>>>>>>>> Block Size: 1b+
>>>>>>>> No. of Reads: 0
>>>>>>>> No. of Writes: 794080
>>>>>>>> %-latency Avg-latency Min-Latency Max-Latency
No. of
>>>>>>>> calls Fop
>>>>>>>> --------- ----------- ----------- -----------
>>>>>>>> ------------ ----
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 2093 RELEASE
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 1093 RELEASEDIR
>>>>>>>> 0.00 70.31 us 26.00 us 125.00 us
>>>>>>>> 16 FLUSH
>>>>>>>> 0.00 193.10 us 103.00 us 713.00 us
>>>>>>>> 71 REMOVEXATTR
>>>>>>>> 0.01 227.32 us 133.00 us 513.00 us
>>>>>>>> 71 SETATTR
>>>>>>>> 0.01 79.32 us 33.00 us 189.00 us
>>>>>>>> 228 LK
>>>>>>>> 0.01 259.83 us 35.00 us 1138.00 us
>>>>>>>> 89 READDIR
>>>>>>>> 0.03 318.26 us 145.00 us 2047.00 us
>>>>>>>> 253 XATTROP
>>>>>>>> 0.04 112.67 us 3.00 us 1258.00 us
>>>>>>>> 1093 OPENDIR
>>>>>>>> 0.06 167.98 us 23.00 us 1951.00 us
>>>>>>>> 1014 GETXATTR
>>>>>>>> 0.08 70.97 us 22.00 us 431.00 us
>>>>>>>> 3053 DISCARD
>>>>>>>> 0.13 183.78 us 66.00 us 2844.00 us
>>>>>>>> 2091 OPEN
>>>>>>>> 1.01 303.82 us 90.00 us 3085.00 us
>>>>>>>> 9610 FSTAT
>>>>>>>> 3.27 316.59 us 30.00 us 3716.00 us
>>>>>>>> 29820 LOOKUP
>>>>>>>> 5.83 265.79 us 59.00 us 53424.00 us
>>>>>>>> 63296 FXATTROP
>>>>>>>> 7.95 1373.89 us 51.00 us 64822.00 us
>>>>>>>> 16717 FSYNC
>>>>>>>> 23.17 851.99 us 14.00 us 16210553.00 us
>>>>>>>> 78555 FINODELK
>>>>>>>> 24.04 87.44 us 27.00 us 6917.00 us
>>>>>>>> 794081 WRITE
>>>>>>>> 34.36 111.91 us 14.00 us 984871.00 us
>>>>>>>> 886790 INODELK
>>>>>>>>
>>>>>>>> Duration: 7485 seconds
>>>>>>>> Data Read: 0 bytes
>>>>>>>> Data Written: 794080 bytes
>>>>>>>>
>>>>>>>>
>>>>>>>> -----------------------
>>>>>>>> Here is the data from the volume that is backed by the
SHDDs and
>>>>>>>> has one failed disk:
>>>>>>>> [root@ovirt2 yum.repos.d]# gluster volume profile
data-hdd info
>>>>>>>> Brick: 172.172.1.12:/gluster/brick3/data-hdd
>>>>>>>> --------------------------------------------
>>>>>>>> Cumulative Stats:
>>>>>>>> Block Size: 256b+
512b+
>>>>>>>> 1024b+
>>>>>>>> No. of Reads: 1702
86
>>>>>>>> 16
>>>>>>>> No. of Writes: 0
767
>>>>>>>> 71
>>>>>>>>
>>>>>>>> Block Size: 2048b+
4096b+
>>>>>>>> 8192b+
>>>>>>>> No. of Reads: 19
51841
>>>>>>>> 2049
>>>>>>>> No. of Writes: 76
60668
>>>>>>>> 35727
>>>>>>>>
>>>>>>>> Block Size: 16384b+
32768b+
>>>>>>>> 65536b+
>>>>>>>> No. of Reads: 1744
639
>>>>>>>> 1088
>>>>>>>> No. of Writes: 8524
2410
>>>>>>>> 1285
>>>>>>>>
>>>>>>>> Block Size: 131072b+
>>>>>>>> No. of Reads: 771999
>>>>>>>> No. of Writes: 29584
>>>>>>>> %-latency Avg-latency Min-Latency Max-Latency
No. of
>>>>>>>> calls Fop
>>>>>>>> --------- ----------- ----------- -----------
>>>>>>>> ------------ ----
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 2902 RELEASE
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 1517 RELEASEDIR
>>>>>>>> 0.00 197.00 us 197.00 us 197.00 us
>>>>>>>> 1 FTRUNCATE
>>>>>>>> 0.00 70.24 us 16.00 us 758.00 us
>>>>>>>> 51 FLUSH
>>>>>>>> 0.00 143.93 us 82.00 us 305.00 us
>>>>>>>> 57 REMOVEXATTR
>>>>>>>> 0.00 178.63 us 105.00 us 712.00 us
>>>>>>>> 60 SETATTR
>>>>>>>> 0.00 67.30 us 19.00 us 572.00 us
>>>>>>>> 555 LK
>>>>>>>> 0.00 322.80 us 23.00 us 4673.00 us
>>>>>>>> 138 READDIR
>>>>>>>> 0.00 336.56 us 106.00 us 11994.00 us
>>>>>>>> 237 XATTROP
>>>>>>>> 0.00 84.70 us 28.00 us 1071.00 us
>>>>>>>> 3469 STATFS
>>>>>>>> 0.01 387.75 us 2.00 us 146017.00 us
>>>>>>>> 1467 OPENDIR
>>>>>>>> 0.01 148.59 us 21.00 us 64374.00 us
>>>>>>>> 4454 STAT
>>>>>>>> 0.02 783.02 us 16.00 us 93502.00 us
>>>>>>>> 1902 GETXATTR
>>>>>>>> 0.03 1516.10 us 17.00 us 210690.00 us
>>>>>>>> 1364 ENTRYLK
>>>>>>>> 0.03 2555.47 us 300.00 us 674454.00 us
>>>>>>>> 1064 READDIRP
>>>>>>>> 0.07 85.74 us 19.00 us 68340.00 us
>>>>>>>> 62849 FSTAT
>>>>>>>> 0.07 1978.12 us 59.00 us 202596.00 us
>>>>>>>> 2729 OPEN
>>>>>>>> 0.22 708.57 us 15.00 us 394799.00 us
>>>>>>>> 25447 LOOKUP
>>>>>>>> 5.94 2331.74 us 15.00 us 1099530.00 us
>>>>>>>> 207534 FINODELK
>>>>>>>> 7.31 8311.75 us 58.00 us 1800216.00 us
>>>>>>>> 71668 FXATTROP
>>>>>>>> 12.49 7735.19 us 51.00 us 3595513.00 us
>>>>>>>> 131642 WRITE
>>>>>>>> 17.70 957.08 us 16.00 us 13700466.00 us
>>>>>>>> 1508160 INODELK
>>>>>>>> 24.55 2546.43 us 26.00 us 5077347.00 us
>>>>>>>> 786060 READ
>>>>>>>> 31.56 49699.15 us 47.00 us 3746331.00 us
>>>>>>>> 51777 FSYNC
>>>>>>>>
>>>>>>>> Duration: 10101 seconds
>>>>>>>> Data Read: 101562897361 bytes
>>>>>>>> Data Written: 4834450432 bytes
>>>>>>>>
>>>>>>>> Interval 0 Stats:
>>>>>>>> Block Size: 256b+
512b+
>>>>>>>> 1024b+
>>>>>>>> No. of Reads: 1702
86
>>>>>>>> 16
>>>>>>>> No. of Writes: 0
767
>>>>>>>> 71
>>>>>>>>
>>>>>>>> Block Size: 2048b+
4096b+
>>>>>>>> 8192b+
>>>>>>>> No. of Reads: 19
51841
>>>>>>>> 2049
>>>>>>>> No. of Writes: 76
60668
>>>>>>>> 35727
>>>>>>>>
>>>>>>>> Block Size: 16384b+
32768b+
>>>>>>>> 65536b+
>>>>>>>> No. of Reads: 1744
639
>>>>>>>> 1088
>>>>>>>> No. of Writes: 8524
2410
>>>>>>>> 1285
>>>>>>>>
>>>>>>>> Block Size: 131072b+
>>>>>>>> No. of Reads: 771999
>>>>>>>> No. of Writes: 29584
>>>>>>>> %-latency Avg-latency Min-Latency Max-Latency
No. of
>>>>>>>> calls Fop
>>>>>>>> --------- ----------- ----------- -----------
>>>>>>>> ------------ ----
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 2902 RELEASE
>>>>>>>> 0.00 0.00 us 0.00 us 0.00 us
>>>>>>>> 1517 RELEASEDIR
>>>>>>>> 0.00 197.00 us 197.00 us 197.00 us
>>>>>>>> 1 FTRUNCATE
>>>>>>>> 0.00 70.24 us 16.00 us 758.00 us
>>>>>>>> 51 FLUSH
>>>>>>>> 0.00 143.93 us 82.00 us 305.00 us
>>>>>>>> 57 REMOVEXATTR
>>>>>>>> 0.00 178.63 us 105.00 us 712.00 us
>>>>>>>> 60 SETATTR
>>>>>>>> 0.00 67.30 us 19.00 us 572.00 us
>>>>>>>> 555 LK
>>>>>>>> 0.00 322.80 us 23.00 us 4673.00 us
>>>>>>>> 138 READDIR
>>>>>>>> 0.00 336.56 us 106.00 us 11994.00 us
>>>>>>>> 237 XATTROP
>>>>>>>> 0.00 84.70 us 28.00 us 1071.00 us
>>>>>>>> 3469 STATFS
>>>>>>>> 0.01 387.75 us 2.00 us 146017.00 us
>>>>>>>> 1467 OPENDIR
>>>>>>>> 0.01 148.59 us 21.00 us 64374.00 us
>>>>>>>> 4454 STAT
>>>>>>>> 0.02 783.02 us 16.00 us 93502.00 us
>>>>>>>> 1902 GETXATTR
>>>>>>>> 0.03 1516.10 us 17.00 us 210690.00 us
>>>>>>>> 1364 ENTRYLK
>>>>>>>> 0.03 2555.47 us 300.00 us 674454.00 us
>>>>>>>> 1064 READDIRP
>>>>>>>> 0.07 85.73 us 19.00 us 68340.00 us
>>>>>>>> 62849 FSTAT
>>>>>>>> 0.07 1978.12 us 59.00 us 202596.00 us
>>>>>>>> 2729 OPEN
>>>>>>>> 0.22 708.57 us 15.00 us 394799.00 us
>>>>>>>> 25447 LOOKUP
>>>>>>>> 5.94 2334.57 us 15.00 us 1099530.00 us
>>>>>>>> 207534 FINODELK
>>>>>>>> 7.31 8311.49 us 58.00 us 1800216.00 us
>>>>>>>> 71668 FXATTROP
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>>>>> oVirt Code of Conduct:
>>>>>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>> List Archives:
>>>>>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YJ5J7LHDNM6...
>>>>>>>>
>>>>>>>