New setup - Failing to Activate storage domain on NFS shared storage

I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the same problem with both versions. The issue occurs in both cockpit UI and tmux'd CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is created and running. I tried to do some debugging on my own before reaching out to this list. Any help is much appreciated! ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4 cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but I believe it meets the minimum requirements. NFS server: * Ubuntu 19.10 w/ ZFS share w/ 17TB available space. * NFS share settings are just 'rw=@172.16.1.0/24' but have also tried 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure' * The target directory is always empty and chown'd 36:36 with 0755 permissions. * I have tried using both IP and DNS names. forward and reverse DNS works from ovirt host and other systems on the network. * The NFS share always gets mounted successfully on the ovirt node system. * I have tried auto and v3 NFS versions in other various combinations. * I have also tried setting up an NFS server on a non-ZFS backed storage system that is open to any host and get the same errors as shown below. * I ran nfs-check.py script without issue against both NFS servers and followed other verification steps listed on https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html ***Snip from ovirt-hosted-engine-setup*** Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]: Please specify the full shared storage connection path to use (example: host:/path): stumpy.mydomain.com:/tanker/ovirt/host_storage If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: rw [ INFO ] Creating Storage Domain [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Force facts gathering] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the storage interface to be up] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check local VM dir stat] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Enforce local VM dir existence] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch host facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter name] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add NFS storage domain] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add glusterfs storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add iSCSI storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add Fibre Channel storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get storage domain details] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the appliance OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Parse OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get required size] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Remove unsuitable storage domain] [ INFO ] skipping: [localhost] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check storage domain free space] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: ***End snippet** The relevant section from /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-ansible-create_storage_domain-20210116195220-bnfg1w.log: **Snip*** 2021-01-16 19:53:56,010-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a04c8320> kwargs 2021-01-16 19:53:57,219-0700 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Activate storage domain'} 2021-01-16 19:53:57,220-0700 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs is_conditional:False 2021-01-16 19:53:57,221-0700 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs 2021-01-16 19:54:00,346-0700 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,347-0700 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,348-0700 DEBUG var changed: host "localhost" var "play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,349-0700 ERROR ansible failed { "ansible_host": "localhost", "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_result": { "_ansible_no_log": false, "changed": false, "exception": "Traceback (most recent call last):\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 783, in main\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 638, in post_create_check\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 3647, in add\n return self._internal_add(storage_domain, headers, query, wait)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n return future.wait() if wait else future\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in wait\n return self._code(response)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in callback\n self._check_fault(respon se)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n self._raise_error(response, body)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n", "invocation": { "module_args": { "backup": null, "comment": null, "critical_space_action_blocker": null, "data_center": "Default", "description": null, "destroy": null, "discard_after_delete": null, "domain_function": "data", "fcp": null, "fetch_nested": false, "format": null, "glusterfs": null, "host": "brick.mydomain.com", "id": null, "iscsi": null, "localfs": null, "managed_block_storage": null, "name": "hosted_storage", "nested_attributes": [], "nfs": null, "poll_interval": 3, "posixfs": null, "state": "present", "timeout": 180, "wait": true, "warning_low_space": null, "wipe_after_delete": null } }, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400." }, "ansible_task": "Activate storage domain", "ansible_type": "task", "status": "FAILED", "task_duration": 4 } 2021-01-16 19:54:00,350-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a05f2d30> kwargs ignore_errors:None 2021-01-16 19:54:00,358-0700 INFO ansible stats { "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_playbook_duration": "01:35 Minutes", "ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 23, 'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 7, 'rescued': 0, 'ignored': 0}}", "ansible_type": "finish", "status": "FAILED" } 2021-01-16 19:54:00,359-0700 INFO SUMMARY: Duration Task Name -------- -------- [ < 1 sec ] Execute just a specific set of steps [ 00:05 ] Force facts gathering [ 00:03 ] Check local VM dir stat [ 00:04 ] Obtain SSO token using username/password credentials [ 00:04 ] Fetch host facts [ 00:02 ] Fetch cluster ID [ 00:04 ] Fetch cluster facts [ 00:04 ] Fetch Datacenter facts [ 00:02 ] Fetch Datacenter ID [ 00:02 ] Fetch Datacenter name [ 00:04 ] Add NFS storage domain [ 00:04 ] Get storage domain details [ 00:03 ] Find the appliance OVF [ 00:03 ] Parse OVF [ 00:02 ] Get required size [ FAILED ] Activate storage domain 2021-01-16 19:54:00,359-0700 DEBUG ansible on_any args <ansible.executor.stats.AggregateStats object at 0x7f11a2e3b8d0> kwargs **End snip** The relevantsection from /var/log/vdsm/vdsm.log: ***begin snip*** 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'status': 0}]} from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:54) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [vdsm.api] START getStorageDomainInfo(sdUUID='54532dd4-3e5b-4885-b88e-599c81efb146', options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:48) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [vdsm.api] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:52) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-131>", line 2, in getStorageDomainInfo File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2796, in getStorageDomainInfo dom = self.validateSdUUID(sdUUID) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 312, in validateSdUUID sdDom.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.Dispatcher] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,919-0700 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call StorageDomain.getInfo failed (error 379) in 0.00 seconds (__init__:312) 2021-01-16 19:53:58,956-0700 INFO (jsonrpc/1) [vdsm.api] START createStoragePool(poolType=None, spUUID='91fb3f22-5795-11eb-ad9f-00163e3f4683', poolName='Default', masterDom='54532dd4-3e5b-4885-b88e-599c81efb146', domList=['54532dd4-3e5b-4885-b88e-599c81efb146'], masterVersion=7, lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, ioOpTimeoutSec=10, leaseRetries=3, options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:48) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] updating pool 91fb3f22-5795-11eb-ad9f-00163e3f4683 backend from type NoneType instance 0x7f86e45f89d0 to type StoragePoolDiskBackend instance 0x7f8680738408 (sp:157) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] spUUID=91fb3f22-5795-11eb-ad9f-00163e3f4683 poolName=Default master_sd=54532dd4-3e5b-4885-b88e-599c81efb146 domList=['54532dd4-3e5b-4885-b88e-599c81efb146'] masterVersion=7 {'LEASERETRIES': 3, 'LEASETIMESEC': 60, 'LOCKRENEWALINTERVALSEC': 5, 'IOOPTIMEOUTSEC': 10} (sp:602) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.StoragePool] Unexpected error (sp:618) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,963-0700 INFO (jsonrpc/1) [vdsm.api] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:52) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-31>", line 2, in createStoragePool File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 1027, in createStoragePool leaseParams) File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 619, in create raise se.StorageDomainAccessError(sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,964-0700 INFO (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,964-0700 ERROR (jsonrpc/1) [storage.Dispatcher] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,965-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call StoragePool.create failed (error 379) in 0.01 seconds (__init__:312) 2021-01-16 19:54:01,923-0700 INFO (jsonrpc/5) [api.host] START getStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,964-0700 INFO (jsonrpc/7) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,966-0700 INFO (jsonrpc/7) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) 2021-01-16 19:54:01,977-0700 INFO (jsonrpc/5) [vdsm.api] START repoStats(domains=()) from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:48) 2021-01-16 19:54:01,978-0700 INFO (jsonrpc/5) [vdsm.api] FINISH repoStats return={} from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:54) 2021-01-16 19:54:01,980-0700 INFO (jsonrpc/5) [vdsm.api] START multipath_health() from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:48) 2021-01-16 19:54:01,981-0700 INFO (jsonrpc/5) [vdsm.api] FINISH multipath_health return={} from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:54) 2021-01-16 19:54:01,996-0700 INFO (jsonrpc/5) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) ***end snip***

On Mon, Jan 18, 2021 at 8:58 AM Matt Snow <mattsnow@gmail.com> wrote:
I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the same problem with both versions. The issue occurs in both cockpit UI and tmux'd CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is created and running. I tried to do some debugging on my own before reaching out to this list. Any help is much appreciated!
ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4 cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but I believe it meets the minimum requirements.
NFS server: * Ubuntu 19.10 w/ ZFS share w/ 17TB available space. * NFS share settings are just 'rw=@172.16.1.0/24' but have also tried 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure' * The target directory is always empty and chown'd 36:36 with 0755 permissions. * I have tried using both IP and DNS names. forward and reverse DNS works from ovirt host and other systems on the network. * The NFS share always gets mounted successfully on the ovirt node system. * I have tried auto and v3 NFS versions in other various combinations. * I have also tried setting up an NFS server on a non-ZFS backed storage system that is open to any host and get the same errors as shown below. * I ran nfs-check.py script without issue against both NFS servers and followed other verification steps listed on https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html
***Snip from ovirt-hosted-engine-setup***
Removing snippets of other logs, as I think vdsm should be enough. Thanks for the investigation so far!
The relevantsection from /var/log/vdsm/vdsm.log: ***begin snip*** 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48)
Are you sure this is the first place that deals with this issue? Perhaps search the log earlier, e.g. for 'host_storage' (part of your path). Also, please check supervdsm.log (in the same log directory). Does it manage to mount it? Write anything there? Best regards, -- Didi

Hi Didi, I did log clean up and am re-running ovirt-hosted-engine-cleanup && ovirt-hosted-engine-setup to get you cleaner log files. searching for host_storage in vdsm.log... **snip** 2021-01-18 08:43:18,842-0700 INFO (jsonrpc/3) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,39612 (api:54) 2021-01-18 08:43:19,963-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=fb80c883-2447-4ed2-b344-aa0c0fb65809 (api:48) 2021-01-18 08:43:19,963-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=fb80c883-2447-4ed2-b344-aa0c0fb65809 (api:54) 2021-01-18 08:43:19,964-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:20,441-0700 INFO (jsonrpc/4) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '00000000-0000-0000-0000-000000000000', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=032afa50-381a-44af-a067-d25bcc224355 (api:48) 2021-01-18 08:43:20,446-0700 INFO (jsonrpc/4) [storage.StorageServer.MountConnection] Creating directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage' (storageServer:167) 2021-01-18 08:43:20,446-0700 INFO (jsonrpc/4) [storage.fileUtils] Creating directory: /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage mode: None (fileUtils:201) 2021-01-18 08:43:20,447-0700 INFO (jsonrpc/4) [storage.Mount] mounting stumpy:/tanker/ovirt/host_storage at /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (mount:207) 2021-01-18 08:43:21,271-0700 INFO (jsonrpc/4) [IOProcessClient] (Global) Starting client (__init__:340) 2021-01-18 08:43:21,313-0700 INFO (ioprocess/51124) [IOProcess] (Global) Starting ioprocess (__init__:465) 2021-01-18 08:43:21,373-0700 INFO (jsonrpc/4) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-18 08:43:21,373-0700 INFO (jsonrpc/4) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '00000000-0000-0000-0000-000000000000', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=032afa50-381a-44af-a067-d25bcc224355 (api:54) 2021-01-18 08:43:21,497-0700 INFO (jsonrpc/5) [vdsm.api] START getStorageDomainsList(spUUID='00000000-0000-0000-0000-000000000000', domainClass=1, storageType='', remotePath='stumpy:/tanker/ovirt/host_storage', options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=e37eb000-13da-440f-9197-07495e53ce52 (api:48) 2021-01-18 08:43:21,497-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Refreshing storage domain cache (resize=True) (sdc:80) 2021-01-18 08:43:21,498-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:21,628-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI devices: 0.13 seconds (utils:390) 2021-01-18 08:43:21,629-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:21,908-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC devices: 0.28 seconds (utils:390) 2021-01-18 08:43:21,969-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing multipath devices (multipath:104) 2021-01-18 08:43:21,975-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing multipath devices: 0.01 seconds (utils:390) 2021-01-18 08:43:21,975-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Refreshing storage domain cache: 0.48 seconds (utils:390) 2021-01-18 08:43:22,167-0700 INFO (tmap-0/0) [IOProcessClient] (stumpy:_tanker_ovirt_host__storage) Starting client (__init__:340) 2021-01-18 08:43:22,204-0700 INFO (ioprocess/51144) [IOProcess] (stumpy:_tanker_ovirt_host__storage) Starting ioprocess (__init__:465) 2021-01-18 08:43:22,208-0700 INFO (jsonrpc/5) [vdsm.api] FINISH getStorageDomainsList return={'domlist': []} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=e37eb000-13da-440f-9197-07495e53ce52 (api:54) 2021-01-18 08:43:22,999-0700 INFO (jsonrpc/7) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=51daa36a-e1cf-479d-a93c-1c87f21ce934 (api:48) 2021-01-18 08:43:23,007-0700 INFO (jsonrpc/7) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-18 08:43:23,007-0700 INFO (jsonrpc/7) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=51daa36a-e1cf-479d-a93c-1c87f21ce934 (api:54) 2021-01-18 08:43:23,130-0700 INFO (jsonrpc/0) [vdsm.api] START createStorageDomain(storageType=1, sdUUID='4b3fb9a9-6975-4b80-a2c1-af4e30865088', domainName='hosted_storage', typeSpecificArg='stumpy:/tanker/ovirt/host_storage', domClass=1, domVersion='5', block_size=512, max_hosts=250, options=None) from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:48) 2021-01-18 08:43:23,130-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Refreshing storage domain cache (resize=True) (sdc:80) 2021-01-18 08:43:23,131-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:23,238-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI devices: 0.11 seconds (utils:390) 2021-01-18 08:43:23,239-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:23,522-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC devices: 0.28 seconds (utils:390) 2021-01-18 08:43:23,576-0700 INFO (jsonrpc/0) [storage.Multipath] Resizing multipath devices (multipath:104) 2021-01-18 08:43:23,582-0700 INFO (jsonrpc/0) [storage.Multipath] Resizing multipath devices: 0.01 seconds (utils:390) 2021-01-18 08:43:23,583-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Refreshing storage domain cache: 0.45 seconds (utils:390) 2021-01-18 08:43:23,583-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Looking up domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 (sdc:171) 2021-01-18 08:43:23,760-0700 WARN (jsonrpc/0) [storage.LVM] All 1 tries have failed: cmd=['/sbin/lvm', 'vgs', '--config', 'devices { preferred_names=["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter=["r|.*|"] hints="none" obtain_device_list_from_udev=0 } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { retain_min=50 retain_days=0 }', '--noheadings', '--units', 'b', '--nosuffix', '--separator', '|', '--ignoreskippedcluster', '-o', 'uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name', '4b3fb9a9-6975-4b80-a2c1-af4e30865088'] rc=5 err=[' Volume group "4b3fb9a9-6975-4b80-a2c1-af4e30865088" not found', ' Cannot process volume group 4b3fb9a9-6975-4b80-a2c1-af4e30865088'] (lvm:538) 2021-01-18 08:43:23,771-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Looking up domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088: 0.19 seconds (utils:390) 2021-01-18 08:43:23,818-0700 INFO (jsonrpc/0) [IOProcessClient] (4b3fb9a9-6975-4b80-a2c1-af4e30865088) Starting client (__init__:340) 2021-01-18 08:43:23,856-0700 INFO (ioprocess/51165) [IOProcess] (4b3fb9a9-6975-4b80-a2c1-af4e30865088) Starting ioprocess (__init__:465) 2021-01-18 08:43:24,166-0700 INFO (jsonrpc/0) [storage.StorageDomain] Creating domain metadata directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md' (fileSD:441) 2021-01-18 08:43:24,970-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=5b920619-cd95-40d0-a253-c39d41d71245 (api:48) 2021-01-18 08:43:24,971-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=5b920619-cd95-40d0-a253-c39d41d71245 (api:54) 2021-01-18 08:43:24,971-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:25,285-0700 INFO (jsonrpc/0) [storage.xlease] Formatting index for lockspace '4b3fb9a9-6975-4b80-a2c1-af4e30865088' (version=1) (xlease:701) 2021-01-18 08:43:25,441-0700 INFO (jsonrpc/0) [storage.StorageDomain] Creating domain images directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/images' (nfsSD:110) 2021-01-18 08:43:25,522-0700 INFO (jsonrpc/0) [storage.StorageDomain] Removing remnants of deleted images [] (fileSD:735) 2021-01-18 08:43:25,524-0700 INFO (jsonrpc/0) [storage.SANLock] Initializing sanlock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 path=/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids alignment=1048576 block_size=512 io_timeout=10 (clusterlock:286) 2021-01-18 08:43:25,533-0700 ERROR (jsonrpc/0) [storage.SANLock] Cannot initialize lock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 (clusterlock:305) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 295, in initLock sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such device') 2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [vdsm.api] FINISH createStorageDomain error=Could not initialize cluster lock: () from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:52) 2021-01-18 08:43:25,534-0700 ERROR (jsonrpc/0) [storage.TaskManager.Task] (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 295, in initLock sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such device') During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-123>", line 2, in createStorageDomain File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2669, in createStorageDomain max_hosts=max_hosts) File "/usr/lib/python3.6/site-packages/vdsm/storage/nfsSD.py", line 120, in create fsd.initSPMlease() File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 1019, in initSPMlease return self._manifest.initDomainLock() File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 627, in initDomainLock self._domainLock.initLock(self.getDomainLease()) File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 306, in initLock raise se.ClusterLockInitError() vdsm.storage.exception.ClusterLockInitError: Could not initialize cluster lock: () 2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [storage.TaskManager.Task] (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') aborting: Task is aborted: 'value=Could not initialize cluster lock: () abortedcode=701' (task:1190) 2021-01-18 08:43:25,535-0700 ERROR (jsonrpc/0) [storage.Dispatcher] FINISH createStorageDomain error=Could not initialize cluster lock: () (dispatcher:83) 2021-01-18 08:43:25,536-0700 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call StorageDomain.create failed (error 701) in 2.41 seconds (__init__:312) 2021-01-18 08:43:26,086-0700 INFO (jsonrpc/6) [vdsm.api] START disconnectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=5e35a85a-4010-425e-81b5-5db3ad6d63f3 (api:48) 2021-01-18 08:43:26,087-0700 INFO (jsonrpc/6) [storage.Mount] unmounting /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (mount:215) 2021-01-18 08:43:26,209-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] Refreshing storage domain cache (resize=False) (sdc:80) 2021-01-18 08:43:26,210-0700 INFO (jsonrpc/6) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:26,346-0700 INFO (jsonrpc/6) [storage.ISCSI] Scanning iSCSI devices: 0.14 seconds (utils:390) 2021-01-18 08:43:26,347-0700 INFO (jsonrpc/6) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:26,635-0700 INFO (jsonrpc/6) [storage.HBA] Scanning FC devices: 0.29 seconds (utils:390) 2021-01-18 08:43:26,693-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] Refreshing storage domain cache: 0.48 seconds (utils:390) 2021-01-18 08:43:26,693-0700 INFO (jsonrpc/6) [vdsm.api] FINISH disconnectStorageServer return={'statuslist': [{'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=5e35a85a-4010-425e-81b5-5db3ad6d63f3 (api:54) 2021-01-18 08:43:27,353-0700 INFO (periodic/0) [vdsm.api] START repoStats(domains=()) from=internal, task_id=236f3956-28f6-407e-b893-e018f90fcd08 (api:48) 2021-01-18 08:43:27,355-0700 INFO (periodic/0) [vdsm.api] FINISH repoStats return={} from=internal, task_id=236f3956-28f6-407e-b893-e018f90fcd08 (api:54) 2021-01-18 08:43:27,694-0700 INFO (jsonrpc/1) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,39612 (api:48) 2021-01-18 08:43:27,697-0700 INFO (jsonrpc/1) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,39612 (api:54) 2021-01-18 08:43:27,800-0700 INFO (jsonrpc/2) [api.host] START getAllVmStats() from=::1,55776 (api:48) 2021-01-18 08:43:27,802-0700 INFO (jsonrpc/2) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::1,55776 (api:54) 2021-01-18 08:43:27,826-0700 INFO (jsonrpc/3) [api.host] START getAllVmIoTunePolicies() from=::1,55776 (api:48) 2021-01-18 08:43:27,826-0700 INFO (jsonrpc/3) [api.host] FINISH getAllVmIoTunePolicies return={'status': {'code': 0, 'message': 'Done'}, 'io_tune_policies_dict': {'ddceefdd-57d3-432f-816e-5dc013249bff': {'policy': [], 'current_values': []}}} from=::1,55776 (api:54) 2021-01-18 08:43:29,975-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=821ce12a-ab6e-4db5-ba22-bd2630f2730c (api:48) 2021-01-18 08:43:29,976-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=821ce12a-ab6e-4db5-ba22-bd2630f2730c (api:54) 2021-01-18 08:43:29,976-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:33,826-0700 INFO (jsonrpc/4) [api.host] START getStats() from=::ffff:192.168.222.53,39612 (api:48) **end snip** from supervdsm.log. This is the only section where 'host_storage' shows up. **snip** MainProcess|mpathhealth::DEBUG::2021-01-18 08:42:55,447::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:42:55,448::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,450::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,450::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,483::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,484::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,515::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,516::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:20,448::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call mount with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'stumpy:/tanker/ovirt/host_storage', '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage') {'mntOpts': 'soft,nosharecache,timeo=100,retrans=3,nolock', 'vfstype': 'nfs', 'cgroup': None} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:20,448::commands::211::root::(execCmd) /usr/bin/taskset --cpu-list 0-3 /usr/bin/mount -t nfs -o soft,nosharecache,timeo=100,retrans=3,nolock stumpy:/tanker/ovirt/host_storage /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (cwd None) MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,198::commands::224::root::(execCmd) SUCCESS: <err> = b''; <rc> = 0 MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,198::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return mount with None MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,320::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'vdsm', ('kvm',), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 7) {} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,340::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,341::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'qemu', ('qemu', 'kvm'), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 5) {} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,368::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,630::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>,) {} MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,630::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,907::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,246::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>,) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,246::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,521::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,778::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'vdsm', ('kvm',), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 7) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,796::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,798::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'qemu', ('qemu', 'kvm'), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 5) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,816::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None **end snip** As for the mount area, it does get mounted and unmounted. Files and directory structure is also created. root@stumpy:/tanker/ovirt/host_storage# ls -l total 1 drwxrwsr-x 4 vdsm 36 4 Jan 18 08:43 4b3fb9a9-6975-4b80-a2c1-af4e30865088 root@stumpy:/tanker/ovirt/host_storage# find . . ./4b3fb9a9-6975-4b80-a2c1-af4e30865088 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/images ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/inbox ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/leases ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/metadata ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/xleases ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/outbox root@stumpy:/tanker/ovirt/host_storage# find . -exec ls -ltrh '{}' \; total 512 drwxrwsr-x 4 vdsm 36 4 Jan 18 08:43 4b3fb9a9-6975-4b80-a2c1-af4e30865088 total 12K drwxrwsr-x 2 vdsm 36 8 Jan 18 08:43 dom_md drwxrwsr-x 2 vdsm 36 2 Jan 18 08:43 images total 0 total 275K -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ids -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 inbox -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 outbox -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 leases -rw-rw---- 1 vdsm 36 1.3M Jan 18 08:43 xleases -rw-rw-r-- 1 vdsm 36 342 Jan 18 08:43 metadata -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/inbox -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/leases -rw-rw-r-- 1 vdsm 36 342 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/metadata -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids -rw-rw---- 1 vdsm 36 1.3M Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/xleases -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/outbox root@stumpy:/tanker/ovirt/host_storage# Thank you, ..Matt

In case its useful here is the mount occurring. [root@brick setup_debugging_logs]# while :; do mount | grep stumpy ; sleep 1; done stumpy:/tanker/ovirt/host_storage on /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50) stumpy:/tanker/ovirt/host_storage on /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50) stumpy:/tanker/ovirt/host_storage on /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50) ^C [root@brick setup_debugging_logs]#

Complete logs can be found here: vdsm.log - https://paste.c-net.org/DownloadPressure supervdsm.log - https://paste.c-net.org/LaterScandals

On Mon, Jan 18, 2021 at 6:01 PM Matt Snow <mattsnow@gmail.com> wrote:
Hi Didi, I did log clean up and am re-running ovirt-hosted-engine-cleanup && ovirt-hosted-engine-setup to get you cleaner log files.
searching for host_storage in vdsm.log... **snip** 2021-01-18 08:43:18,842-0700 INFO (jsonrpc/3) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,39612 (api:54) 2021-01-18 08:43:19,963-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=fb80c883-2447-4ed2-b344-aa0c0fb65809 (api:48) 2021-01-18 08:43:19,963-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=fb80c883-2447-4ed2-b344-aa0c0fb65809 (api:54) 2021-01-18 08:43:19,964-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:20,441-0700 INFO (jsonrpc/4) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '00000000-0000-0000-0000-000000000000', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=032afa50-381a-44af-a067-d25bcc224355 (api:48) 2021-01-18 08:43:20,446-0700 INFO (jsonrpc/4) [storage.StorageServer.MountConnection] Creating directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage' (storageServer:167) 2021-01-18 08:43:20,446-0700 INFO (jsonrpc/4) [storage.fileUtils] Creating directory: /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage mode: None (fileUtils:201) 2021-01-18 08:43:20,447-0700 INFO (jsonrpc/4) [storage.Mount] mounting stumpy:/tanker/ovirt/host_storage at /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (mount:207) 2021-01-18 08:43:21,271-0700 INFO (jsonrpc/4) [IOProcessClient] (Global) Starting client (__init__:340) 2021-01-18 08:43:21,313-0700 INFO (ioprocess/51124) [IOProcess] (Global) Starting ioprocess (__init__:465) 2021-01-18 08:43:21,373-0700 INFO (jsonrpc/4) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-18 08:43:21,373-0700 INFO (jsonrpc/4) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '00000000-0000-0000-0000-000000000000', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=032afa50-381a-44af-a067-d25bcc224355 (api:54) 2021-01-18 08:43:21,497-0700 INFO (jsonrpc/5) [vdsm.api] START getStorageDomainsList(spUUID='00000000-0000-0000-0000-000000000000', domainClass=1, storageType='', remotePath='stumpy:/tanker/ovirt/host_storage', options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=e37eb000-13da-440f-9197-07495e53ce52 (api:48) 2021-01-18 08:43:21,497-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Refreshing storage domain cache (resize=True) (sdc:80) 2021-01-18 08:43:21,498-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:21,628-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI devices: 0.13 seconds (utils:390) 2021-01-18 08:43:21,629-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:21,908-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC devices: 0.28 seconds (utils:390) 2021-01-18 08:43:21,969-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing multipath devices (multipath:104) 2021-01-18 08:43:21,975-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing multipath devices: 0.01 seconds (utils:390) 2021-01-18 08:43:21,975-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Refreshing storage domain cache: 0.48 seconds (utils:390) 2021-01-18 08:43:22,167-0700 INFO (tmap-0/0) [IOProcessClient] (stumpy:_tanker_ovirt_host__storage) Starting client (__init__:340) 2021-01-18 08:43:22,204-0700 INFO (ioprocess/51144) [IOProcess] (stumpy:_tanker_ovirt_host__storage) Starting ioprocess (__init__:465) 2021-01-18 08:43:22,208-0700 INFO (jsonrpc/5) [vdsm.api] FINISH getStorageDomainsList return={'domlist': []} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=e37eb000-13da-440f-9197-07495e53ce52 (api:54) 2021-01-18 08:43:22,999-0700 INFO (jsonrpc/7) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=51daa36a-e1cf-479d-a93c-1c87f21ce934 (api:48) 2021-01-18 08:43:23,007-0700 INFO (jsonrpc/7) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-18 08:43:23,007-0700 INFO (jsonrpc/7) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=51daa36a-e1cf-479d-a93c-1c87f21ce934 (api:54) 2021-01-18 08:43:23,130-0700 INFO (jsonrpc/0) [vdsm.api] START createStorageDomain(storageType=1, sdUUID='4b3fb9a9-6975-4b80-a2c1-af4e30865088', domainName='hosted_storage', typeSpecificArg='stumpy:/tanker/ovirt/host_storage', domClass=1, domVersion='5', block_size=512, max_hosts=250, options=None) from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:48) 2021-01-18 08:43:23,130-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Refreshing storage domain cache (resize=True) (sdc:80) 2021-01-18 08:43:23,131-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:23,238-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI devices: 0.11 seconds (utils:390) 2021-01-18 08:43:23,239-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:23,522-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC devices: 0.28 seconds (utils:390) 2021-01-18 08:43:23,576-0700 INFO (jsonrpc/0) [storage.Multipath] Resizing multipath devices (multipath:104) 2021-01-18 08:43:23,582-0700 INFO (jsonrpc/0) [storage.Multipath] Resizing multipath devices: 0.01 seconds (utils:390) 2021-01-18 08:43:23,583-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Refreshing storage domain cache: 0.45 seconds (utils:390) 2021-01-18 08:43:23,583-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Looking up domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 (sdc:171) 2021-01-18 08:43:23,760-0700 WARN (jsonrpc/0) [storage.LVM] All 1 tries have failed: cmd=['/sbin/lvm', 'vgs', '--config', 'devices { preferred_names=["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter=["r|.*|"] hints="none" obtain_device_list_from_udev=0 } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { retain_min=50 retain_days=0 }', '--noheadings', '--units', 'b', '--nosuffix', '--separator', '|', '--ignoreskippedcluster', '-o', 'uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name', '4b3fb9a9-6975-4b80-a2c1-af4e30865088'] rc=5 err=[' Volume group "4b3fb9a9-6975-4b80-a2c1-af4e30865088" not found', ' Cannot process volume group 4b3fb9a9-6975-4b80-a2c1-af4e30865088'] (lvm:538) 2021-01-18 08:43:23,771-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Looking up domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088: 0.19 seconds (utils:390) 2021-01-18 08:43:23,818-0700 INFO (jsonrpc/0) [IOProcessClient] (4b3fb9a9-6975-4b80-a2c1-af4e30865088) Starting client (__init__:340) 2021-01-18 08:43:23,856-0700 INFO (ioprocess/51165) [IOProcess] (4b3fb9a9-6975-4b80-a2c1-af4e30865088) Starting ioprocess (__init__:465) 2021-01-18 08:43:24,166-0700 INFO (jsonrpc/0) [storage.StorageDomain] Creating domain metadata directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md' (fileSD:441) 2021-01-18 08:43:24,970-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=5b920619-cd95-40d0-a253-c39d41d71245 (api:48) 2021-01-18 08:43:24,971-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=5b920619-cd95-40d0-a253-c39d41d71245 (api:54) 2021-01-18 08:43:24,971-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:25,285-0700 INFO (jsonrpc/0) [storage.xlease] Formatting index for lockspace '4b3fb9a9-6975-4b80-a2c1-af4e30865088' (version=1) (xlease:701) 2021-01-18 08:43:25,441-0700 INFO (jsonrpc/0) [storage.StorageDomain] Creating domain images directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/images' (nfsSD:110) 2021-01-18 08:43:25,522-0700 INFO (jsonrpc/0) [storage.StorageDomain] Removing remnants of deleted images [] (fileSD:735) 2021-01-18 08:43:25,524-0700 INFO (jsonrpc/0) [storage.SANLock] Initializing sanlock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 path=/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids alignment=1048576 block_size=512 io_timeout=10 (clusterlock:286) 2021-01-18 08:43:25,533-0700 ERROR (jsonrpc/0) [storage.SANLock] Cannot initialize lock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 (clusterlock:305) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 295, in initLock sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such device') 2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [vdsm.api] FINISH createStorageDomain error=Could not initialize cluster lock: () from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:52) 2021-01-18 08:43:25,534-0700 ERROR (jsonrpc/0) [storage.TaskManager.Task] (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 295, in initLock sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such device')
I think this ^^ is the issue. Can you please check /var/log/sanlock.log? Adding Nir. Best regards,
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-123>", line 2, in createStorageDomain File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2669, in createStorageDomain max_hosts=max_hosts) File "/usr/lib/python3.6/site-packages/vdsm/storage/nfsSD.py", line 120, in create fsd.initSPMlease() File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 1019, in initSPMlease return self._manifest.initDomainLock() File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 627, in initDomainLock self._domainLock.initLock(self.getDomainLease()) File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 306, in initLock raise se.ClusterLockInitError() vdsm.storage.exception.ClusterLockInitError: Could not initialize cluster lock: () 2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [storage.TaskManager.Task] (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') aborting: Task is aborted: 'value=Could not initialize cluster lock: () abortedcode=701' (task:1190) 2021-01-18 08:43:25,535-0700 ERROR (jsonrpc/0) [storage.Dispatcher] FINISH createStorageDomain error=Could not initialize cluster lock: () (dispatcher:83) 2021-01-18 08:43:25,536-0700 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call StorageDomain.create failed (error 701) in 2.41 seconds (__init__:312) 2021-01-18 08:43:26,086-0700 INFO (jsonrpc/6) [vdsm.api] START disconnectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=5e35a85a-4010-425e-81b5-5db3ad6d63f3 (api:48) 2021-01-18 08:43:26,087-0700 INFO (jsonrpc/6) [storage.Mount] unmounting /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (mount:215) 2021-01-18 08:43:26,209-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] Refreshing storage domain cache (resize=False) (sdc:80) 2021-01-18 08:43:26,210-0700 INFO (jsonrpc/6) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:26,346-0700 INFO (jsonrpc/6) [storage.ISCSI] Scanning iSCSI devices: 0.14 seconds (utils:390) 2021-01-18 08:43:26,347-0700 INFO (jsonrpc/6) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:26,635-0700 INFO (jsonrpc/6) [storage.HBA] Scanning FC devices: 0.29 seconds (utils:390) 2021-01-18 08:43:26,693-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] Refreshing storage domain cache: 0.48 seconds (utils:390) 2021-01-18 08:43:26,693-0700 INFO (jsonrpc/6) [vdsm.api] FINISH disconnectStorageServer return={'statuslist': [{'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=5e35a85a-4010-425e-81b5-5db3ad6d63f3 (api:54) 2021-01-18 08:43:27,353-0700 INFO (periodic/0) [vdsm.api] START repoStats(domains=()) from=internal, task_id=236f3956-28f6-407e-b893-e018f90fcd08 (api:48) 2021-01-18 08:43:27,355-0700 INFO (periodic/0) [vdsm.api] FINISH repoStats return={} from=internal, task_id=236f3956-28f6-407e-b893-e018f90fcd08 (api:54) 2021-01-18 08:43:27,694-0700 INFO (jsonrpc/1) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,39612 (api:48) 2021-01-18 08:43:27,697-0700 INFO (jsonrpc/1) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,39612 (api:54) 2021-01-18 08:43:27,800-0700 INFO (jsonrpc/2) [api.host] START getAllVmStats() from=::1,55776 (api:48) 2021-01-18 08:43:27,802-0700 INFO (jsonrpc/2) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::1,55776 (api:54) 2021-01-18 08:43:27,826-0700 INFO (jsonrpc/3) [api.host] START getAllVmIoTunePolicies() from=::1,55776 (api:48) 2021-01-18 08:43:27,826-0700 INFO (jsonrpc/3) [api.host] FINISH getAllVmIoTunePolicies return={'status': {'code': 0, 'message': 'Done'}, 'io_tune_policies_dict': {'ddceefdd-57d3-432f-816e-5dc013249bff': {'policy': [], 'current_values': []}}} from=::1,55776 (api:54) 2021-01-18 08:43:29,975-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=821ce12a-ab6e-4db5-ba22-bd2630f2730c (api:48) 2021-01-18 08:43:29,976-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=821ce12a-ab6e-4db5-ba22-bd2630f2730c (api:54) 2021-01-18 08:43:29,976-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:33,826-0700 INFO (jsonrpc/4) [api.host] START getStats() from=::ffff:192.168.222.53,39612 (api:48)
**end snip**
from supervdsm.log. This is the only section where 'host_storage' shows up.
**snip** MainProcess|mpathhealth::DEBUG::2021-01-18 08:42:55,447::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:42:55,448::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,450::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,450::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,483::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,484::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,515::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,516::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:20,448::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call mount with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'stumpy:/tanker/ovirt/host_storage', '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage') {'mntOpts': 'soft,nosharecache,timeo=100,retrans=3,nolock', 'vfstype': 'nfs', 'cgroup': None} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:20,448::commands::211::root::(execCmd) /usr/bin/taskset --cpu-list 0-3 /usr/bin/mount -t nfs -o soft,nosharecache,timeo=100,retrans=3,nolock stumpy:/tanker/ovirt/host_storage /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (cwd None) MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,198::commands::224::root::(execCmd) SUCCESS: <err> = b''; <rc> = 0 MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,198::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return mount with None MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,320::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'vdsm', ('kvm',), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 7) {} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,340::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,341::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'qemu', ('qemu', 'kvm'), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 5) {} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,368::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,630::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>,) {} MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,630::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,907::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,246::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>,) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,246::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,521::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,778::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'vdsm', ('kvm',), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 7) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,796::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,798::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'qemu', ('qemu', 'kvm'), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 5) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,816::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None **end snip**
As for the mount area, it does get mounted and unmounted. Files and directory structure is also created.
root@stumpy:/tanker/ovirt/host_storage# ls -l total 1 drwxrwsr-x 4 vdsm 36 4 Jan 18 08:43 4b3fb9a9-6975-4b80-a2c1-af4e30865088 root@stumpy:/tanker/ovirt/host_storage# find . . ./4b3fb9a9-6975-4b80-a2c1-af4e30865088 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/images ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/inbox ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/leases ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/metadata ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/xleases ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/outbox root@stumpy:/tanker/ovirt/host_storage# find . -exec ls -ltrh '{}' \; total 512 drwxrwsr-x 4 vdsm 36 4 Jan 18 08:43 4b3fb9a9-6975-4b80-a2c1-af4e30865088 total 12K drwxrwsr-x 2 vdsm 36 8 Jan 18 08:43 dom_md drwxrwsr-x 2 vdsm 36 2 Jan 18 08:43 images total 0 total 275K -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ids -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 inbox -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 outbox -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 leases -rw-rw---- 1 vdsm 36 1.3M Jan 18 08:43 xleases -rw-rw-r-- 1 vdsm 36 342 Jan 18 08:43 metadata -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/inbox -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/leases -rw-rw-r-- 1 vdsm 36 342 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/metadata -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids -rw-rw---- 1 vdsm 36 1.3M Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/xleases -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/outbox root@stumpy:/tanker/ovirt/host_storage#
Thank you,
..Matt _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/BORUNFGGNC6TKT...
-- Didi

[root@brick log]# cat sanlock.log 2021-01-15 18:18:48 3974 [36280]: sanlock daemon started 3.8.2 host 3b903780-4f79-1018-816e-aeb2724778a7 (brick.co.slakin.net) 2021-01-15 19:17:31 7497 [36293]: s1 lockspace 54532dd4-3e5b-4885-b88e-599c81efb146:250:/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids:0 2021-01-15 19:17:31 7497 [50873]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids 2021-01-15 19:17:31 7497 [50873]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-15 19:17:31 7497 [50873]: s1 open_disk /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids error -13 2021-01-15 19:17:32 7498 [36293]: s1 add_lockspace fail result -19 2021-01-18 07:23:58 18 [1318]: sanlock daemon started 3.8.2 host 3b903780-4f79-1018-816e-aeb2724778a7 (brick.co.slakin.net) 2021-01-18 08:43:25 4786 [1359]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids 2021-01-18 08:43:25 4786 [1359]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-18 09:13:17 6578 [1358]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/d3aec1fd-57cb-4d48-86b9-0a89ae3741a7/dom_md/ids 2021-01-18 09:13:17 6578 [1358]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-18 09:19:45 6966 [1359]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/fa5434cf-3e05-45d5-b32e-4948903ee2b4/dom_md/ids 2021-01-18 09:19:45 6966 [1359]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-18 09:21:16 7057 [1358]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/b0f7b773-7e37-4b6b-a467-64230d5f7391/dom_md/ids 2021-01-18 09:21:16 7057 [1358]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-18 09:49:42 8763 [1359]: s1 lockspace b0f7b773-7e37-4b6b-a467-64230d5f7391:250:/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/b0f7b773-7e37-4b6b-a467-64230d5f7391/dom_md/ids:0 2021-01-18 09:49:42 8763 [54250]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/b0f7b773-7e37-4b6b-a467-64230d5f7391/dom_md/ids 2021-01-18 09:49:42 8763 [54250]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-18 09:49:42 8763 [54250]: s1 open_disk /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/b0f7b773-7e37-4b6b-a467-64230d5f7391/dom_md/ids error -13 2021-01-18 09:49:43 8764 [1359]: s1 add_lockspace fail result -19 [root@brick log]# [root@brick log]# su - sanlock -s /bin/bash Last login: Tue Jan 19 09:05:25 MST 2021 on pts/2 nodectl must be run as root! nodectl must be run as root! [sanlock@brick ~]$ grep sanlock /etc/group disk:x:6:sanlock kvm:x:36:qemu,ovirtimg,sanlock sanlock:x:179:vdsm qemu:x:107:vdsm,ovirtimg,sanlock [sanlock@brick ~]$ cd /rhev/data-center/mnt/stumpy\:_tanker_ovirt_host__storage/ && touch file.txt && ls -l file.txt -rw-rw-rw-. 1 sanlock sanlock 0 Jan 19 09:07 file.txt [sanlock@brick stumpy:_tanker_ovirt_host__storage]$ [root@brick log]# su - sanlock -s /bin/bash Last login: Tue Jan 19 09:05:25 MST 2021 on pts/2 nodectl must be run as root! nodectl must be run as root! [sanlock@brick ~]$ grep sanlock /etc/group disk:x:6:sanlock kvm:x:36:qemu,ovirtimg,sanlock sanlock:x:179:vdsm qemu:x:107:vdsm,ovirtimg,sanlock [sanlock@brick ~]$ cd /rhev/data-center/mnt/stumpy\:_tanker_ovirt_host__storage/ && touch file.txt && ls -l file.txt -rw-rw-rw-. 1 sanlock sanlock 0 Jan 19 09:07 file.txt [sanlock@brick stumpy:_tanker_ovirt_host__storage]$ [sanlock@brick stumpy:_tanker_ovirt_host__storage]$ ls -ltra total 2 drwxr-xr-x. 3 vdsm kvm 48 Jan 18 09:48 .. drwxrwxrwx. 2 vdsm kvm 3 Jan 19 09:07 . -rw-rw-rw-. 1 sanlock sanlock 0 Jan 19 09:07 file.txt [sanlock@brick stumpy:_tanker_ovirt_host__storage]$ On Tue, Jan 19, 2021 at 6:44 AM Yedidyah Bar David <didi@redhat.com> wrote:
On Mon, Jan 18, 2021 at 6:01 PM Matt Snow <mattsnow@gmail.com> wrote:
Hi Didi, I did log clean up and am re-running ovirt-hosted-engine-cleanup &&
ovirt-hosted-engine-setup to get you cleaner log files.
searching for host_storage in vdsm.log... **snip** 2021-01-18 08:43:18,842-0700 INFO (jsonrpc/3) [api.host] FINISH
2021-01-18 08:43:19,963-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=fb80c883-2447-4ed2-b344-aa0c0fb65809 (api:48) 2021-01-18 08:43:19,963-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=fb80c883-2447-4ed2-b344-aa0c0fb65809 (api:54) 2021-01-18 08:43:19,964-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:20,441-0700 INFO (jsonrpc/4) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '00000000-0000-0000-0000-000000000000', 'user': '', 'tpgt': '1'}],
2021-01-18 08:43:20,446-0700 INFO (jsonrpc/4) [storage.StorageServer.MountConnection] Creating directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage' (storageServer:167) 2021-01-18 08:43:20,446-0700 INFO (jsonrpc/4) [storage.fileUtils] Creating directory: /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage mode: None (fileUtils:201) 2021-01-18 08:43:20,447-0700 INFO (jsonrpc/4) [storage.Mount] mounting stumpy:/tanker/ovirt/host_storage at /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (mount:207) 2021-01-18 08:43:21,271-0700 INFO (jsonrpc/4) [IOProcessClient] (Global) Starting client (__init__:340) 2021-01-18 08:43:21,313-0700 INFO (ioprocess/51124) [IOProcess] (Global) Starting ioprocess (__init__:465) 2021-01-18 08:43:21,373-0700 INFO (jsonrpc/4) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-18 08:43:21,373-0700 INFO (jsonrpc/4) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '00000000-0000-0000-0000-000000000000', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=032afa50-381a-44af-a067-d25bcc224355 (api:54) 2021-01-18 08:43:21,497-0700 INFO (jsonrpc/5) [vdsm.api] START getStorageDomainsList(spUUID='00000000-0000-0000-0000-000000000000', domainClass=1, storageType='', remotePath='stumpy:/tanker/ovirt/host_storage', options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=e37eb000-13da-440f-9197-07495e53ce52 (api:48) 2021-01-18 08:43:21,497-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Refreshing storage domain cache (resize=True) (sdc:80) 2021-01-18 08:43:21,498-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:21,628-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI devices: 0.13 seconds (utils:390) 2021-01-18 08:43:21,629-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:21,908-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC devices: 0.28 seconds (utils:390) 2021-01-18 08:43:21,969-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing multipath devices (multipath:104) 2021-01-18 08:43:21,975-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing multipath devices: 0.01 seconds (utils:390) 2021-01-18 08:43:21,975-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Refreshing storage domain cache: 0.48 seconds (utils:390) 2021-01-18 08:43:22,167-0700 INFO (tmap-0/0) [IOProcessClient] (stumpy:_tanker_ovirt_host__storage) Starting client (__init__:340) 2021-01-18 08:43:22,204-0700 INFO (ioprocess/51144) [IOProcess] (stumpy:_tanker_ovirt_host__storage) Starting ioprocess (__init__:465) 2021-01-18 08:43:22,208-0700 INFO (jsonrpc/5) [vdsm.api] FINISH getStorageDomainsList return={'domlist': []} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=e37eb000-13da-440f-9197-07495e53ce52 (api:54) 2021-01-18 08:43:22,999-0700 INFO (jsonrpc/7) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'user': '', 'tpgt': '1'}],
2021-01-18 08:43:23,007-0700 INFO (jsonrpc/7) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-18 08:43:23,007-0700 INFO (jsonrpc/7) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=51daa36a-e1cf-479d-a93c-1c87f21ce934 (api:54) 2021-01-18 08:43:23,130-0700 INFO (jsonrpc/0) [vdsm.api] START createStorageDomain(storageType=1, sdUUID='4b3fb9a9-6975-4b80-a2c1-af4e30865088', domainName='hosted_storage', typeSpecificArg='stumpy:/tanker/ovirt/host_storage', domClass=1, domVersion='5', block_size=512, max_hosts=250, options=None) from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:48) 2021-01-18 08:43:23,130-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Refreshing storage domain cache (resize=True) (sdc:80) 2021-01-18 08:43:23,131-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:23,238-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI devices: 0.11 seconds (utils:390) 2021-01-18 08:43:23,239-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:23,522-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC devices: 0.28 seconds (utils:390) 2021-01-18 08:43:23,576-0700 INFO (jsonrpc/0) [storage.Multipath] Resizing multipath devices (multipath:104) 2021-01-18 08:43:23,582-0700 INFO (jsonrpc/0) [storage.Multipath] Resizing multipath devices: 0.01 seconds (utils:390) 2021-01-18 08:43:23,583-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Refreshing storage domain cache: 0.45 seconds (utils:390) 2021-01-18 08:43:23,583-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Looking up domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 (sdc:171) 2021-01-18 08:43:23,760-0700 WARN (jsonrpc/0) [storage.LVM] All 1 tries have failed: cmd=['/sbin/lvm', 'vgs', '--config', 'devices {
2021-01-18 08:43:23,771-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Looking up domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088: 0.19 seconds (utils:390) 2021-01-18 08:43:23,818-0700 INFO (jsonrpc/0) [IOProcessClient] (4b3fb9a9-6975-4b80-a2c1-af4e30865088) Starting client (__init__:340) 2021-01-18 08:43:23,856-0700 INFO (ioprocess/51165) [IOProcess] (4b3fb9a9-6975-4b80-a2c1-af4e30865088) Starting ioprocess (__init__:465) 2021-01-18 08:43:24,166-0700 INFO (jsonrpc/0) [storage.StorageDomain] Creating domain metadata directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md' (fileSD:441) 2021-01-18 08:43:24,970-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=5b920619-cd95-40d0-a253-c39d41d71245 (api:48) 2021-01-18 08:43:24,971-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=5b920619-cd95-40d0-a253-c39d41d71245 (api:54) 2021-01-18 08:43:24,971-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:25,285-0700 INFO (jsonrpc/0) [storage.xlease] Formatting index for lockspace '4b3fb9a9-6975-4b80-a2c1-af4e30865088' (version=1) (xlease:701) 2021-01-18 08:43:25,441-0700 INFO (jsonrpc/0) [storage.StorageDomain] Creating domain images directory '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/images' (nfsSD:110) 2021-01-18 08:43:25,522-0700 INFO (jsonrpc/0) [storage.StorageDomain] Removing remnants of deleted images [] (fileSD:735) 2021-01-18 08:43:25,524-0700 INFO (jsonrpc/0) [storage.SANLock] Initializing sanlock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088
2021-01-18 08:43:25,533-0700 ERROR (jsonrpc/0) [storage.SANLock] Cannot initialize lock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 (clusterlock:305) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py",
getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,39612 (api:54) options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=032afa50-381a-44af-a067-d25bcc224355 (api:48) options=None) from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=51daa36a-e1cf-479d-a93c-1c87f21ce934 (api:48) preferred_names=["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter=["r|.*|"] hints="none" obtain_device_list_from_udev=0 } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { retain_min=50 retain_days=0 }', '--noheadings', '--units', 'b', '--nosuffix', '--separator', '|', '--ignoreskippedcluster', '-o', 'uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name', '4b3fb9a9-6975-4b80-a2c1-af4e30865088'] rc=5 err=[' Volume group "4b3fb9a9-6975-4b80-a2c1-af4e30865088" not found', ' Cannot process volume group 4b3fb9a9-6975-4b80-a2c1-af4e30865088'] (lvm:538) path=/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids alignment=1048576 block_size=512 io_timeout=10 (clusterlock:286) line 295, in initLock
sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No
2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [vdsm.api] FINISH createStorageDomain error=Could not initialize cluster lock: () from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:52) 2021-01-18 08:43:25,534-0700 ERROR (jsonrpc/0) [storage.TaskManager.Task] (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py",
such device') line 295, in initLock
sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No
such device')
I think this ^^ is the issue. Can you please check /var/log/sanlock.log?
Adding Nir.
Best regards,
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line
887, in _run
return fn(*args, **kargs) File "<decorator-gen-123>", line 2, in createStorageDomain File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50,
in method
ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line
2669, in createStorageDomain
max_hosts=max_hosts) File "/usr/lib/python3.6/site-packages/vdsm/storage/nfsSD.py", line
120, in create
fsd.initSPMlease() File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 1019,
in initSPMlease
return self._manifest.initDomainLock() File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 627,
in initDomainLock
self._domainLock.initLock(self.getDomainLease()) File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py",
line 306, in initLock
raise se.ClusterLockInitError() vdsm.storage.exception.ClusterLockInitError: Could not initialize
2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [storage.TaskManager.Task] (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') aborting: Task is aborted: 'value=Could not initialize cluster lock: () abortedcode=701' (task:1190) 2021-01-18 08:43:25,535-0700 ERROR (jsonrpc/0) [storage.Dispatcher] FINISH createStorageDomain error=Could not initialize cluster lock: () (dispatcher:83) 2021-01-18 08:43:25,536-0700 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call StorageDomain.create failed (error 701) in 2.41 seconds (__init__:312) 2021-01-18 08:43:26,086-0700 INFO (jsonrpc/6) [vdsm.api] START disconnectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'user': '', 'tpgt': '1'}],
2021-01-18 08:43:26,087-0700 INFO (jsonrpc/6) [storage.Mount] unmounting /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (mount:215) 2021-01-18 08:43:26,209-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] Refreshing storage domain cache (resize=False) (sdc:80) 2021-01-18 08:43:26,210-0700 INFO (jsonrpc/6) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-18 08:43:26,346-0700 INFO (jsonrpc/6) [storage.ISCSI] Scanning iSCSI devices: 0.14 seconds (utils:390) 2021-01-18 08:43:26,347-0700 INFO (jsonrpc/6) [storage.HBA] Scanning FC devices (hba:60) 2021-01-18 08:43:26,635-0700 INFO (jsonrpc/6) [storage.HBA] Scanning FC devices: 0.29 seconds (utils:390) 2021-01-18 08:43:26,693-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] Refreshing storage domain cache: 0.48 seconds (utils:390) 2021-01-18 08:43:26,693-0700 INFO (jsonrpc/6) [vdsm.api] FINISH disconnectStorageServer return={'statuslist': [{'id': 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'status': 0}]} from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=5e35a85a-4010-425e-81b5-5db3ad6d63f3 (api:54) 2021-01-18 08:43:27,353-0700 INFO (periodic/0) [vdsm.api] START repoStats(domains=()) from=internal, task_id=236f3956-28f6-407e-b893-e018f90fcd08 (api:48) 2021-01-18 08:43:27,355-0700 INFO (periodic/0) [vdsm.api] FINISH repoStats return={} from=internal, task_id=236f3956-28f6-407e-b893-e018f90fcd08 (api:54) 2021-01-18 08:43:27,694-0700 INFO (jsonrpc/1) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,39612 (api:48) 2021-01-18 08:43:27,697-0700 INFO (jsonrpc/1) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,39612 (api:54) 2021-01-18 08:43:27,800-0700 INFO (jsonrpc/2) [api.host] START getAllVmStats() from=::1,55776 (api:48) 2021-01-18 08:43:27,802-0700 INFO (jsonrpc/2) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::1,55776 (api:54) 2021-01-18 08:43:27,826-0700 INFO (jsonrpc/3) [api.host] START getAllVmIoTunePolicies() from=::1,55776 (api:48) 2021-01-18 08:43:27,826-0700 INFO (jsonrpc/3) [api.host] FINISH getAllVmIoTunePolicies return={'status': {'code': 0, 'message': 'Done'}, 'io_tune_policies_dict': {'ddceefdd-57d3-432f-816e-5dc013249bff': {'policy': [], 'current_values': []}}} from=::1,55776 (api:54) 2021-01-18 08:43:29,975-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=821ce12a-ab6e-4db5-ba22-bd2630f2730c (api:48) 2021-01-18 08:43:29,976-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=821ce12a-ab6e-4db5-ba22-bd2630f2730c (api:54) 2021-01-18 08:43:29,976-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-18 08:43:33,826-0700 INFO (jsonrpc/4) [api.host] START getStats() from=::ffff:192.168.222.53,39612 (api:48)
**end snip**
from supervdsm.log. This is the only section where 'host_storage' shows up.
**snip** MainProcess|mpathhealth::DEBUG::2021-01-18 08:42:55,447::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:42:55,448::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,450::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,450::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,483::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:05,484::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::logutils::354::root::(_report_stats) ThreadedHandler is ok in
MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,485::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,515::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-18 08:43:15,516::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:20,448::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call mount with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'stumpy:/tanker/ovirt/host_storage', '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage') {'mntOpts': 'soft,nosharecache,timeo=100,retrans=3,nolock', 'vfstype': 'nfs', 'cgroup': None} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:20,448::commands::211::root::(execCmd) /usr/bin/taskset --cpu-list 0-3 /usr/bin/mount -t nfs -o soft,nosharecache,timeo=100,retrans=3,nolock stumpy:/tanker/ovirt/host_storage /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (cwd None) MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,198::commands::224::root::(execCmd) SUCCESS: <err> = b''; <rc> = 0 MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,198::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return mount with None MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,320::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'vdsm', ('kvm',), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 7) {} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,340::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,341::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'qemu', ('qemu', 'kvm'), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 5) {} MainProcess|jsonrpc/4::DEBUG::2021-01-18 08:43:21,368::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,630::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>,) {} MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,630::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) MainProcess|jsonrpc/5::DEBUG::2021-01-18 08:43:21,907::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,246::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>,) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,246::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,521::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,778::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'vdsm', ('kvm',), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 7) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,796::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,798::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, 'qemu', ('qemu', 'kvm'), '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 5) {} MainProcess|jsonrpc/0::DEBUG::2021-01-18 08:43:23,816::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None **end snip**
As for the mount area, it does get mounted and unmounted. Files and
cluster lock: () options=None) from=::ffff:192.168.222.53,39612, flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, task_id=5e35a85a-4010-425e-81b5-5db3ad6d63f3 (api:48) the last 60 seconds (max pending: 2) directory structure is also created.
root@stumpy:/tanker/ovirt/host_storage# ls -l total 1 drwxrwsr-x 4 vdsm 36 4 Jan 18 08:43 4b3fb9a9-6975-4b80-a2c1-af4e30865088 root@stumpy:/tanker/ovirt/host_storage# find . . ./4b3fb9a9-6975-4b80-a2c1-af4e30865088 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/images ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/inbox ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/leases ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/metadata ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/xleases ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/outbox root@stumpy:/tanker/ovirt/host_storage# find . -exec ls -ltrh '{}' \; total 512 drwxrwsr-x 4 vdsm 36 4 Jan 18 08:43 4b3fb9a9-6975-4b80-a2c1-af4e30865088 total 12K drwxrwsr-x 2 vdsm 36 8 Jan 18 08:43 dom_md drwxrwsr-x 2 vdsm 36 2 Jan 18 08:43 images total 0 total 275K -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ids -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 inbox -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 outbox -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 leases -rw-rw---- 1 vdsm 36 1.3M Jan 18 08:43 xleases -rw-rw-r-- 1 vdsm 36 342 Jan 18 08:43 metadata -rw-rw---- 1 vdsm 36 16M Jan 18 08:43
./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/inbox
-rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/leases -rw-rw-r-- 1 vdsm 36 342 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/metadata -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids -rw-rw---- 1 vdsm 36 1.3M Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/xleases -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/outbox root@stumpy:/tanker/ovirt/host_storage#
Thank you,
..Matt _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/BORUNFGGNC6TKT...
-- Didi

On Tue, Jan 19, 2021 at 6:13 PM Matt Snow <mattsnow@gmail.com> wrote:
[root@brick log]# cat sanlock.log
2021-01-15 18:18:48 3974 [36280]: sanlock daemon started 3.8.2 host 3b903780-4f79-1018-816e-aeb2724778a7 (brick.co.slakin.net) 2021-01-15 19:17:31 7497 [36293]: s1 lockspace 54532dd4-3e5b-4885-b88e-599c81efb146:250:/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids:0 2021-01-15 19:17:31 7497 [50873]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids
Smells like selinux issue. What do you see in "ausearch -m avc"?

[root@brick ~]# ps -efz | grep sanlock error: unsupported SysV option Usage: ps [options] Try 'ps --help <simple|list|output|threads|misc|all>' or 'ps --help <s|l|o|t|m|a>' for additional help text. For more details see ps(1). [root@brick ~]# ps -ef | grep sanlock sanlock 1308 1 0 10:21 ? 00:00:01 /usr/sbin/sanlock daemon root 1309 1308 0 10:21 ? 00:00:00 /usr/sbin/sanlock daemon root 68086 67674 0 13:38 pts/4 00:00:00 tail -f sanlock.log root 73724 68214 0 18:49 pts/5 00:00:00 grep --color=auto sanlock [root@brick ~]# ausearch -m avc <no matches> [root@brick ~]# ls -lhZ /rhev/data-center/mnt/stumpy\:_tanker_ovirt_host__storage/8fd5420f-61fd-41af-8575-f61853a18d91/dom_md total 278K -rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 0 Jan 19 13:38 ids -rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 16M Jan 19 13:38 inbox -rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 0 Jan 19 13:38 leases -rw-rw-r--. 1 vdsm kvm system_u:object_r:nfs_t:s0 342 Jan 19 13:38 metadata -rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 16M Jan 19 13:38 outbox -rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 1.3M Jan 19 13:38 xleases [root@brick ~]# On Tue, Jan 19, 2021 at 4:13 PM Nir Soffer <nsoffer@redhat.com> wrote:
On Tue, Jan 19, 2021 at 6:13 PM Matt Snow <mattsnow@gmail.com> wrote:
[root@brick log]# cat sanlock.log
2021-01-15 18:18:48 3974 [36280]: sanlock daemon started 3.8.2 host
3b903780-4f79-1018-816e-aeb2724778a7 (brick.co.slakin.net)
2021-01-15 19:17:31 7497 [36293]: s1 lockspace 54532dd4-3e5b-4885-b88e-599c81efb146:250:/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids:0 2021-01-15 19:17:31 7497 [50873]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids
Smells like selinux issue.
What do you see in "ausearch -m avc"?

Hi Nir, Yedidyah, for what it's worth I ran through the steps outlined here: https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/8/htm... and eventually got to running `setenforce 0` and the issue still persists. On Tue, Jan 19, 2021 at 6:52 PM Matt Snow <mattsnow@gmail.com> wrote:
[root@brick ~]# ps -efz | grep sanlock
error: unsupported SysV option
Usage:
ps [options]
Try 'ps --help <simple|list|output|threads|misc|all>'
or 'ps --help <s|l|o|t|m|a>'
for additional help text.
For more details see ps(1).
[root@brick ~]# ps -ef | grep sanlock
sanlock 1308 1 0 10:21 ? 00:00:01 /usr/sbin/sanlock daemon
root 1309 1308 0 10:21 ? 00:00:00 /usr/sbin/sanlock daemon
root 68086 67674 0 13:38 pts/4 00:00:00 tail -f sanlock.log
root 73724 68214 0 18:49 pts/5 00:00:00 grep --color=auto sanlock
[root@brick ~]# ausearch -m avc
<no matches>
[root@brick ~]# ls -lhZ /rhev/data-center/mnt/stumpy\:_tanker_ovirt_host__storage/8fd5420f-61fd-41af-8575-f61853a18d91/dom_md
total 278K
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 0 Jan 19 13:38 ids
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 16M Jan 19 13:38 inbox
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 0 Jan 19 13:38 leases
-rw-rw-r--. 1 vdsm kvm system_u:object_r:nfs_t:s0 342 Jan 19 13:38 metadata
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 16M Jan 19 13:38 outbox
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 1.3M Jan 19 13:38 xleases
[root@brick ~]#
On Tue, Jan 19, 2021 at 4:13 PM Nir Soffer <nsoffer@redhat.com> wrote:
On Tue, Jan 19, 2021 at 6:13 PM Matt Snow <mattsnow@gmail.com> wrote:
[root@brick log]# cat sanlock.log
2021-01-15 18:18:48 3974 [36280]: sanlock daemon started 3.8.2 host
3b903780-4f79-1018-816e-aeb2724778a7 (brick.co.slakin.net)
2021-01-15 19:17:31 7497 [36293]: s1 lockspace 54532dd4-3e5b-4885-b88e-599c81efb146:250:/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids:0 2021-01-15 19:17:31 7497 [50873]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids
Smells like selinux issue.
What do you see in "ausearch -m avc"?

On Wed, Jan 20, 2021 at 6:45 AM Matt Snow <mattsnow@gmail.com> wrote:
Hi Nir, Yedidyah, for what it's worth I ran through the steps outlined here: https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/8/htm... and eventually got to running `setenforce 0` and the issue still persists.
Good, this and no reports in "ausearch -m avc" show that this is not selinux issue.
On Tue, Jan 19, 2021 at 6:52 PM Matt Snow <mattsnow@gmail.com> wrote:
[root@brick ~]# ps -efz | grep sanlock
error: unsupported SysV option
Usage:
ps [options]
Try 'ps --help <simple|list|output|threads|misc|all>'
or 'ps --help <s|l|o|t|m|a>'
for additional help text.
For more details see ps(1).
[root@brick ~]# ps -ef | grep sanlock
sanlock 1308 1 0 10:21 ? 00:00:01 /usr/sbin/sanlock daemon
root 1309 1308 0 10:21 ? 00:00:00 /usr/sbin/sanlock daemon
root 68086 67674 0 13:38 pts/4 00:00:00 tail -f sanlock.log
root 73724 68214 0 18:49 pts/5 00:00:00 grep --color=auto sanlock
[root@brick ~]# ausearch -m avc
<no matches>
[root@brick ~]# ls -lhZ /rhev/data-center/mnt/stumpy\:_tanker_ovirt_host__storage/8fd5420f-61fd-41af-8575-f61853a18d91/dom_md
total 278K
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 0 Jan 19 13:38 ids
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 16M Jan 19 13:38 inbox
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 0 Jan 19 13:38 leases
-rw-rw-r--. 1 vdsm kvm system_u:object_r:nfs_t:s0 342 Jan 19 13:38 metadata
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 16M Jan 19 13:38 outbox
-rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 1.3M Jan 19 13:38 xleases
[root@brick ~]#
On Tue, Jan 19, 2021 at 4:13 PM Nir Soffer <nsoffer@redhat.com> wrote:
On Tue, Jan 19, 2021 at 6:13 PM Matt Snow <mattsnow@gmail.com> wrote:
[root@brick log]# cat sanlock.log
2021-01-15 18:18:48 3974 [36280]: sanlock daemon started 3.8.2 host 3b903780-4f79-1018-816e-aeb2724778a7 (brick.co.slakin.net) 2021-01-15 19:17:31 7497 [36293]: s1 lockspace 54532dd4-3e5b-4885-b88e-599c81efb146:250:/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids:0 2021-01-15 19:17:31 7497 [50873]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/54532dd4-3e5b-4885-b88e-599c81efb146/dom_md/ids
Smells like selinux issue.
What do you see in "ausearch -m avc"?

On Wed, Jan 20, 2021 at 3:52 AM Matt Snow <mattsnow@gmail.com> wrote:
[root@brick ~]# ps -efz | grep sanlock
Sorry, its "ps -efZ", but we already know its not selinux.
[root@brick ~]# ps -ef | grep sanlock sanlock 1308 1 0 10:21 ? 00:00:01 /usr/sbin/sanlock daemon
Does sanlock run with the right groups? On a working system: $ ps -efZ | grep sanlock | grep -v grep system_u:system_r:sanlock_t:s0-s0:c0.c1023 sanlock 983 1 0 11:23 ? 00:00:03 /usr/sbin/sanlock daemon system_u:system_r:sanlock_t:s0-s0:c0.c1023 root 986 983 0 11:23 ? 00:00:00 /usr/sbin/sanlock daemon The sanlock process running with "sanlock" user (pid=983) is the interesting one. The other one is a helper that never accesses storage. $ grep Groups: /proc/983/status Groups: 6 36 107 179 Vdsm verify this on startup using vdsm-tool is-configured. On a working system: $ sudo vdsm-tool is-configured lvm is configured for vdsm libvirt is already configured for vdsm sanlock is configured for vdsm Managed volume database is already configured Current revision of multipath.conf detected, preserving abrt is already configured for vdsm
[root@brick ~]# ausearch -m avc <no matches>
Looks good.
[root@brick ~]# ls -lhZ /rhev/data-center/mnt/stumpy\:_tanker_ovirt_host__storage/8fd5420f-61fd-41af-8575-f61853a18d91/dom_md total 278K -rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 0 Jan 19 13:38 ids
Looks correct.

On Wed, Jan 20, 2021 at 3:00 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Wed, Jan 20, 2021 at 3:52 AM Matt Snow <mattsnow@gmail.com> wrote:
[root@brick ~]# ps -efz | grep sanlock
Sorry, its "ps -efZ", but we already know its not selinux.
[root@brick ~]# ps -ef | grep sanlock sanlock 1308 1 0 10:21 ? 00:00:01 /usr/sbin/sanlock daemon
Does sanlock run with the right groups?
It appears so.
On a working system:
$ ps -efZ | grep sanlock | grep -v grep system_u:system_r:sanlock_t:s0-s0:c0.c1023 sanlock 983 1 0 11:23 ? 00:00:03 /usr/sbin/sanlock daemon system_u:system_r:sanlock_t:s0-s0:c0.c1023 root 986 983 0 11:23 ? 00:00:00 /usr/sbin/sanlock daemon
[root@brick audit]# ps -efZ | grep sanlock | grep -v grep system_u:system_r:sanlock_t:s0-s0:c0.c1023 sanlock 1308 1 0 Jan19 ? 00:00:04 /usr/sbin/sanlock daemon system_u:system_r:sanlock_t:s0-s0:c0.c1023 root 1309 1308 0 Jan19 ? 00:00:00 /usr/sbin/sanlock daemon
The sanlock process running with "sanlock" user (pid=983) is the interesting one. The other one is a helper that never accesses storage.
$ grep Groups: /proc/983/status Groups: 6 36 107 179
[root@brick audit]# grep Groups: /proc/1308/status Groups: 6 36 107 179
Vdsm verify this on startup using vdsm-tool is-configured. On a working system:
$ sudo vdsm-tool is-configured lvm is configured for vdsm libvirt is already configured for vdsm sanlock is configured for vdsm Managed volume database is already configured Current revision of multipath.conf detected, preserving abrt is already configured for vdsm
[root@brick audit]# vdsm-tool is-configured
lvm is configured for vdsm libvirt is already configured for vdsm sanlock is configured for vdsm Current revision of multipath.conf detected, preserving abrt is already configured for vdsm Managed volume database is already configured
[root@brick ~]# ausearch -m avc <no matches>
Looks good.
[root@brick ~]# ls -lhZ /rhev/data-center/mnt/stumpy\:_tanker_ovirt_host__storage/8fd5420f-61fd-41af-8575-f61853a18d91/dom_md total 278K -rw-rw----. 1 vdsm kvm system_u:object_r:nfs_t:s0 0 Jan 19 13:38 ids
Looks correct.

Hallo All, I'm not sure if this was mentioned but I had some problems with NFS, using Synology, a while back and the solution for me was to use a subfolder instead of the root of the NFS mount. My exports: /volume1/nfs *(rw,sec=sys,anonuid=36,anongid=36) And I mount data in oVirt as /volume1/nfs/data. Before I re-installed I used /volume1/nfs/data as an export. Worth a try? Regards, Joop

Thanks, Joop. I saw this solution on the list and tried it as well. It did not work for me. On Fri, Jan 22, 2021 at 3:48 AM Joop <jvdwege@xs4all.nl> wrote:
Hallo All,
I'm not sure if this was mentioned but I had some problems with NFS, using Synology, a while back and the solution for me was to use a subfolder instead of the root of the NFS mount. My exports: /volume1/nfs *(rw,sec=sys,anonuid=36,anongid=36)
And I mount data in oVirt as /volume1/nfs/data. Before I re-installed I used /volume1/nfs/data as an export. Worth a try?
Regards,
Joop _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/BXK27HXYNB2X6H...

On Tue, Jan 19, 2021 at 3:43 PM Yedidyah Bar David <didi@redhat.com> wrote: ...
2021-01-18 08:43:25,524-0700 INFO (jsonrpc/0) [storage.SANLock] Initializing sanlock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 path=/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids alignment=1048576 block_size=512 io_timeout=10 (clusterlock:286) 2021-01-18 08:43:25,533-0700 ERROR (jsonrpc/0) [storage.SANLock] Cannot initialize lock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 (clusterlock:305) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 295, in initLock sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such device') 2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [vdsm.api] FINISH createStorageDomain error=Could not initialize cluster lock: () from=::ffff:192.168.222.53,39612, flow_id=5618fb28, task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:52) 2021-01-18 08:43:25,534-0700 ERROR (jsonrpc/0) [storage.TaskManager.Task] (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 295, in initLock sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such device')
I think this ^^ is the issue. Can you please check /var/log/sanlock.log?
This is a symptom of inaccessible storage. Sanlock failed to write to the "ids" file. We need to understand why sanlock failed. If the partialy created domain is still available, this may explain the issue: ls -lhZ /rhv/data-center/mnt/server:_path/storage-domain-uuid/dom_md ps -efz | grep sanlock ausearch -m avc Nir

On Mon, Jan 18, 2021 at 8:58 AM Matt Snow <mattsnow@gmail.com> wrote:
I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the same problem with both versions. The issue occurs in both cockpit UI and tmux'd CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is created and running. I tried to do some debugging on my own before reaching out to this list. Any help is much appreciated!
ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4 cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but I believe it meets the minimum requirements.
NFS server: * Ubuntu 19.10 w/ ZFS share w/ 17TB available space.
We don't test ZFS, but since you tried also non-ZFS server with same issue, the issue is probably not ZFS.
* NFS share settings are just 'rw=@172.16.1.0/24' but have also tried 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure'
You are missing anonuid=36,anongid=36. This should not affect sanlock, but you will have issues with libvirt without this. Here is working export from my test system: /export/00 *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
* The target directory is always empty and chown'd 36:36 with 0755 permissions.
Correct.
* I have tried using both IP and DNS names. forward and reverse DNS works from ovirt host and other systems on the network. * The NFS share always gets mounted successfully on the ovirt node system. * I have tried auto and v3 NFS versions in other various combinations. * I have also tried setting up an NFS server on a non-ZFS backed storage system that is open to any host and get the same errors as shown below. * I ran nfs-check.py script without issue against both NFS servers and followed other verification steps listed on https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html
***Snip from ovirt-hosted-engine-setup*** Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]: Please specify the full shared storage connection path to use (example: host:/path): stumpy.mydomain.com:/tanker/ovirt/host_storage If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: rw [ INFO ] Creating Storage Domain [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Force facts gathering] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the storage interface to be up] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check local VM dir stat] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Enforce local VM dir existence] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch host facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter name] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add NFS storage domain] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add glusterfs storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add iSCSI storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add Fibre Channel storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get storage domain details] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the appliance OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Parse OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get required size] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Remove unsuitable storage domain] [ INFO ] skipping: [localhost] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check storage domain free space] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
***End snippet**
The relevant section from /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-ansible-create_storage_domain-20210116195220-bnfg1w.log: **Snip*** 2021-01-16 19:53:56,010-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a04c8320> kwargs 2021-01-16 19:53:57,219-0700 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Activate storage domain'} 2021-01-16 19:53:57,220-0700 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs is_conditional:False 2021-01-16 19:53:57,221-0700 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs 2021-01-16 19:54:00,346-0700 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,347-0700 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,348-0700 DEBUG var changed: host "localhost" var "play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,349-0700 ERROR ansible failed { "ansible_host": "localhost", "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_result": { "_ansible_no_log": false, "changed": false, "exception": "Traceback (most recent call last):\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 783, in main\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 638, in post_create_check\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 3647, in add\n return self._internal_add(storage_domain, headers, query, wait)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n return future.wait() if wait else future\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in wait\n return self._code(response)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in callback\n self._check_fault(respon se)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n self._raise_error(response, body)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n", "invocation": { "module_args": { "backup": null, "comment": null, "critical_space_action_blocker": null, "data_center": "Default", "description": null, "destroy": null, "discard_after_delete": null, "domain_function": "data", "fcp": null, "fetch_nested": false, "format": null, "glusterfs": null, "host": "brick.mydomain.com", "id": null, "iscsi": null, "localfs": null, "managed_block_storage": null, "name": "hosted_storage", "nested_attributes": [], "nfs": null, "poll_interval": 3, "posixfs": null, "state": "present", "timeout": 180, "wait": true, "warning_low_space": null, "wipe_after_delete": null } }, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400." }, "ansible_task": "Activate storage domain", "ansible_type": "task", "status": "FAILED", "task_duration": 4 } 2021-01-16 19:54:00,350-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a05f2d30> kwargs ignore_errors:None 2021-01-16 19:54:00,358-0700 INFO ansible stats { "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_playbook_duration": "01:35 Minutes", "ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 23, 'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 7, 'rescued': 0, 'ignored': 0}}", "ansible_type": "finish", "status": "FAILED" } 2021-01-16 19:54:00,359-0700 INFO SUMMARY: Duration Task Name -------- -------- [ < 1 sec ] Execute just a specific set of steps [ 00:05 ] Force facts gathering [ 00:03 ] Check local VM dir stat [ 00:04 ] Obtain SSO token using username/password credentials [ 00:04 ] Fetch host facts [ 00:02 ] Fetch cluster ID [ 00:04 ] Fetch cluster facts [ 00:04 ] Fetch Datacenter facts [ 00:02 ] Fetch Datacenter ID [ 00:02 ] Fetch Datacenter name [ 00:04 ] Add NFS storage domain [ 00:04 ] Get storage domain details [ 00:03 ] Find the appliance OVF [ 00:03 ] Parse OVF [ 00:02 ] Get required size [ FAILED ] Activate storage domain 2021-01-16 19:54:00,359-0700 DEBUG ansible on_any args <ansible.executor.stats.AggregateStats object at 0x7f11a2e3b8d0> kwargs **End snip**
The relevantsection from /var/log/vdsm/vdsm.log: ***begin snip*** 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'status': 0}]} from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:54) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [vdsm.api] START getStorageDomainInfo(sdUUID='54532dd4-3e5b-4885-b88e-599c81efb146', options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:48) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [vdsm.api] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:52) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-131>", line 2, in getStorageDomainInfo File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2796, in getStorageDomainInfo dom = self.validateSdUUID(sdUUID) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 312, in validateSdUUID sdDom.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.Dispatcher] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,919-0700 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call StorageDomain.getInfo failed (error 379) in 0.00 seconds (__init__:312) 2021-01-16 19:53:58,956-0700 INFO (jsonrpc/1) [vdsm.api] START createStoragePool(poolType=None, spUUID='91fb3f22-5795-11eb-ad9f-00163e3f4683', poolName='Default', masterDom='54532dd4-3e5b-4885-b88e-599c81efb146', domList=['54532dd4-3e5b-4885-b88e-599c81efb146'], masterVersion=7, lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, ioOpTimeoutSec=10, leaseRetries=3, options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:48) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] updating pool 91fb3f22-5795-11eb-ad9f-00163e3f4683 backend from type NoneType instance 0x7f86e45f89d0 to type StoragePoolDiskBackend instance 0x7f8680738408 (sp:157) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] spUUID=91fb3f22-5795-11eb-ad9f-00163e3f4683 poolName=Default master_sd=54532dd4-3e5b-4885-b88e-599c81efb146 domList=['54532dd4-3e5b-4885-b88e-599c81efb146'] masterVersion=7 {'LEASERETRIES': 3, 'LEASETIMESEC': 60, 'LOCKRENEWALINTERVALSEC': 5, 'IOOPTIMEOUTSEC': 10} (sp:602) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.StoragePool] Unexpected error (sp:618) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,963-0700 INFO (jsonrpc/1) [vdsm.api] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:52) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',)
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-31>", line 2, in createStoragePool File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 1027, in createStoragePool leaseParams) File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 619, in create raise se.StorageDomainAccessError(sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,964-0700 INFO (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,964-0700 ERROR (jsonrpc/1) [storage.Dispatcher] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,965-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call StoragePool.create failed (error 379) in 0.01 seconds (__init__:312) 2021-01-16 19:54:01,923-0700 INFO (jsonrpc/5) [api.host] START getStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,964-0700 INFO (jsonrpc/7) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,966-0700 INFO (jsonrpc/7) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) 2021-01-16 19:54:01,977-0700 INFO (jsonrpc/5) [vdsm.api] START repoStats(domains=()) from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:48) 2021-01-16 19:54:01,978-0700 INFO (jsonrpc/5) [vdsm.api] FINISH repoStats return={} from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:54) 2021-01-16 19:54:01,980-0700 INFO (jsonrpc/5) [vdsm.api] START multipath_health() from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:48) 2021-01-16 19:54:01,981-0700 INFO (jsonrpc/5) [vdsm.api] FINISH multipath_health return={} from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:54) 2021-01-16 19:54:01,996-0700 INFO (jsonrpc/5) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) ***end snip*** _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/G2TQC6XTJTIBAI...

On Wed, Jan 20, 2021 at 2:46 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 18, 2021 at 8:58 AM Matt Snow <mattsnow@gmail.com> wrote:
I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the
same problem with both versions. The issue occurs in both cockpit UI and tmux'd CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is created and running.
I tried to do some debugging on my own before reaching out to this list. Any help is much appreciated!
ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4 cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but I believe it meets the minimum requirements.
NFS server: * Ubuntu 19.10 w/ ZFS share w/ 17TB available space.
We don't test ZFS, but since you tried also non-ZFS server with same issue, the issue is probably not ZFS.
* NFS share settings are just 'rw=@172.16.1.0/24' but have also tried 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure'
You are missing anonuid=36,anongid=36. This should not affect sanlock, but you will have issues with libvirt without this.
I found this setting in another email thread and have applied it root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs tanker/ovirt sharenfs rw,anonuid=36,anongid=36 local root@stumpy:/tanker/ovirt/host_storage#
Here is working export from my test system:
/export/00 *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
I have updated both ZFS server and the Ubuntu 20.04 NFS server to use the
above settings. # ZFS Server root@stumpy:/tanker/ovirt/host_storage# zfs set sharenfs="rw,sync,anonuid=36,anongid=36,no_subtree_check" tanker/ovirt root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs tanker/ovirt sharenfs rw,sync,anonuid=36,anongid=36,no_subtree_check local root@stumpy:/tanker/ovirt/host_storage# # Ubuntu laptop ls -ld /exports/host_storage drwxr-xr-x 2 36 36 4096 Jan 16 19:42 /exports/host_storage root@msnowubntlt:/exports/host_storage# showmount -e dongle Export list for dongle: /exports/host_storage * root@msnowubntlt:/exports/host_storage# grep host_storage /etc/exports /exports/host_storage *(rw,sync,anonuid=36,anongid=36,no_subtree_check) root@msnowubntlt:/exports/host_storage# Interesting point - upon being re-prompted to configure storage after initial failure I provided *different* NFS server information (the ubuntu laptop) **snip** [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]: Please specify the full shared storage connection path to use (example: host:/path): dongle:/exports/host_storage If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: [ INFO ] Creating Storage Domain [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps] **snip** [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: **snip** Upon checking mounts I see that the original server is still being used. [root@brick ~]# mount | grep nfs sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime) stumpy:/tanker/ovirt/host_storage on /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50) [root@brick ~]#
* The target directory is always empty and chown'd 36:36 with 0755 permissions.
Correct.
* I have tried using both IP and DNS names. forward and reverse DNS works from ovirt host and other systems on the network. * The NFS share always gets mounted successfully on the ovirt node system. * I have tried auto and v3 NFS versions in other various combinations. * I have also tried setting up an NFS server on a non-ZFS backed storage system that is open to any host and get the same errors as shown below. * I ran nfs-check.py script without issue against both NFS servers and followed other verification steps listed on https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html
***Snip from ovirt-hosted-engine-setup*** Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]: Please specify the full shared storage connection path to use (example: host:/path): stumpy.mydomain.com:/tanker/ovirt/host_storage If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: rw [ INFO ] Creating Storage Domain [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Force facts gathering] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the storage interface to be up] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check local VM dir stat] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Enforce local VM dir existence] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch host facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter name] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add NFS storage domain] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add glusterfs storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add iSCSI storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add Fibre Channel storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get storage domain details] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the appliance OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Parse OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get required size] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Remove unsuitable storage domain] [ INFO ] skipping: [localhost] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check storage domain free space] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
***End snippet**
The relevant section from /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-ansible-create_storage_domain-20210116195220-bnfg1w.log: **Snip*** 2021-01-16 19:53:56,010-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a04c8320> kwargs 2021-01-16 19:53:57,219-0700 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Activate storage domain'} 2021-01-16 19:53:57,220-0700 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs is_conditional:False 2021-01-16 19:53:57,221-0700 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs 2021-01-16 19:54:00,346-0700 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,347-0700 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,348-0700 DEBUG var changed: host "localhost" var "play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,349-0700 ERROR ansible failed { "ansible_host": "localhost", "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_result": { "_ansible_no_log": false, "changed": false, "exception": "Traceback (most recent call last):\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 783, in main\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 638, in post_create_check\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 3647, in add\n return self._internal_add(storage_domain, headers, query, wait)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n return future.wait() if wait else future\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in wait\n return self._code(response)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in callback\n self._check_fault(respon se)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n self._raise_error(response, body)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n", "invocation": { "module_args": { "backup": null, "comment": null, "critical_space_action_blocker": null, "data_center": "Default", "description": null, "destroy": null, "discard_after_delete": null, "domain_function": "data", "fcp": null, "fetch_nested": false, "format": null, "glusterfs": null, "host": "brick.mydomain.com", "id": null, "iscsi": null, "localfs": null, "managed_block_storage": null, "name": "hosted_storage", "nested_attributes": [], "nfs": null, "poll_interval": 3, "posixfs": null, "state": "present", "timeout": 180, "wait": true, "warning_low_space": null, "wipe_after_delete": null } }, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400." }, "ansible_task": "Activate storage domain", "ansible_type": "task", "status": "FAILED", "task_duration": 4 } 2021-01-16 19:54:00,350-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a05f2d30> kwargs ignore_errors:None 2021-01-16 19:54:00,358-0700 INFO ansible stats { "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_playbook_duration": "01:35 Minutes", "ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 23, 'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 7, 'rescued': 0, 'ignored': 0}}", "ansible_type": "finish", "status": "FAILED" } 2021-01-16 19:54:00,359-0700 INFO SUMMARY: Duration Task Name -------- -------- [ < 1 sec ] Execute just a specific set of steps [ 00:05 ] Force facts gathering [ 00:03 ] Check local VM dir stat [ 00:04 ] Obtain SSO token using username/password credentials [ 00:04 ] Fetch host facts [ 00:02 ] Fetch cluster ID [ 00:04 ] Fetch cluster facts [ 00:04 ] Fetch Datacenter facts [ 00:02 ] Fetch Datacenter ID [ 00:02 ] Fetch Datacenter name [ 00:04 ] Add NFS storage domain [ 00:04 ] Get storage domain details [ 00:03 ] Find the appliance OVF [ 00:03 ] Parse OVF [ 00:02 ] Get required size [ FAILED ] Activate storage domain 2021-01-16 19:54:00,359-0700 DEBUG ansible on_any args <ansible.executor.stats.AggregateStats object at 0x7f11a2e3b8d0> kwargs **End snip**
The relevantsection from /var/log/vdsm/vdsm.log: ***begin snip*** 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'status': 0}]} from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:54) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [vdsm.api] START getStorageDomainInfo(sdUUID='54532dd4-3e5b-4885-b88e-599c81efb146', options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:48) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [vdsm.api] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:52) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-131>", line 2, in getStorageDomainInfo File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2796, in getStorageDomainInfo dom = self.validateSdUUID(sdUUID) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 312, in validateSdUUID sdDom.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.Dispatcher] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,919-0700 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call StorageDomain.getInfo failed (error 379) in 0.00 seconds (__init__:312) 2021-01-16 19:53:58,956-0700 INFO (jsonrpc/1) [vdsm.api] START createStoragePool(poolType=None, spUUID='91fb3f22-5795-11eb-ad9f-00163e3f4683', poolName='Default', masterDom='54532dd4-3e5b-4885-b88e-599c81efb146', domList=['54532dd4-3e5b-4885-b88e-599c81efb146'], masterVersion=7, lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, ioOpTimeoutSec=10, leaseRetries=3, options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:48) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] updating pool 91fb3f22-5795-11eb-ad9f-00163e3f4683 backend from type NoneType instance 0x7f86e45f89d0 to type StoragePoolDiskBackend instance 0x7f8680738408 (sp:157) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] spUUID=91fb3f22-5795-11eb-ad9f-00163e3f4683 poolName=Default master_sd=54532dd4-3e5b-4885-b88e-599c81efb146 domList=['54532dd4-3e5b-4885-b88e-599c81efb146'] masterVersion=7 {'LEASERETRIES': 3, 'LEASETIMESEC': 60, 'LOCKRENEWALINTERVALSEC': 5, 'IOOPTIMEOUTSEC': 10} (sp:602) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.StoragePool] Unexpected error (sp:618) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,963-0700 INFO (jsonrpc/1) [vdsm.api] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:52) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',)
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-31>", line 2, in createStoragePool File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 1027, in createStoragePool leaseParams) File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 619, in create raise se.StorageDomainAccessError(sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,964-0700 INFO (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,964-0700 ERROR (jsonrpc/1) [storage.Dispatcher] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,965-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call StoragePool.create failed (error 379) in 0.01 seconds (__init__:312) 2021-01-16 19:54:01,923-0700 INFO (jsonrpc/5) [api.host] START getStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,964-0700 INFO (jsonrpc/7) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,966-0700 INFO (jsonrpc/7) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) 2021-01-16 19:54:01,977-0700 INFO (jsonrpc/5) [vdsm.api] START repoStats(domains=()) from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:48) 2021-01-16 19:54:01,978-0700 INFO (jsonrpc/5) [vdsm.api] FINISH repoStats return={} from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:54) 2021-01-16 19:54:01,980-0700 INFO (jsonrpc/5) [vdsm.api] START multipath_health() from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:48) 2021-01-16 19:54:01,981-0700 INFO (jsonrpc/5) [vdsm.api] FINISH multipath_health return={} from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:54) 2021-01-16 19:54:01,996-0700 INFO (jsonrpc/5) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) ***end snip*** _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/G2TQC6XTJTIBAI...

On Wed, Jan 20, 2021 at 7:36 PM Matt Snow <mattsnow@gmail.com> wrote:
On Wed, Jan 20, 2021 at 2:46 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 18, 2021 at 8:58 AM Matt Snow <mattsnow@gmail.com> wrote:
I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the same problem with both versions. The issue occurs in both cockpit UI and tmux'd CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is created and running. I tried to do some debugging on my own before reaching out to this list. Any help is much appreciated!
ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4 cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but I believe it meets the minimum requirements.
NFS server: * Ubuntu 19.10 w/ ZFS share w/ 17TB available space.
We don't test ZFS, but since you tried also non-ZFS server with same issue, the issue is probably not ZFS.
* NFS share settings are just 'rw=@172.16.1.0/24' but have also tried 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure'
You are missing anonuid=36,anongid=36. This should not affect sanlock, but you will have issues with libvirt without this.
I found this setting in another email thread and have applied it
root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs
tanker/ovirt sharenfs rw,anonuid=36,anongid=36 local
root@stumpy:/tanker/ovirt/host_storage#
Here is working export from my test system:
/export/00 *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
I have updated both ZFS server and the Ubuntu 20.04 NFS server to use the above settings. # ZFS Server
root@stumpy:/tanker/ovirt/host_storage# zfs set sharenfs="rw,sync,anonuid=36,anongid=36,no_subtree_check" tanker/ovirt
root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs
tanker/ovirt sharenfs rw,sync,anonuid=36,anongid=36,no_subtree_check local
root@stumpy:/tanker/ovirt/host_storage#
# Ubuntu laptop
ls -ld /exports/host_storage
drwxr-xr-x 2 36 36 4096 Jan 16 19:42 /exports/host_storage
root@msnowubntlt:/exports/host_storage# showmount -e dongle
Export list for dongle:
/exports/host_storage *
root@msnowubntlt:/exports/host_storage# grep host_storage /etc/exports
/exports/host_storage *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
root@msnowubntlt:/exports/host_storage#
Interesting point - upon being re-prompted to configure storage after initial failure I provided *different* NFS server information (the ubuntu laptop) **snip**
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."}
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs
Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]:
Please specify the full shared storage connection path to use (example: host:/path): dongle:/exports/host_storage
If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []:
[ INFO ] Creating Storage Domain
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps]
**snip**
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."}
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
**snip** Upon checking mounts I see that the original server is still being used.
[root@brick ~]# mount | grep nfs
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime)
stumpy:/tanker/ovirt/host_storage on /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50)
Using node and hosted engine makes everything more complicated and harder to debug. I would create an engine vm for testing on some machine, and add this host as regular host. It will be easier to debug this way, since you don't have to reinstall everything from scratch on every failure. Once we sort out the issue with the NFS/ZFS storage, using node and hosted engine makes sense for better availability and saving hardware resources. Also, can you explain how to build and configure identical ubuntu storage server? I can try to test a vm if you give detailed description. Maybe this an issue with AppArmor? https://linuxconfig.org/how-to-disable-enable-selinux-on-ubuntu-20-04-focal-... Did you try to disable it on the storage server, or enable selinux?
[root@brick ~]#
* The target directory is always empty and chown'd 36:36 with 0755 permissions.
Correct.
* I have tried using both IP and DNS names. forward and reverse DNS works from ovirt host and other systems on the network. * The NFS share always gets mounted successfully on the ovirt node system. * I have tried auto and v3 NFS versions in other various combinations. * I have also tried setting up an NFS server on a non-ZFS backed storage system that is open to any host and get the same errors as shown below. * I ran nfs-check.py script without issue against both NFS servers and followed other verification steps listed on https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html
***Snip from ovirt-hosted-engine-setup*** Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]: Please specify the full shared storage connection path to use (example: host:/path): stumpy.mydomain.com:/tanker/ovirt/host_storage If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: rw [ INFO ] Creating Storage Domain [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Force facts gathering] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the storage interface to be up] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check local VM dir stat] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Enforce local VM dir existence] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch host facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter name] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add NFS storage domain] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add glusterfs storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add iSCSI storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add Fibre Channel storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get storage domain details] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the appliance OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Parse OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get required size] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Remove unsuitable storage domain] [ INFO ] skipping: [localhost] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check storage domain free space] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
***End snippet**
The relevant section from /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-ansible-create_storage_domain-20210116195220-bnfg1w.log: **Snip*** 2021-01-16 19:53:56,010-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a04c8320> kwargs 2021-01-16 19:53:57,219-0700 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Activate storage domain'} 2021-01-16 19:53:57,220-0700 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs is_conditional:False 2021-01-16 19:53:57,221-0700 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs 2021-01-16 19:54:00,346-0700 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,347-0700 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,348-0700 DEBUG var changed: host "localhost" var "play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,349-0700 ERROR ansible failed { "ansible_host": "localhost", "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_result": { "_ansible_no_log": false, "changed": false, "exception": "Traceback (most recent call last):\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 783, in main\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 638, in post_create_check\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 3647, in add\n return self._internal_add(storage_domain, headers, query, wait)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n return future.wait() if wait else future\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in wait\n return self._code(response)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in callback\n self._check_fault(respon se)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n self._raise_error(response, body)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n", "invocation": { "module_args": { "backup": null, "comment": null, "critical_space_action_blocker": null, "data_center": "Default", "description": null, "destroy": null, "discard_after_delete": null, "domain_function": "data", "fcp": null, "fetch_nested": false, "format": null, "glusterfs": null, "host": "brick.mydomain.com", "id": null, "iscsi": null, "localfs": null, "managed_block_storage": null, "name": "hosted_storage", "nested_attributes": [], "nfs": null, "poll_interval": 3, "posixfs": null, "state": "present", "timeout": 180, "wait": true, "warning_low_space": null, "wipe_after_delete": null } }, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400." }, "ansible_task": "Activate storage domain", "ansible_type": "task", "status": "FAILED", "task_duration": 4 } 2021-01-16 19:54:00,350-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a05f2d30> kwargs ignore_errors:None 2021-01-16 19:54:00,358-0700 INFO ansible stats { "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_playbook_duration": "01:35 Minutes", "ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 23, 'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 7, 'rescued': 0, 'ignored': 0}}", "ansible_type": "finish", "status": "FAILED" } 2021-01-16 19:54:00,359-0700 INFO SUMMARY: Duration Task Name -------- -------- [ < 1 sec ] Execute just a specific set of steps [ 00:05 ] Force facts gathering [ 00:03 ] Check local VM dir stat [ 00:04 ] Obtain SSO token using username/password credentials [ 00:04 ] Fetch host facts [ 00:02 ] Fetch cluster ID [ 00:04 ] Fetch cluster facts [ 00:04 ] Fetch Datacenter facts [ 00:02 ] Fetch Datacenter ID [ 00:02 ] Fetch Datacenter name [ 00:04 ] Add NFS storage domain [ 00:04 ] Get storage domain details [ 00:03 ] Find the appliance OVF [ 00:03 ] Parse OVF [ 00:02 ] Get required size [ FAILED ] Activate storage domain 2021-01-16 19:54:00,359-0700 DEBUG ansible on_any args <ansible.executor.stats.AggregateStats object at 0x7f11a2e3b8d0> kwargs **End snip**
The relevantsection from /var/log/vdsm/vdsm.log: ***begin snip*** 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'status': 0}]} from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:54) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [vdsm.api] START getStorageDomainInfo(sdUUID='54532dd4-3e5b-4885-b88e-599c81efb146', options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:48) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [vdsm.api] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:52) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-131>", line 2, in getStorageDomainInfo File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2796, in getStorageDomainInfo dom = self.validateSdUUID(sdUUID) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 312, in validateSdUUID sdDom.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.Dispatcher] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,919-0700 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call StorageDomain.getInfo failed (error 379) in 0.00 seconds (__init__:312) 2021-01-16 19:53:58,956-0700 INFO (jsonrpc/1) [vdsm.api] START createStoragePool(poolType=None, spUUID='91fb3f22-5795-11eb-ad9f-00163e3f4683', poolName='Default', masterDom='54532dd4-3e5b-4885-b88e-599c81efb146', domList=['54532dd4-3e5b-4885-b88e-599c81efb146'], masterVersion=7, lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, ioOpTimeoutSec=10, leaseRetries=3, options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:48) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] updating pool 91fb3f22-5795-11eb-ad9f-00163e3f4683 backend from type NoneType instance 0x7f86e45f89d0 to type StoragePoolDiskBackend instance 0x7f8680738408 (sp:157) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] spUUID=91fb3f22-5795-11eb-ad9f-00163e3f4683 poolName=Default master_sd=54532dd4-3e5b-4885-b88e-599c81efb146 domList=['54532dd4-3e5b-4885-b88e-599c81efb146'] masterVersion=7 {'LEASERETRIES': 3, 'LEASETIMESEC': 60, 'LOCKRENEWALINTERVALSEC': 5, 'IOOPTIMEOUTSEC': 10} (sp:602) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.StoragePool] Unexpected error (sp:618) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,963-0700 INFO (jsonrpc/1) [vdsm.api] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:52) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line 538, in validate raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',)
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-31>", line 2, in createStoragePool File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 1027, in createStoragePool leaseParams) File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 619, in create raise se.StorageDomainAccessError(sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) 2021-01-16 19:53:58,964-0700 INFO (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,964-0700 ERROR (jsonrpc/1) [storage.Dispatcher] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,965-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call StoragePool.create failed (error 379) in 0.01 seconds (__init__:312) 2021-01-16 19:54:01,923-0700 INFO (jsonrpc/5) [api.host] START getStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,964-0700 INFO (jsonrpc/7) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,966-0700 INFO (jsonrpc/7) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) 2021-01-16 19:54:01,977-0700 INFO (jsonrpc/5) [vdsm.api] START repoStats(domains=()) from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:48) 2021-01-16 19:54:01,978-0700 INFO (jsonrpc/5) [vdsm.api] FINISH repoStats return={} from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:54) 2021-01-16 19:54:01,980-0700 INFO (jsonrpc/5) [vdsm.api] START multipath_health() from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:48) 2021-01-16 19:54:01,981-0700 INFO (jsonrpc/5) [vdsm.api] FINISH multipath_health return={} from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:54) 2021-01-16 19:54:01,996-0700 INFO (jsonrpc/5) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) ***end snip*** _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/G2TQC6XTJTIBAI...

Hi Nir, I will check selinux on both NFS servers. I've created a centos8.3 cloud image KVM on the Ubuntu 20.04 laptop. I am stepping through the installing ovirt as a standalone manager with local databases instructions <https://www.ovirt.org/documentation/installing_ovirt_as_a_standalone_manager_with_local_databases/#Installing_the_Red_Hat_Virtualization_Manager_SM_localDB_deploy> . Right now I am waiting for the host add to finish in the ovirt manager and will try to add the NFS shares once that is done. It would be slick if the self hosted engine worked out of the box with node. Maybe in the future. :) I will let you know what I find out tomorrow. Thank you for your help so far! ..Matt On Thu, Jan 21, 2021 at 6:27 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Wed, Jan 20, 2021 at 7:36 PM Matt Snow <mattsnow@gmail.com> wrote:
On Wed, Jan 20, 2021 at 2:46 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 18, 2021 at 8:58 AM Matt Snow <mattsnow@gmail.com> wrote:
I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the
I tried to do some debugging on my own before reaching out to this
ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4
cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but I believe it meets the minimum requirements.
NFS server: * Ubuntu 19.10 w/ ZFS share w/ 17TB available space.
We don't test ZFS, but since you tried also non-ZFS server with same issue, the issue is probably not ZFS.
* NFS share settings are just 'rw=@172.16.1.0/24' but have also
You are missing anonuid=36,anongid=36. This should not affect sanlock,
but
you will have issues with libvirt without this.
I found this setting in another email thread and have applied it
root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs
tanker/ovirt sharenfs rw,anonuid=36,anongid=36 local
root@stumpy:/tanker/ovirt/host_storage#
Here is working export from my test system:
/export/00 *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
I have updated both ZFS server and the Ubuntu 20.04 NFS server to use
same problem with both versions. The issue occurs in both cockpit UI and tmux'd CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is created and running. list. Any help is much appreciated! tried 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure' the above settings.
# ZFS Server
root@stumpy:/tanker/ovirt/host_storage# zfs set sharenfs="rw,sync,anonuid=36,anongid=36,no_subtree_check" tanker/ovirt
root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs
tanker/ovirt sharenfs rw,sync,anonuid=36,anongid=36,no_subtree_check local
root@stumpy:/tanker/ovirt/host_storage#
# Ubuntu laptop
ls -ld /exports/host_storage
drwxr-xr-x 2 36 36 4096 Jan 16 19:42 /exports/host_storage
root@msnowubntlt:/exports/host_storage# showmount -e dongle
Export list for dongle:
/exports/host_storage *
root@msnowubntlt:/exports/host_storage# grep host_storage /etc/exports
/exports/host_storage *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
root@msnowubntlt:/exports/host_storage#
Interesting point - upon being re-prompted to configure storage after initial failure I provided *different* NFS server information (the ubuntu laptop) **snip**
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."}
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs
Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]:
Please specify the full shared storage connection path to use (example: host:/path): dongle:/exports/host_storage
If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []:
[ INFO ] Creating Storage Domain
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps]
**snip**
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."}
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
**snip** Upon checking mounts I see that the original server is still being used.
[root@brick ~]# mount | grep nfs
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime)
stumpy:/tanker/ovirt/host_storage on /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50)
Using node and hosted engine makes everything more complicated and harder to debug.
I would create an engine vm for testing on some machine, and add this host as regular host. It will be easier to debug this way, since you don't have to reinstall everything from scratch on every failure.
Once we sort out the issue with the NFS/ZFS storage, using node and hosted engine makes sense for better availability and saving hardware resources.
Also, can you explain how to build and configure identical ubuntu storage server? I can try to test a vm if you give detailed description.
Maybe this an issue with AppArmor?
https://linuxconfig.org/how-to-disable-enable-selinux-on-ubuntu-20-04-focal-...
Did you try to disable it on the storage server, or enable selinux?
[root@brick ~]#
* The target directory is always empty and chown'd 36:36 with 0755
Correct.
* I have tried using both IP and DNS names. forward and reverse DNS
works from ovirt host and other systems on the network.
* The NFS share always gets mounted successfully on the ovirt node system. * I have tried auto and v3 NFS versions in other various combinations. * I have also tried setting up an NFS server on a non-ZFS backed storage system that is open to any host and get the same errors as shown below. * I ran nfs-check.py script without issue against both NFS servers and followed other verification steps listed on https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html
***Snip from ovirt-hosted-engine-setup*** Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]: Please specify the full shared storage connection path to use (example: host:/path): stumpy.mydomain.com:/tanker/ovirt/host_storage If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: rw [ INFO ] Creating Storage Domain [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Force facts gathering] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the storage interface to be up] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check local VM dir stat] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Enforce local VM
[ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch host facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter name] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add NFS storage domain] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add glusterfs storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add iSCSI storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add Fibre Channel storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get storage domain
[ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the appliance OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Parse OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get required size] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Remove unsuitable storage domain] [ INFO ] skipping: [localhost] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check storage domain free space] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
***End snippet**
The relevant section from /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-ansible-create_storage_domain-20210116195220-bnfg1w.log: **Snip*** 2021-01-16 19:53:56,010-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a04c8320> kwargs 2021-01-16 19:53:57,219-0700 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Activate storage domain'} 2021-01-16 19:53:57,220-0700 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs is_conditional:False 2021-01-16 19:53:57,221-0700 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs 2021-01-16 19:54:00,346-0700 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,347-0700 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,348-0700 DEBUG var changed: host "localhost" var "play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,349-0700 ERROR ansible failed { "ansible_host": "localhost", "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_result": { "_ansible_no_log": false, "changed": false, "exception": "Traceback (most recent call last):\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\",
se)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n self._raise_error(response, body)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n", "invocation": { "module_args": { "backup": null, "comment": null, "critical_space_action_blocker": null, "data_center": "Default", "description": null, "destroy": null, "discard_after_delete": null, "domain_function": "data", "fcp": null, "fetch_nested": false, "format": null, "glusterfs": null, "host": "brick.mydomain.com", "id": null, "iscsi": null, "localfs": null, "managed_block_storage": null, "name": "hosted_storage", "nested_attributes": [], "nfs": null, "poll_interval": 3, "posixfs": null, "state": "present", "timeout": 180, "wait": true, "warning_low_space": null, "wipe_after_delete": null } }, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400." }, "ansible_task": "Activate storage domain", "ansible_type": "task", "status": "FAILED", "task_duration": 4 } 2021-01-16 19:54:00,350-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a05f2d30> kwargs ignore_errors:None 2021-01-16 19:54:00,358-0700 INFO ansible stats { "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_playbook_duration": "01:35 Minutes", "ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 23, 'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 7, 'rescued': 0, 'ignored': 0}}", "ansible_type": "finish", "status": "FAILED" } 2021-01-16 19:54:00,359-0700 INFO SUMMARY: Duration Task Name -------- -------- [ < 1 sec ] Execute just a specific set of steps [ 00:05 ] Force facts gathering [ 00:03 ] Check local VM dir stat [ 00:04 ] Obtain SSO token using username/password credentials [ 00:04 ] Fetch host facts [ 00:02 ] Fetch cluster ID [ 00:04 ] Fetch cluster facts [ 00:04 ] Fetch Datacenter facts [ 00:02 ] Fetch Datacenter ID [ 00:02 ] Fetch Datacenter name [ 00:04 ] Add NFS storage domain [ 00:04 ] Get storage domain details [ 00:03 ] Find the appliance OVF [ 00:03 ] Parse OVF [ 00:02 ] Get required size [ FAILED ] Activate storage domain 2021-01-16 19:54:00,359-0700 DEBUG ansible on_any args <ansible.executor.stats.AggregateStats object at 0x7f11a2e3b8d0> kwargs **End snip**
The relevantsection from /var/log/vdsm/vdsm.log: ***begin snip*** 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': '1'}],
2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'status': 0}]} from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:54) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [vdsm.api] START getStorageDomainInfo(sdUUID='54532dd4-3e5b-4885-b88e-599c81efb146',
2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [vdsm.api] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:52) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-131>", line 2, in getStorageDomainInfo File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2796, in getStorageDomainInfo dom = self.validateSdUUID(sdUUID) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 312, in validateSdUUID sdDom.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.Dispatcher] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,919-0700 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call StorageDomain.getInfo failed (error 379) in 0.00 seconds (__init__:312) 2021-01-16 19:53:58,956-0700 INFO (jsonrpc/1) [vdsm.api] START createStoragePool(poolType=None, spUUID='91fb3f22-5795-11eb-ad9f-00163e3f4683', poolName='Default', masterDom='54532dd4-3e5b-4885-b88e-599c81efb146', domList=['54532dd4-3e5b-4885-b88e-599c81efb146'], masterVersion=7, lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, ioOpTimeoutSec=10, leaseRetries=3, options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:48) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] updating pool 91fb3f22-5795-11eb-ad9f-00163e3f4683 backend from type NoneType instance 0x7f86e45f89d0 to type StoragePoolDiskBackend instance 0x7f8680738408 (sp:157) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] spUUID=91fb3f22-5795-11eb-ad9f-00163e3f4683 poolName=Default master_sd=54532dd4-3e5b-4885-b88e-599c81efb146 domList=['54532dd4-3e5b-4885-b88e-599c81efb146'] masterVersion=7 {'LEASERETRIES': 3, 'LEASETIMESEC': 60, 'LOCKRENEWALINTERVALSEC': 5, 'IOOPTIMEOUTSEC': 10} (sp:602) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.StoragePool] Unexpected error (sp:618) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
2021-01-16 19:53:58,963-0700 INFO (jsonrpc/1) [vdsm.api] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:52) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line
887, in _run
return fn(*args, **kargs) File "<decorator-gen-31>", line 2, in createStoragePool File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line
50, in method
ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line
1027, in createStoragePool
leaseParams) File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line
619, in create
raise se.StorageDomainAccessError(sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
permissions. dir existence] details] line 783, in main\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 638, in post_create_check\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 3647, in add\n return self._internal_add(storage_domain, headers, query, wait)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n return future.wait() if wait else future\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in wait\n return self._code(response)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in callback\n self._check_fault(respon options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48) options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:48) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',)
2021-01-16 19:53:58,964-0700 INFO (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,964-0700 ERROR (jsonrpc/1) [storage.Dispatcher] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,965-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call StoragePool.create failed (error 379) in 0.01 seconds (__init__:312) 2021-01-16 19:54:01,923-0700 INFO (jsonrpc/5) [api.host] START getStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,964-0700 INFO (jsonrpc/7) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,966-0700 INFO (jsonrpc/7) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) 2021-01-16 19:54:01,977-0700 INFO (jsonrpc/5) [vdsm.api] START repoStats(domains=()) from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:48) 2021-01-16 19:54:01,978-0700 INFO (jsonrpc/5) [vdsm.api] FINISH repoStats return={} from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:54) 2021-01-16 19:54:01,980-0700 INFO (jsonrpc/5) [vdsm.api] START multipath_health() from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:48) 2021-01-16 19:54:01,981-0700 INFO (jsonrpc/5) [vdsm.api] FINISH multipath_health return={} from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:54) 2021-01-16 19:54:01,996-0700 INFO (jsonrpc/5) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) ***end snip*** _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/G2TQC6XTJTIBAI...

I completed disabled appArmor and selinux on the Ubuntu NFS server (non-ZFS server) root@dongle:/exports/00/storage# grep 'apparmor=0' /etc/default/grub GRUB_CMDLINE_LINUX_DEFAULT="quiet splash apparmor=0" root@msnowubntlt:/exports/00/storage# aa-status apparmor module is loaded. apparmor filesystem is not mounted. root@msnowubntlt:/exports/00/storage# sestatus SELinux status: disabled root@dongle:/exports/00/storage# I went ahead and disabled UFW and Firewalld for good measure. root@dongle:/exports/00# systemctl status ufw.service ● ufw.service - Uncomplicated firewall Loaded: loaded (/lib/systemd/system/ufw.service; enabled; vendor preset: enabled) Active: inactive (dead) since Sun 2021-01-24 14:32:54 MST; 15min ago Docs: man:ufw(8) Process: 353 ExecStart=/lib/ufw/ufw-init start quiet (code=exited, status=0/SUCCESS) Process: 157355 ExecStop=/lib/ufw/ufw-init stop (code=exited, status=0/SUCCESS) Main PID: 353 (code=exited, status=0/SUCCESS) Jan 24 14:32:54 dongle.co.slakin.net systemd[1]: Stopping Uncomplicated firewall... Jan 24 14:32:54 dongle.co.slakin.net ufw-init[157355]: Skip stopping firewall: ufw (not enabled) Jan 24 14:32:54 dongle.co.slakin.net systemd[1]: ufw.service: Succeeded. Jan 24 14:32:54 dongle.co.slakin.net systemd[1]: Stopped Uncomplicated firewall. Warning: journal has been rotated since unit was started, output may be incomplete. root@dongle:/exports/00# systemctl status firewalld.service ● firewalld.service - firewalld - dynamic firewall daemon Loaded: loaded (/lib/systemd/system/firewalld.service; enabled; vendor preset: enabled) Active: inactive (dead) since Sun 2021-01-24 14:33:00 MST; 15min ago Docs: man:firewalld(1) Process: 817 ExecStart=/usr/sbin/firewalld --nofork --nopid (code=exited, status=0/SUCCESS) Main PID: 817 (code=exited, status=0/SUCCESS) Jan 24 14:17:06 dongle.co.slakin.net systemd[1]: Starting firewalld - dynamic firewall daemon... Jan 24 14:17:06 dongle.co.slakin.net systemd[1]: Started firewalld - dynamic firewall daemon. Jan 24 14:32:58 dongle.co.slakin.net systemd[1]: Stopping firewalld - dynamic firewall daemon... Jan 24 14:33:00 dongle.co.slakin.net systemd[1]: firewalld.service: Succeeded. Jan 24 14:33:00 dongle.co.slakin.net systemd[1]: Stopped firewalld - dynamic firewall daemon. root@dongle:/exports/00# I started with a new fresh mount that is empty and owned by 36:36 root@dongle:/exports/00/storage# ls -lan total 8 drwxr-xr-x 2 36 36 4096 Jan 23 10:09 . drwxrwxrwx 3 36 36 4096 Jan 23 09:41 .. root@msnowubntlt:/exports/00/storage# showmount -e $HOSTNAME Export list for msnowubntlt: /exports/00 * /exports/host_storage * root@dongle:/exports/00/storage# I tried to create an NFS domain from the ovirt gui but it still fails. brick(ovirt node host) successfully mounts the NFS share and creates files: root@dongle:/exports/00# find . -exec ls -ln '{}' \; total 4 drwxrwxr-x 3 36 36 4096 Jan 24 14:43 storage total 4 drwxr-xr-x 4 36 36 4096 Jan 24 14:43 a9ab8829-14bd-45cc-a57b-03b01e96f407 total 8 drwxr-xr-x 2 36 36 4096 Jan 24 14:43 dom_md drwxr-xr-x 2 36 36 4096 Jan 24 14:43 images total 0 total 252 -rw-rw---- 1 36 36 0 Jan 24 14:43 ids -rw-rw---- 1 36 36 16777216 Jan 24 14:43 inbox -rw-rw---- 1 36 36 0 Jan 24 14:43 leases -rw-r--r-- 1 36 36 345 Jan 24 14:43 metadata -rw-rw---- 1 36 36 16777216 Jan 24 14:43 outbox -rw-rw---- 1 36 36 1302528 Jan 24 14:43 xleases -rw-rw---- 1 36 36 16777216 Jan 24 14:43 ./storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md/outbox -rw-r--r-- 1 36 36 345 Jan 24 14:43 ./storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md/metadata -rw-rw---- 1 36 36 0 Jan 24 14:43 ./storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md/leases -rw-rw---- 1 36 36 1302528 Jan 24 14:43 ./storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md/xleases -rw-rw---- 1 36 36 16777216 Jan 24 14:43 ./storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md/inbox -rw-rw---- 1 36 36 0 Jan 24 14:43 ./storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md/ids root@dongle:/exports/00# vdsm.log still shows a failure... [root@brick vdsm]# ./reset.sh -rw-r--r--. 1 vdsm kvm 6.9K Jan 24 14:42 mom.log -rw-r--r--. 1 root root 32K Jan 24 14:43 supervdsm.log -rw-r--r--. 1 vdsm kvm 64K Jan 24 14:43 vdsm.log -rw-r--r--. 1 vdsm kvm 2 Jan 24 14:35 vdsm-schema-error.log -rw-r--r--. 1 vdsm kvm 2 Jan 24 14:43 mom.log -rw-r--r--. 1 root root 2 Jan 24 14:43 supervdsm.log -rw-r--r--. 1 vdsm kvm 2 Jan 24 14:43 vdsm.log -rw-r--r--. 1 vdsm kvm 2 Jan 24 14:43 vdsm-schema-error.log 2021-01-24 14:43:09,764-0700 INFO (jsonrpc/4) [api.host] START getAllVmStats() from=::ffff:172.16.1.29,45646 (api:48) 2021-01-24 14:43:09,765-0700 INFO (jsonrpc/4) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:172.16.1.29,45646 (api:54) 2021-01-24 14:43:12,680-0700 INFO (periodic/1) [vdsm.api] START repoStats(domains=()) from=internal, task_id=1f14dd7b-e4e5-4b11-b889-fec1cfa82a53 (api:48) 2021-01-24 14:43:12,682-0700 INFO (periodic/1) [vdsm.api] FINISH repoStats return={} from=internal, task_id=1f14dd7b-e4e5-4b11-b889-fec1cfa82a53 (api:54) 2021-01-24 14:43:13,439-0700 INFO (jsonrpc/7) [api.host] START getStats() from=::ffff:172.16.1.29,45646 (api:48) 2021-01-24 14:43:13,472-0700 INFO (jsonrpc/7) [vdsm.api] START repoStats(domains=()) from=::ffff:172.16.1.29,45646, task_id=daf4ce0b-f8e1-4067-af3a-47620f5b9491 (api:48) 2021-01-24 14:43:13,474-0700 INFO (jsonrpc/7) [vdsm.api] FINISH repoStats return={} from=::ffff:172.16.1.29,45646, task_id=daf4ce0b-f8e1-4067-af3a-47620f5b9491 (api:54) 2021-01-24 14:43:13,475-0700 INFO (jsonrpc/7) [vdsm.api] START multipath_health() from=::ffff:172.16.1.29,45646, task_id=14c53e2e-e304-46e6-b2bc-df8642fd31fd (api:48) 2021-01-24 14:43:13,476-0700 INFO (jsonrpc/7) [vdsm.api] FINISH multipath_health return={} from=::ffff:172.16.1.29,45646, task_id=14c53e2e-e304-46e6-b2bc-df8642fd31fd (api:54) 2021-01-24 14:43:13,488-0700 INFO (jsonrpc/7) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:172.16.1.29,45646 (api:54) 2021-01-24 14:43:13,898-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'dongle.co.slakin.net:/exports/00/storage', 'ipv6_enabled': 'false', 'id': '00000000-0000-0000-0000-000000000000', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:172.16.1.29,45646, flow_id=30b1a760-9117-4f1f-aea3-64b849f5e38c, task_id=e3c977b4-9994-438c-9fa8-f0af5ae49909 (api:48) 2021-01-24 14:43:13,902-0700 INFO (jsonrpc/3) [storage.StorageServer.MountConnection] Creating directory '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage' (storageServer:167) 2021-01-24 14:43:13,902-0700 INFO (jsonrpc/3) [storage.fileUtils] Creating directory: /rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage mode: None (fileUtils:201) 2021-01-24 14:43:13,903-0700 INFO (jsonrpc/3) [storage.Mount] mounting dongle.co.slakin.net:/exports/00/storage at /rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage (mount:207) 2021-01-24 14:43:14,082-0700 INFO (jsonrpc/3) [IOProcessClient] (dongle.co.slakin.net:_exports_00_storage) Closing client (__init__:641) 2021-01-24 14:43:14,083-0700 INFO (jsonrpc/3) [IOProcessClient] (23673598-92e5-4f9e-91de-1a38a03d7511) Closing client (__init__:641) 2021-01-24 14:43:14,086-0700 INFO (ioprocess/109656) [IOProcessClient] (dongle.co.slakin.net:_exports_00_storage) Killing ioprocess (__init__:191) 2021-01-24 14:43:14,086-0700 INFO (ioprocess/109665) [IOProcessClient] (23673598-92e5-4f9e-91de-1a38a03d7511) Killing ioprocess (__init__:191) 2021-01-24 14:43:14,089-0700 INFO (ioprocess/109665) [IOProcessClient] (23673598-92e5-4f9e-91de-1a38a03d7511) ioprocess was terminated by signal 9 (__init__:200) 2021-01-24 14:43:14,090-0700 INFO (ioprocess/109656) [IOProcessClient] (dongle.co.slakin.net:_exports_00_storage) ioprocess was terminated by signal 9 (__init__:200) 2021-01-24 14:43:14,147-0700 INFO (jsonrpc/3) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-24 14:43:14,147-0700 INFO (jsonrpc/3) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '00000000-0000-0000-0000-000000000000', 'status': 0}]} from=::ffff:172.16.1.29,45646, flow_id=30b1a760-9117-4f1f-aea3-64b849f5e38c, task_id=e3c977b4-9994-438c-9fa8-f0af5ae49909 (api:54) 2021-01-24 14:43:14,313-0700 INFO (jsonrpc/6) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'dongle.co.slakin.net:/exports/00/storage', 'ipv6_enabled': 'false', 'id': 'ca248090-f338-49b2-9a4d-9b3987a2f75e', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:172.16.1.29,45646, flow_id=1c30523d, task_id=31bd7156-d08a-47aa-8e90-a56175d188ba (api:48) 2021-01-24 14:43:14,319-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-24 14:43:14,319-0700 INFO (jsonrpc/6) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': 'ca248090-f338-49b2-9a4d-9b3987a2f75e', 'status': 0}]} from=::ffff:172.16.1.29,45646, flow_id=1c30523d, task_id=31bd7156-d08a-47aa-8e90-a56175d188ba (api:54) 2021-01-24 14:43:14,377-0700 INFO (jsonrpc/5) [vdsm.api] START createStorageDomain(storageType=1, sdUUID='a9ab8829-14bd-45cc-a57b-03b01e96f407', domainName='dongle_nfs', typeSpecificArg='dongle.co.slakin.net:/exports/00/storage', domClass=1, domVersion='5', block_size=512, max_hosts=250, options=None) from=::ffff:172.16.1.29,45646, flow_id=1c30523d, task_id=b9870145-e55b-4fa9-a07b-5167f763ec0d (api:48) 2021-01-24 14:43:14,378-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Refreshing storage domain cache (resize=True) (sdc:80) 2021-01-24 14:43:14,378-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-24 14:43:14,496-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI devices: 0.12 seconds (utils:390) 2021-01-24 14:43:14,497-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC devices (hba:60) 2021-01-24 14:43:14,767-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC devices: 0.27 seconds (utils:390) 2021-01-24 14:43:14,825-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing multipath devices (multipath:104) 2021-01-24 14:43:14,831-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing multipath devices: 0.01 seconds (utils:390) 2021-01-24 14:43:14,831-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Refreshing storage domain cache: 0.45 seconds (utils:390) 2021-01-24 14:43:14,832-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Looking up domain a9ab8829-14bd-45cc-a57b-03b01e96f407 (sdc:171) 2021-01-24 14:43:15,026-0700 WARN (jsonrpc/5) [storage.LVM] All 1 tries have failed: cmd=['/sbin/lvm', 'vgs', '--config', 'devices { preferred_names=["^/dev/mapper/"] ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 filter=["r|.*|"] hints="none" obtain_device_list_from_udev=0 } global { locking_type=1 prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { retain_min=50 retain_days=0 }', '--noheadings', '--units', 'b', '--nosuffix', '--separator', '|', '--ignoreskippedcluster', '-o', 'uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name', 'a9ab8829-14bd-45cc-a57b-03b01e96f407'] rc=5 err=[' Volume group "a9ab8829-14bd-45cc-a57b-03b01e96f407" not found', ' Cannot process volume group a9ab8829-14bd-45cc-a57b-03b01e96f407'] (lvm:538) 2021-01-24 14:43:15,032-0700 INFO (tmap-56/0) [IOProcessClient] (dongle.co.slakin.net:_exports_00_storage) Starting client (__init__:340) 2021-01-24 14:43:15,068-0700 INFO (ioprocess/109792) [IOProcess] (dongle.co.slakin.net:_exports_00_storage) Starting ioprocess (__init__:465) 2021-01-24 14:43:15,073-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] Looking up domain a9ab8829-14bd-45cc-a57b-03b01e96f407: 0.24 seconds (utils:390) 2021-01-24 14:43:15,124-0700 INFO (jsonrpc/5) [IOProcessClient] (a9ab8829-14bd-45cc-a57b-03b01e96f407) Starting client (__init__:340) 2021-01-24 14:43:15,164-0700 INFO (ioprocess/109802) [IOProcess] (a9ab8829-14bd-45cc-a57b-03b01e96f407) Starting ioprocess (__init__:465) 2021-01-24 14:43:15,193-0700 INFO (jsonrpc/5) [storage.StorageDomain] Creating domain metadata directory '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md' (fileSD:441) 2021-01-24 14:43:15,321-0700 INFO (jsonrpc/5) [storage.xlease] Formatting index for lockspace 'a9ab8829-14bd-45cc-a57b-03b01e96f407' (version=1) (xlease:701) 2021-01-24 14:43:15,460-0700 INFO (jsonrpc/5) [storage.StorageDomain] Creating domain images directory '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/images' (nfsSD:110) 2021-01-24 14:43:15,529-0700 INFO (jsonrpc/5) [storage.StorageDomain] Removing remnants of deleted images [] (fileSD:735) 2021-01-24 14:43:15,529-0700 INFO (jsonrpc/5) [storage.SANLock] Initializing sanlock for domain a9ab8829-14bd-45cc-a57b-03b01e96f407 path=/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md/ids alignment=1048576 block_size=512 io_timeout=10 (clusterlock:286) 2021-01-24 14:43:15,535-0700 ERROR (jsonrpc/5) [storage.SANLock] Cannot initialize lock for domain a9ab8829-14bd-45cc-a57b-03b01e96f407 (clusterlock:305) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 295, in initLock sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such device') 2021-01-24 14:43:15,535-0700 INFO (jsonrpc/5) [vdsm.api] FINISH createStorageDomain error=Could not initialize cluster lock: () from=::ffff:172.16.1.29,45646, flow_id=1c30523d, task_id=b9870145-e55b-4fa9-a07b-5167f763ec0d (api:52) 2021-01-24 14:43:15,536-0700 ERROR (jsonrpc/5) [storage.TaskManager.Task] (Task='b9870145-e55b-4fa9-a07b-5167f763ec0d') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 295, in initLock sector=self._block_size) sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such device') During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-123>", line 2, in createStorageDomain File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2669, in createStorageDomain max_hosts=max_hosts) File "/usr/lib/python3.6/site-packages/vdsm/storage/nfsSD.py", line 120, in create fsd.initSPMlease() File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 1019, in initSPMlease return self._manifest.initDomainLock() File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 627, in initDomainLock self._domainLock.initLock(self.getDomainLease()) File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line 306, in initLock raise se.ClusterLockInitError() vdsm.storage.exception.ClusterLockInitError: Could not initialize cluster lock: () 2021-01-24 14:43:15,536-0700 INFO (jsonrpc/5) [storage.TaskManager.Task] (Task='b9870145-e55b-4fa9-a07b-5167f763ec0d') aborting: Task is aborted: 'value=Could not initialize cluster lock: () abortedcode=701' (task:1190) 2021-01-24 14:43:15,537-0700 ERROR (jsonrpc/5) [storage.Dispatcher] FINISH createStorageDomain error=Could not initialize cluster lock: () (dispatcher:83) 2021-01-24 14:43:15,537-0700 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC call StorageDomain.create failed (error 701) in 1.16 seconds (__init__:312) 2021-01-24 14:43:15,708-0700 INFO (jsonrpc/0) [vdsm.api] START disconnectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'dongle.co.slakin.net:/exports/00/storage', 'ipv6_enabled': 'false', 'id': 'ca248090-f338-49b2-9a4d-9b3987a2f75e', 'user': '', 'tpgt': '1'}], options=None) from=::ffff:172.16.1.29,45646, flow_id=169aaae0-c187-4089-9b90-a8ddeec48dc2, task_id=4363b864-5954-4bac-99d3-e197e6d12400 (api:48) 2021-01-24 14:43:15,709-0700 INFO (jsonrpc/0) [storage.Mount] unmounting /rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage (mount:215) 2021-01-24 14:43:15,835-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Refreshing storage domain cache (resize=False) (sdc:80) 2021-01-24 14:43:15,836-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI devices (iscsi:442) 2021-01-24 14:43:15,952-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI devices: 0.12 seconds (utils:390) 2021-01-24 14:43:15,952-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC devices (hba:60) 2021-01-24 14:43:16,223-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC devices: 0.27 seconds (utils:390) 2021-01-24 14:43:16,283-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] Refreshing storage domain cache: 0.45 seconds (utils:390) 2021-01-24 14:43:16,284-0700 INFO (jsonrpc/0) [vdsm.api] FINISH disconnectStorageServer return={'statuslist': [{'id': 'ca248090-f338-49b2-9a4d-9b3987a2f75e', 'status': 0}]} from=::ffff:172.16.1.29,45646, flow_id=169aaae0-c187-4089-9b90-a8ddeec48dc2, task_id=4363b864-5954-4bac-99d3-e197e6d12400 (api:54) ^C supervdsm.log shows a bit more info [root@brick vdsm]# cat supervdsm.log MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:10,237::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:10,237::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:10,269::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:10,270::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|jsonrpc/3::DEBUG::2021-01-24 14:43:13,904::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call mount with (<vdsm.supervdsm_server._SuperVdsm object at 0x7ff15f423588>, 'dongle.co.slakin.net:/exports/00/storage', '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage') {'mntOpts': 'soft,nosharecache,timeo=100,retrans=3,nolock', 'vfstype': 'nfs', 'cgroup': None} MainProcess|jsonrpc/3::DEBUG::2021-01-24 14:43:13,904::commands::211::root::(execCmd) /usr/bin/taskset --cpu-list 0-3 /usr/bin/mount -t nfs -o soft,nosharecache,timeo=100,retrans=3,nolock dongle.co.slakin.net:/exports/00/storage /rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage (cwd None) MainProcess|jsonrpc/3::DEBUG::2021-01-24 14:43:14,020::commands::224::root::(execCmd) SUCCESS: <err> = b''; <rc> = 0 MainProcess|jsonrpc/3::DEBUG::2021-01-24 14:43:14,021::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return mount with None MainProcess|jsonrpc/3::DEBUG::2021-01-24 14:43:14,093::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7ff15f423588>, 'vdsm', ('kvm',), '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage', 7) {} MainProcess|jsonrpc/3::DEBUG::2021-01-24 14:43:14,114::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/3::DEBUG::2021-01-24 14:43:14,116::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7ff15f423588>, 'qemu', ('qemu', 'kvm'), '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage', 5) {} MainProcess|jsonrpc/3::DEBUG::2021-01-24 14:43:14,142::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/5::DEBUG::2021-01-24 14:43:14,498::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7ff15f423588>,) {} MainProcess|jsonrpc/5::DEBUG::2021-01-24 14:43:14,498::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) MainProcess|jsonrpc/5::DEBUG::2021-01-24 14:43:14,766::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None MainProcess|jsonrpc/5::DEBUG::2021-01-24 14:43:15,080::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7ff15f423588>, 'vdsm', ('kvm',), '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage', 7) {} MainProcess|jsonrpc/5::DEBUG::2021-01-24 14:43:15,100::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/5::DEBUG::2021-01-24 14:43:15,102::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at 0x7ff15f423588>, 'qemu', ('qemu', 'kvm'), '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage', 5) {} MainProcess|jsonrpc/5::DEBUG::2021-01-24 14:43:15,123::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return validateAccess with None MainProcess|jsonrpc/0::DEBUG::2021-01-24 14:43:15,710::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call umount with (<vdsm.supervdsm_server._SuperVdsm object at 0x7ff15f423588>, '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage') {'force': True, 'lazy': True, 'freeloop': False} MainProcess|jsonrpc/0::DEBUG::2021-01-24 14:43:15,711::commands::211::root::(execCmd) /usr/bin/taskset --cpu-list 0-3 /usr/bin/umount -f -l /rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage (cwd None) MainProcess|jsonrpc/0::DEBUG::2021-01-24 14:43:15,781::commands::224::root::(execCmd) SUCCESS: <err> = b''; <rc> = 0 MainProcess|jsonrpc/0::DEBUG::2021-01-24 14:43:15,781::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return umount with None MainProcess|jsonrpc/0::DEBUG::2021-01-24 14:43:15,953::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7ff15f423588>,) {} MainProcess|jsonrpc/0::DEBUG::2021-01-24 14:43:15,954::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) MainProcess|jsonrpc/0::DEBUG::2021-01-24 14:43:16,222::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return hbaRescan with None MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:20,272::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:20,272::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:20,304::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:20,305::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:30,307::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:30,307::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:30,307::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:30,340::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:30,340::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:40,342::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:40,342::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:40,374::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:40,374::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:50,376::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:50,376::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:50,408::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:43:50,408::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:00,410::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:00,410::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:00,446::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:00,447::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:10,449::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:10,449::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:10,481::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:10,482::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:20,484::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:20,484::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:20,517::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:20,517::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:30,519::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:30,519::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:30,519::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:30,550::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:30,550::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:40,552::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:40,552::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:40,586::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:40,587::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:50,588::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:50,589::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:50,619::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:44:50,620::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:00,621::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:00,622::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:00,656::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:00,657::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:10,659::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:10,659::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:10,692::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:10,693::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:20,694::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:20,695::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:20,726::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:20,726::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:30,728::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:30,728::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:30,729::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:30,759::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:30,759::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:40,761::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:40,761::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:40,791::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:40,792::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:50,794::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:50,794::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:50,826::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:45:50,827::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:00,828::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:00,829::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:00,864::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:00,865::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:10,867::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:10,867::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:10,900::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:10,900::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:20,902::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:20,903::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:20,932::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:20,933::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:30,934::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:30,935::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:30,935::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:30,967::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:30,968::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:40,969::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:40,970::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:41,001::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:41,001::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:51,003::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:51,003::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:51,036::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:46:51,036::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:01,038::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:01,038::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:01,073::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:01,074::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:11,075::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:11,076::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:11,107::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:11,108::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:21,110::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:21,110::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:21,140::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:21,140::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:31,142::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:31,142::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:31,142::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:31,174::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:31,175::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:41,176::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:41,177::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:41,207::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:41,207::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:51,209::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:51,210::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:51,241::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:47:51,242::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:01,243::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:01,244::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:01,279::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:01,280::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:11,282::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:11,282::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:11,314::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:11,314::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:21,316::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:21,316::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:21,346::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:21,347::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:31,349::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:31,349::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:31,349::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:31,382::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:31,383::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:41,385::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:41,385::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:41,415::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:41,415::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:51,417::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:51,418::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:51,449::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:48:51,449::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:01,451::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:01,451::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:01,487::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:01,488::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:11,490::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:11,490::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:11,521::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:11,521::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:21,523::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:21,524::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:21,554::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:21,554::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:31,556::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call dmsetup_run_status with ('multipath',) {} MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:31,556::logutils::354::root::(_report_stats) ThreadedHandler is ok in the last 60 seconds (max pending: 2) MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:31,557::commands::153::common.commands::(start) /usr/bin/taskset --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:31,589::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> = 0 MainProcess|mpathhealth::DEBUG::2021-01-24 14:49:31,590::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) return dmsetup_run_status with b'' And last but not least is sanlock.log which shows the smoking gun: [root@brick log]# tail sanlock.log 2021-01-24 14:34:40 226677 [1366]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage/9aa78da8-c391-46eb-9fcd-426734b6f431/dom_md/ids 2021-01-24 14:34:40 226677 [1366]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-24 14:43:15 227192 [1365]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage/a9ab8829-14bd-45cc-a57b-03b01e96f407/dom_md/ids 2021-01-24 14:43:15 227192 [1365]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-24 14:55:43 227940 [1366]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage/adecf2ee-53d0-4ccd-a02c-ac1f4edf3ffa/dom_md/ids 2021-01-24 14:55:43 227940 [1366]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. 2021-01-24 15:05:47 228543 [1328]: helper pid 1333 term signal 15 2021-01-24 15:05:47 228544 [120124]: sanlock daemon started 3.8.2 host 3b903780-4f79-1018-816e-aeb2724778a7 (brick.co.slakin.net) 2021-01-24 15:07:52 228669 [120129]: open error -13 EACCES: no permission to open /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage/01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids 2021-01-24 15:07:52 228669 [120129]: check that daemon user sanlock 179 group sanlock 179 has access to disk or file. [root@brick log]# I re-mounted the NFS share; switched to sanlock user; I can see sanlock is in the kvm group but cannot open the file. [root@brick log]# mkdir /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage [root@brick log]# mount -t nfs -o rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.58 dongle.co.slakin.net:/exports/00/storage /rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage [root@brick log]# su - sanlock -s /bin/bash Last login: Sun Jan 24 14:58:42 MST 2021 on pts/0 nodectl must be run as root! nodectl must be run as root! [sanlock@brick ~]$ cd /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage [sanlock@brick dongle.co.slakin.net:_exports_00_storage]$ ls -l /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage/01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids -rw-rw----. 1 vdsm kvm 0 Jan 24 15:07 /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage/01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids [sanlock@brick dongle.co.slakin.net:_exports_00_storage]$ id uid=179(sanlock) gid=179(sanlock) groups=179(sanlock),6(disk),36(kvm),107(qemu) context=unconfined_u:unconfined_r:unconfined_t:s0-s0:c0.c1023 [sanlock@brick dongle.co.slakin.net:_exports_00_storage]$ cat /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage/01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids cat: '/rhev/data-center/mnt/dongle.co.slakin.net:_exports_00_storage/01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids': Permission denied [sanlock@brick dongle.co.slakin.net:_exports_00_storage]$ Switch over to the NFS server (dongle) and change other/world permissions: root@dongle:/exports/00/storage# chmod o+r 01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids root@dongle:/exports/00/storage# ls -l !$ ls -l 01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids -rw-rw-r-- 1 36 36 0 Jan 24 15:07 01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids root@dongle:/exports/00/storage# After changing permissions on the files, sanlock on brick(the oVirt node) can now cat the file [sanlock@brick dongle.co.slakin.net:_exports_00_storage]$ cat /rhev/data-center/mnt/dongle.co.slakin.net: _exports_00_storage/01bfd183-b46e-432d-a5e4-f0cf51e5df6e/dom_md/ids [sanlock@brick dongle.co.slakin.net:_exports_00_storage]$ echo $? 0 [sanlock@brick dongle.co.slakin.net:_exports_00_storage]$ So in short, even though sanlock is member to gid 36(kvm) and the file ownership and permissions appear to be set correctly, sanlock cannot read files that it should be able to. ..Matt On Thu, Jan 21, 2021 at 6:27 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Wed, Jan 20, 2021 at 7:36 PM Matt Snow <mattsnow@gmail.com> wrote:
On Wed, Jan 20, 2021 at 2:46 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 18, 2021 at 8:58 AM Matt Snow <mattsnow@gmail.com> wrote:
I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the
I tried to do some debugging on my own before reaching out to this
ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4
cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but I believe it meets the minimum requirements.
NFS server: * Ubuntu 19.10 w/ ZFS share w/ 17TB available space.
We don't test ZFS, but since you tried also non-ZFS server with same issue, the issue is probably not ZFS.
* NFS share settings are just 'rw=@172.16.1.0/24' but have also
You are missing anonuid=36,anongid=36. This should not affect sanlock,
but
you will have issues with libvirt without this.
I found this setting in another email thread and have applied it
root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs
tanker/ovirt sharenfs rw,anonuid=36,anongid=36 local
root@stumpy:/tanker/ovirt/host_storage#
Here is working export from my test system:
/export/00 *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
I have updated both ZFS server and the Ubuntu 20.04 NFS server to use
same problem with both versions. The issue occurs in both cockpit UI and tmux'd CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is created and running. list. Any help is much appreciated! tried 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure' the above settings.
# ZFS Server
root@stumpy:/tanker/ovirt/host_storage# zfs set sharenfs="rw,sync,anonuid=36,anongid=36,no_subtree_check" tanker/ovirt
root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs
tanker/ovirt sharenfs rw,sync,anonuid=36,anongid=36,no_subtree_check local
root@stumpy:/tanker/ovirt/host_storage#
# Ubuntu laptop
ls -ld /exports/host_storage
drwxr-xr-x 2 36 36 4096 Jan 16 19:42 /exports/host_storage
root@msnowubntlt:/exports/host_storage# showmount -e dongle
Export list for dongle:
/exports/host_storage *
root@msnowubntlt:/exports/host_storage# grep host_storage /etc/exports
/exports/host_storage *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
root@msnowubntlt:/exports/host_storage#
Interesting point - upon being re-prompted to configure storage after initial failure I provided *different* NFS server information (the ubuntu laptop) **snip**
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."}
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs
Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]:
Please specify the full shared storage connection path to use (example: host:/path): dongle:/exports/host_storage
If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []:
[ INFO ] Creating Storage Domain
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps]
**snip**
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."}
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
**snip** Upon checking mounts I see that the original server is still being used.
[root@brick ~]# mount | grep nfs
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime)
stumpy:/tanker/ovirt/host_storage on /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50)
Using node and hosted engine makes everything more complicated and harder to debug.
I would create an engine vm for testing on some machine, and add this host as regular host. It will be easier to debug this way, since you don't have to reinstall everything from scratch on every failure.
Once we sort out the issue with the NFS/ZFS storage, using node and hosted engine makes sense for better availability and saving hardware resources.
Also, can you explain how to build and configure identical ubuntu storage server? I can try to test a vm if you give detailed description.
Maybe this an issue with AppArmor?
https://linuxconfig.org/how-to-disable-enable-selinux-on-ubuntu-20-04-focal-...
Did you try to disable it on the storage server, or enable selinux?
[root@brick ~]#
* The target directory is always empty and chown'd 36:36 with 0755
Correct.
* I have tried using both IP and DNS names. forward and reverse DNS
works from ovirt host and other systems on the network.
* The NFS share always gets mounted successfully on the ovirt node system. * I have tried auto and v3 NFS versions in other various combinations. * I have also tried setting up an NFS server on a non-ZFS backed storage system that is open to any host and get the same errors as shown below. * I ran nfs-check.py script without issue against both NFS servers and followed other verification steps listed on https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html
***Snip from ovirt-hosted-engine-setup*** Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]: Please specify the full shared storage connection path to use (example: host:/path): stumpy.mydomain.com:/tanker/ovirt/host_storage If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: rw [ INFO ] Creating Storage Domain [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Force facts gathering] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the storage interface to be up] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check local VM dir stat] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Enforce local VM
[ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch host facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter name] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add NFS storage domain] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add glusterfs storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add iSCSI storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add Fibre Channel storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get storage domain
[ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the appliance OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Parse OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get required size] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Remove unsuitable storage domain] [ INFO ] skipping: [localhost] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check storage domain free space] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
***End snippet**
The relevant section from /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-ansible-create_storage_domain-20210116195220-bnfg1w.log: **Snip*** 2021-01-16 19:53:56,010-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a04c8320> kwargs 2021-01-16 19:53:57,219-0700 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Activate storage domain'} 2021-01-16 19:53:57,220-0700 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs is_conditional:False 2021-01-16 19:53:57,221-0700 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs 2021-01-16 19:54:00,346-0700 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,347-0700 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,348-0700 DEBUG var changed: host "localhost" var "play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,349-0700 ERROR ansible failed { "ansible_host": "localhost", "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_result": { "_ansible_no_log": false, "changed": false, "exception": "Traceback (most recent call last):\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\",
se)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n self._raise_error(response, body)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n", "invocation": { "module_args": { "backup": null, "comment": null, "critical_space_action_blocker": null, "data_center": "Default", "description": null, "destroy": null, "discard_after_delete": null, "domain_function": "data", "fcp": null, "fetch_nested": false, "format": null, "glusterfs": null, "host": "brick.mydomain.com", "id": null, "iscsi": null, "localfs": null, "managed_block_storage": null, "name": "hosted_storage", "nested_attributes": [], "nfs": null, "poll_interval": 3, "posixfs": null, "state": "present", "timeout": 180, "wait": true, "warning_low_space": null, "wipe_after_delete": null } }, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400." }, "ansible_task": "Activate storage domain", "ansible_type": "task", "status": "FAILED", "task_duration": 4 } 2021-01-16 19:54:00,350-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a05f2d30> kwargs ignore_errors:None 2021-01-16 19:54:00,358-0700 INFO ansible stats { "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_playbook_duration": "01:35 Minutes", "ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 23, 'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 7, 'rescued': 0, 'ignored': 0}}", "ansible_type": "finish", "status": "FAILED" } 2021-01-16 19:54:00,359-0700 INFO SUMMARY: Duration Task Name -------- -------- [ < 1 sec ] Execute just a specific set of steps [ 00:05 ] Force facts gathering [ 00:03 ] Check local VM dir stat [ 00:04 ] Obtain SSO token using username/password credentials [ 00:04 ] Fetch host facts [ 00:02 ] Fetch cluster ID [ 00:04 ] Fetch cluster facts [ 00:04 ] Fetch Datacenter facts [ 00:02 ] Fetch Datacenter ID [ 00:02 ] Fetch Datacenter name [ 00:04 ] Add NFS storage domain [ 00:04 ] Get storage domain details [ 00:03 ] Find the appliance OVF [ 00:03 ] Parse OVF [ 00:02 ] Get required size [ FAILED ] Activate storage domain 2021-01-16 19:54:00,359-0700 DEBUG ansible on_any args <ansible.executor.stats.AggregateStats object at 0x7f11a2e3b8d0> kwargs **End snip**
The relevantsection from /var/log/vdsm/vdsm.log: ***begin snip*** 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': '1'}],
2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'status': 0}]} from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:54) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [vdsm.api] START getStorageDomainInfo(sdUUID='54532dd4-3e5b-4885-b88e-599c81efb146',
2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [vdsm.api] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:52) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-131>", line 2, in getStorageDomainInfo File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2796, in getStorageDomainInfo dom = self.validateSdUUID(sdUUID) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 312, in validateSdUUID sdDom.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.Dispatcher] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,919-0700 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call StorageDomain.getInfo failed (error 379) in 0.00 seconds (__init__:312) 2021-01-16 19:53:58,956-0700 INFO (jsonrpc/1) [vdsm.api] START createStoragePool(poolType=None, spUUID='91fb3f22-5795-11eb-ad9f-00163e3f4683', poolName='Default', masterDom='54532dd4-3e5b-4885-b88e-599c81efb146', domList=['54532dd4-3e5b-4885-b88e-599c81efb146'], masterVersion=7, lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, ioOpTimeoutSec=10, leaseRetries=3, options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:48) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] updating pool 91fb3f22-5795-11eb-ad9f-00163e3f4683 backend from type NoneType instance 0x7f86e45f89d0 to type StoragePoolDiskBackend instance 0x7f8680738408 (sp:157) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] spUUID=91fb3f22-5795-11eb-ad9f-00163e3f4683 poolName=Default master_sd=54532dd4-3e5b-4885-b88e-599c81efb146 domList=['54532dd4-3e5b-4885-b88e-599c81efb146'] masterVersion=7 {'LEASERETRIES': 3, 'LEASETIMESEC': 60, 'LOCKRENEWALINTERVALSEC': 5, 'IOOPTIMEOUTSEC': 10} (sp:602) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.StoragePool] Unexpected error (sp:618) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
2021-01-16 19:53:58,963-0700 INFO (jsonrpc/1) [vdsm.api] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:52) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line
887, in _run
return fn(*args, **kargs) File "<decorator-gen-31>", line 2, in createStoragePool File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line
50, in method
ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line
1027, in createStoragePool
leaseParams) File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line
619, in create
raise se.StorageDomainAccessError(sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
permissions. dir existence] details] line 783, in main\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 638, in post_create_check\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 3647, in add\n return self._internal_add(storage_domain, headers, query, wait)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n return future.wait() if wait else future\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in wait\n return self._code(response)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in callback\n self._check_fault(respon options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48) options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:48) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',)
2021-01-16 19:53:58,964-0700 INFO (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,964-0700 ERROR (jsonrpc/1) [storage.Dispatcher] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,965-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call StoragePool.create failed (error 379) in 0.01 seconds (__init__:312) 2021-01-16 19:54:01,923-0700 INFO (jsonrpc/5) [api.host] START getStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,964-0700 INFO (jsonrpc/7) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,966-0700 INFO (jsonrpc/7) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) 2021-01-16 19:54:01,977-0700 INFO (jsonrpc/5) [vdsm.api] START repoStats(domains=()) from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:48) 2021-01-16 19:54:01,978-0700 INFO (jsonrpc/5) [vdsm.api] FINISH repoStats return={} from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:54) 2021-01-16 19:54:01,980-0700 INFO (jsonrpc/5) [vdsm.api] START multipath_health() from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:48) 2021-01-16 19:54:01,981-0700 INFO (jsonrpc/5) [vdsm.api] FINISH multipath_health return={} from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:54) 2021-01-16 19:54:01,996-0700 INFO (jsonrpc/5) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) ***end snip*** _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/G2TQC6XTJTIBAI...

Can you try to alter the share options like this (check for typos as I am typing by memory):anonuid=36,anongid=36, all_squash And of course export a fresh one: Best Regards,Strahil Nikolov Sent from Yahoo Mail on Android On Mon, Jan 25, 2021 at 0:51, Matt Snow<mattsnow@gmail.com> wrote: _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/BLFODDM63I6IRU...

On Mon, Jan 25, 2021 at 12:33 AM Matt Snow <mattsnow@gmail.com> wrote: I reproduced the issue with ubuntu server 20.04 nfs server. The root cause is this setting in /etc/default/nfs-kernel-server: RPCMOUNTDOPTS="--manage-gids" Looking at https://bugs.launchpad.net/ubuntu/+source/nfs-utils/+bug/1454112 it seems that the purpose of this option is to ignore client groups, which breaks ovirt. After removing this option: RPCMOUNTDOPTS="" And restarting nfs-kernel-server service, creating storage domain works. You can check with Ubuntu folks why they are enabling this configuration by default, and if disabling it has any unwanted side effects. Nir

I can confirm that removing "--manage-gids" flag from RPCMOUNTDOPTS in /etc/default/nfs-kernel-server now allows me to add the ZFS backed NFS share. From the man page <https://linux.die.net/man/8/rpc.mountd>: -g or --manage-gids Accept requests from the kernel to map user id numbers into lists of group id numbers for use in access control. An NFS request will normally (except when using Kerberos or other cryptographic authentication) contains a user-id and a list of group-ids. Due to a limitation in the NFS protocol, at most 16 groups ids can be listed. If you use the -g flag, then the list of group ids received from the client will be replaced by a list of group ids determined by an appropriate lookup on the server. Note that the 'primary' group id is not affected so a newgroup command on the client will still be effective. This function requires a Linux Kernel with version at least 2.6.21. I speculate that if I had directory services setup and the NFS server directed there, this would be a non-issue. Thank you so much for your help on this, Nir & Team! On Mon, Jan 25, 2021 at 1:00 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 25, 2021 at 12:33 AM Matt Snow <mattsnow@gmail.com> wrote:
I reproduced the issue with ubuntu server 20.04 nfs server.
The root cause is this setting in /etc/default/nfs-kernel-server: RPCMOUNTDOPTS="--manage-gids"
Looking at https://bugs.launchpad.net/ubuntu/+source/nfs-utils/+bug/1454112 it seems that the purpose of this option is to ignore client groups, which breaks ovirt.
After removing this option: RPCMOUNTDOPTS=""
And restarting nfs-kernel-server service, creating storage domain works.
You can check with Ubuntu folks why they are enabling this configuration by default, and if disabling it has any unwanted side effects.
Nir

On Mon, Jan 25, 2021 at 7:23 PM Matt Snow <mattsnow@gmail.com> wrote:
I can confirm that removing "--manage-gids" flag from RPCMOUNTDOPTS in /etc/default/nfs-kernel-server now allows me to add the ZFS backed NFS share.
From the man page:
-g or --manage-gids Accept requests from the kernel to map user id numbers into lists of group id numbers for use in access control. An NFS request will normally (except when using Kerberos or other cryptographic authentication) contains a user-id and a list of group-ids. Due to a limitation in the NFS protocol, at most 16 groups ids can be listed. If you use the -g flag, then the list of group ids received from the client will be replaced by a list of group ids determined by an appropriate lookup on the server. Note that the 'primary' group id is not affected so a newgroup command on the client will still be effective. This function requires a Linux Kernel with version at least 2.6.21.
I speculate that if I had directory services setup and the NFS server directed there, this would be a non-issue.
Thank you so much for your help on this, Nir & Team!
You can contribute by updating the nfs troubleshooting page with this info: https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html See the link "Edit this page" in the bottom of the page.
On Mon, Jan 25, 2021 at 1:00 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 25, 2021 at 12:33 AM Matt Snow <mattsnow@gmail.com> wrote:
I reproduced the issue with ubuntu server 20.04 nfs server.
The root cause is this setting in /etc/default/nfs-kernel-server: RPCMOUNTDOPTS="--manage-gids"
Looking at https://bugs.launchpad.net/ubuntu/+source/nfs-utils/+bug/1454112 it seems that the purpose of this option is to ignore client groups, which breaks ovirt.
After removing this option: RPCMOUNTDOPTS=""
And restarting nfs-kernel-server service, creating storage domain works.
You can check with Ubuntu folks why they are enabling this configuration by default, and if disabling it has any unwanted side effects.
Nir

Pull request sent. On Mon, Jan 25, 2021 at 11:21 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 25, 2021 at 7:23 PM Matt Snow <mattsnow@gmail.com> wrote:
I can confirm that removing "--manage-gids" flag from RPCMOUNTDOPTS in
/etc/default/nfs-kernel-server now allows me to add the ZFS backed NFS share.
From the man page:
-g or --manage-gids Accept requests from the kernel to map user id numbers into lists of
group id numbers for use in access control. An NFS request will normally (except when using Kerberos or other cryptographic authentication) contains a user-id and a list of group-ids. Due to a limitation in the NFS protocol, at most 16 groups ids can be listed. If you use the -g flag, then the list of group ids received from the client will be replaced by a list of group ids determined by an appropriate lookup on the server. Note that the 'primary' group id is not affected so a newgroup command on the client will still be effective. This function requires a Linux Kernel with version at least 2.6.21.
I speculate that if I had directory services setup and the NFS server
directed there, this would be a non-issue.
Thank you so much for your help on this, Nir & Team!
You can contribute by updating the nfs troubleshooting page with this info: https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html
See the link "Edit this page" in the bottom of the page.
On Mon, Jan 25, 2021 at 1:00 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 25, 2021 at 12:33 AM Matt Snow <mattsnow@gmail.com> wrote:
I reproduced the issue with ubuntu server 20.04 nfs server.
The root cause is this setting in /etc/default/nfs-kernel-server: RPCMOUNTDOPTS="--manage-gids"
Looking at
https://bugs.launchpad.net/ubuntu/+source/nfs-utils/+bug/1454112
it seems that the purpose of this option is to ignore client groups, which breaks ovirt.
After removing this option: RPCMOUNTDOPTS=""
And restarting nfs-kernel-server service, creating storage domain works.
You can check with Ubuntu folks why they are enabling this configuration by default, and if disabling it has any unwanted side effects.
Nir

Hi Nir, as a test I setup a CentOS 8.3 cloud VM and exported an NFS share. It worked as expected! Clearly something is wrong with how Ubuntu exports NFS shares regardless of ZFS or other filesystems. Unfortunately that is where my storage is. =\ On Thu, Jan 21, 2021 at 6:27 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Wed, Jan 20, 2021 at 7:36 PM Matt Snow <mattsnow@gmail.com> wrote:
On Wed, Jan 20, 2021 at 2:46 AM Nir Soffer <nsoffer@redhat.com> wrote:
On Mon, Jan 18, 2021 at 8:58 AM Matt Snow <mattsnow@gmail.com> wrote:
I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the
I tried to do some debugging on my own before reaching out to this
ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4
cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but I believe it meets the minimum requirements.
NFS server: * Ubuntu 19.10 w/ ZFS share w/ 17TB available space.
We don't test ZFS, but since you tried also non-ZFS server with same issue, the issue is probably not ZFS.
* NFS share settings are just 'rw=@172.16.1.0/24' but have also
You are missing anonuid=36,anongid=36. This should not affect sanlock,
but
you will have issues with libvirt without this.
I found this setting in another email thread and have applied it
root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs
tanker/ovirt sharenfs rw,anonuid=36,anongid=36 local
root@stumpy:/tanker/ovirt/host_storage#
Here is working export from my test system:
/export/00 *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
I have updated both ZFS server and the Ubuntu 20.04 NFS server to use
same problem with both versions. The issue occurs in both cockpit UI and tmux'd CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is created and running. list. Any help is much appreciated! tried 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure' the above settings.
# ZFS Server
root@stumpy:/tanker/ovirt/host_storage# zfs set sharenfs="rw,sync,anonuid=36,anongid=36,no_subtree_check" tanker/ovirt
root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs
tanker/ovirt sharenfs rw,sync,anonuid=36,anongid=36,no_subtree_check local
root@stumpy:/tanker/ovirt/host_storage#
# Ubuntu laptop
ls -ld /exports/host_storage
drwxr-xr-x 2 36 36 4096 Jan 16 19:42 /exports/host_storage
root@msnowubntlt:/exports/host_storage# showmount -e dongle
Export list for dongle:
/exports/host_storage *
root@msnowubntlt:/exports/host_storage# grep host_storage /etc/exports
/exports/host_storage *(rw,sync,anonuid=36,anongid=36,no_subtree_check)
root@msnowubntlt:/exports/host_storage#
Interesting point - upon being re-prompted to configure storage after initial failure I provided *different* NFS server information (the ubuntu laptop) **snip**
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."}
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs
Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]:
Please specify the full shared storage connection path to use (example: host:/path): dongle:/exports/host_storage
If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []:
[ INFO ] Creating Storage Domain
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps]
**snip**
[ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."}
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
**snip** Upon checking mounts I see that the original server is still being used.
[root@brick ~]# mount | grep nfs
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime)
stumpy:/tanker/ovirt/host_storage on /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50)
Using node and hosted engine makes everything more complicated and harder to debug.
I would create an engine vm for testing on some machine, and add this host as regular host. It will be easier to debug this way, since you don't have to reinstall everything from scratch on every failure.
Once we sort out the issue with the NFS/ZFS storage, using node and hosted engine makes sense for better availability and saving hardware resources.
Also, can you explain how to build and configure identical ubuntu storage server? I can try to test a vm if you give detailed description.
Maybe this an issue with AppArmor?
https://linuxconfig.org/how-to-disable-enable-selinux-on-ubuntu-20-04-focal-...
Did you try to disable it on the storage server, or enable selinux?
[root@brick ~]#
* The target directory is always empty and chown'd 36:36 with 0755
Correct.
* I have tried using both IP and DNS names. forward and reverse DNS
works from ovirt host and other systems on the network.
* The NFS share always gets mounted successfully on the ovirt node system. * I have tried auto and v3 NFS versions in other various combinations. * I have also tried setting up an NFS server on a non-ZFS backed storage system that is open to any host and get the same errors as shown below. * I ran nfs-check.py script without issue against both NFS servers and followed other verification steps listed on https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html
***Snip from ovirt-hosted-engine-setup*** Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs Please specify the nfs version you would like to use (auto, v3, v4, v4_0, v4_1, v4_2)[auto]: Please specify the full shared storage connection path to use (example: host:/path): stumpy.mydomain.com:/tanker/ovirt/host_storage If needed, specify additional mount options for the connection to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: rw [ INFO ] Creating Storage Domain [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set of steps] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Force facts gathering] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the storage interface to be up] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check local VM dir stat] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Enforce local VM
[ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using username/password credentials] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch host facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter facts] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter ID] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter name] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add NFS storage domain] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add glusterfs storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add iSCSI storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add Fibre Channel storage domain] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get storage domain
[ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the appliance OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Parse OVF] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get required size] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Remove unsuitable storage domain] [ INFO ] skipping: [localhost] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check storage domain free space] [ INFO ] skipping: [localhost] [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]:
***End snippet**
The relevant section from /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-ansible-create_storage_domain-20210116195220-bnfg1w.log: **Snip*** 2021-01-16 19:53:56,010-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a04c8320> kwargs 2021-01-16 19:53:57,219-0700 INFO ansible task start {'status': 'OK', 'ansible_type': 'task', 'ansible_playbook': '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Activate storage domain'} 2021-01-16 19:53:57,220-0700 DEBUG ansible on_any args TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs is_conditional:False 2021-01-16 19:53:57,221-0700 DEBUG ansible on_any args localhost TASK: ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs 2021-01-16 19:54:00,346-0700 DEBUG var changed: host "localhost" var "ansible_play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,347-0700 DEBUG var changed: host "localhost" var "ansible_play_batch" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,348-0700 DEBUG var changed: host "localhost" var "play_hosts" type "<class 'list'>" value: "[]" 2021-01-16 19:54:00,349-0700 ERROR ansible failed { "ansible_host": "localhost", "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_result": { "_ansible_no_log": false, "changed": false, "exception": "Traceback (most recent call last):\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\",
se)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in _check_fault\n self._raise_error(response, body)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in _raise_error\n raise error\novirtsdk4.Error: Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400.\n", "invocation": { "module_args": { "backup": null, "comment": null, "critical_space_action_blocker": null, "data_center": "Default", "description": null, "destroy": null, "discard_after_delete": null, "domain_function": "data", "fcp": null, "fetch_nested": false, "format": null, "glusterfs": null, "host": "brick.mydomain.com", "id": null, "iscsi": null, "localfs": null, "managed_block_storage": null, "name": "hosted_storage", "nested_attributes": [], "nfs": null, "poll_interval": 3, "posixfs": null, "state": "present", "timeout": 180, "wait": true, "warning_low_space": null, "wipe_after_delete": null } }, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400." }, "ansible_task": "Activate storage domain", "ansible_type": "task", "status": "FAILED", "task_duration": 4 } 2021-01-16 19:54:00,350-0700 DEBUG ansible on_any args <ansible.executor.task_result.TaskResult object at 0x7f11a05f2d30> kwargs ignore_errors:None 2021-01-16 19:54:00,358-0700 INFO ansible stats { "ansible_playbook": "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", "ansible_playbook_duration": "01:35 Minutes", "ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 23, 'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 7, 'rescued': 0, 'ignored': 0}}", "ansible_type": "finish", "status": "FAILED" } 2021-01-16 19:54:00,359-0700 INFO SUMMARY: Duration Task Name -------- -------- [ < 1 sec ] Execute just a specific set of steps [ 00:05 ] Force facts gathering [ 00:03 ] Check local VM dir stat [ 00:04 ] Obtain SSO token using username/password credentials [ 00:04 ] Fetch host facts [ 00:02 ] Fetch cluster ID [ 00:04 ] Fetch cluster facts [ 00:04 ] Fetch Datacenter facts [ 00:02 ] Fetch Datacenter ID [ 00:02 ] Fetch Datacenter name [ 00:04 ] Add NFS storage domain [ 00:04 ] Get storage domain details [ 00:03 ] Find the appliance OVF [ 00:03 ] Parse OVF [ 00:02 ] Get required size [ FAILED ] Activate storage domain 2021-01-16 19:54:00,359-0700 DEBUG ansible on_any args <ansible.executor.stats.AggregateStats object at 0x7f11a2e3b8d0> kwargs **End snip**
The relevantsection from /var/log/vdsm/vdsm.log: ***begin snip*** 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:726) 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START connectStorageServer(domType=1, spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': '1'}],
2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [vdsm.api] FINISH connectStorageServer return={'statuslist': [{'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'status': 0}]} from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:54) 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [vdsm.api] START getStorageDomainInfo(sdUUID='54532dd4-3e5b-4885-b88e-599c81efb146',
2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [vdsm.api] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:52) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in _run return fn(*args, **kargs) File "<decorator-gen-131>", line 2, in getStorageDomainInfo File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2796, in getStorageDomainInfo dom = self.validateSdUUID(sdUUID) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 312, in validateSdUUID sdDom.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [storage.TaskManager.Task] (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.Dispatcher] FINISH getStorageDomainInfo error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,919-0700 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call StorageDomain.getInfo failed (error 379) in 0.00 seconds (__init__:312) 2021-01-16 19:53:58,956-0700 INFO (jsonrpc/1) [vdsm.api] START createStoragePool(poolType=None, spUUID='91fb3f22-5795-11eb-ad9f-00163e3f4683', poolName='Default', masterDom='54532dd4-3e5b-4885-b88e-599c81efb146', domList=['54532dd4-3e5b-4885-b88e-599c81efb146'], masterVersion=7, lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, ioOpTimeoutSec=10, leaseRetries=3, options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:48) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] updating pool 91fb3f22-5795-11eb-ad9f-00163e3f4683 backend from type NoneType instance 0x7f86e45f89d0 to type StoragePoolDiskBackend instance 0x7f8680738408 (sp:157) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] spUUID=91fb3f22-5795-11eb-ad9f-00163e3f4683 poolName=Default master_sd=54532dd4-3e5b-4885-b88e-599c81efb146 domList=['54532dd4-3e5b-4885-b88e-599c81efb146'] masterVersion=7 {'LEASERETRIES': 3, 'LEASETIMESEC': 60, 'LOCKRENEWALINTERVALSEC': 5, 'IOOPTIMEOUTSEC': 10} (sp:602) 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StorageDomain] sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.StoragePool] Unexpected error (sp:618) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
2021-01-16 19:53:58,963-0700 INFO (jsonrpc/1) [vdsm.api] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:52) 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') Unexpected error (task:880) Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in create domain.validate() File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py",
raise se.StorageDomainAccessError(self.sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
During handling of the above exception, another exception occurred:
Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line
887, in _run
return fn(*args, **kargs) File "<decorator-gen-31>", line 2, in createStoragePool File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line
50, in method
ret = func(*args, **kwargs) File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line
1027, in createStoragePool
leaseParams) File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line
619, in create
raise se.StorageDomainAccessError(sdUUID) vdsm.storage.exception.StorageDomainAccessError: Domain is either
permissions. dir existence] details] line 783, in main\n File \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", line 638, in post_create_check\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 3647, in add\n return self._internal_add(storage_domain, headers, query, wait)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232, in _internal_add\n return future.wait() if wait else future\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in wait\n return self._code(response)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in callback\n self._check_fault(respon options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48) options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:48) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) line 538, in validate partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',)
2021-01-16 19:53:58,964-0700 INFO (jsonrpc/1) [storage.TaskManager.Task] (Task='a946357f-537a-4c9c-9040-e70a98ee5643') aborting: Task is aborted: "value=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) 2021-01-16 19:53:58,964-0700 ERROR (jsonrpc/1) [storage.Dispatcher] FINISH createStoragePool error=Domain is either partially accessible or entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) 2021-01-16 19:53:58,965-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call StoragePool.create failed (error 379) in 0.01 seconds (__init__:312) 2021-01-16 19:54:01,923-0700 INFO (jsonrpc/5) [api.host] START getStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,964-0700 INFO (jsonrpc/7) [api.host] START getAllVmStats() from=::ffff:192.168.222.53,41192 (api:48) 2021-01-16 19:54:01,966-0700 INFO (jsonrpc/7) [api.host] FINISH getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) 2021-01-16 19:54:01,977-0700 INFO (jsonrpc/5) [vdsm.api] START repoStats(domains=()) from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:48) 2021-01-16 19:54:01,978-0700 INFO (jsonrpc/5) [vdsm.api] FINISH repoStats return={} from=::ffff:192.168.222.53,41192, task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:54) 2021-01-16 19:54:01,980-0700 INFO (jsonrpc/5) [vdsm.api] START multipath_health() from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:48) 2021-01-16 19:54:01,981-0700 INFO (jsonrpc/5) [vdsm.api] FINISH multipath_health return={} from=::ffff:192.168.222.53,41192, task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:54) 2021-01-16 19:54:01,996-0700 INFO (jsonrpc/5) [api.host] FINISH getStats return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) ***end snip*** _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/G2TQC6XTJTIBAI...
participants (5)
-
Joop
-
Matt Snow
-
Nir Soffer
-
Strahil Nikolov
-
Yedidyah Bar David