oVirt HCI issue: GlusterD refusing to start due to quorum issues for volumes where it isn’t contributing any brick
by thomas@hoberg.net
Sorry if it's a duplicate: I go an error on post... And yes I posted it on the Gluster slack first, but I am using Gluster only because the marketing on oVirt HCI worked so well...
I got 3 recycled servers for an oVirt test environment first and set those up as 3-node HCI using defaults mostly, 2 replica + 1 arbiter, 'engine', 'vmstore' and 'data' volumes with single bricks for each node. I call these group A.
Then I got another set of five machines, let's call them group B, with somewhat different hardware characteristics than group A, but nicely similar between themselves. I wanted to add these to the farm as compute nodes but also use their storage as general GlusterFS storage for a wider use.
Group B machines were added as hosts and set up to run hosted-engine, but they do not contribute bricks to the normal oVirt volumes 'engine', 'vmstore' or 'data'. With some Ansible trickery I managed to set up two dispersed volumes (4 data: 1 redundancy) on group B 'scratch' and 'tape', mostly for external GlusterFS use. oVirt picked them up automagically, so I guess they could also be used with VMs.
I expect to get more machines and adding them one-by-one to dispersed volumes with a fine balance between capacity and redundancy made me so enthusiastic about oVirt HCI in the first place...
After some weeks of fine operation I had to restart a machine from group B for maintenance. When it came back up, GlusterD refuses to come online, because it doesn't have "quorum for volumes 'engine', 'vmstore' and 'data'"....
It's a small surprise it doesn't *have* quorum, what's a bigger surprise is that it *asks* for quorum in a volume where it's not contributing any bricks. What's worse is that it then refuses to start serving its bricks for 'scratch' and 'tape', which are now growing apart without any chance of healing.
How do I fix this?
Is this a bug (my interpretation) or do I fundamentlly misunderstand how Gluster as a hyper scale out file system is supposed to work with potentially thousands of hosts contributing each dozens of bricks to each of hundreds of volumes in a single name space?
4 years, 2 months
Probably dns problem: Internal JSON-RPC error
by jan.kleefeld@jan-home.de
I have a clean installed CentOS 8.2 2004 on my server. The self hosted engine deploy (ovirt-4.4) aborts with the following message:
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "The host has been set in non_operational status, deployment errors: code 505: Host red.colors.ovirt.local installation failed. Failed to configure management network on the host., code 1120: Failed to configure management network on host red.colors.ovirt.local due to setup networks failure., code 9000: Failed to verify Power Management configuration for Host red.colors.ovirt.local., code 10802: VDSM red.colors.ovirt.local command HostSetupNetworksVDS failed: Internal JSON-RPC error:
{'reason': '
desired
=======
---
dns-resolver:
search: []
server:
- 192.168.2.150
- fe80::1%eno1
current
=======
---
dns-resolver:
search: []
server:
- 192.168.2.150
difference
==========
--- desired
+++ current
@@ -3,4 +3,3 @@
search: []
server:
- 192.168.2.150
- - fe80::1%eno1
'}, fix accordingly and re-deploy."}
~# cat /etc/resolv.conf
# Generated by NetworkManager
search colors.ovirt.local
nameserver 192.168.2.150
nameserver fe80::1%eno1
I am confused, because the probalby missing line is present. Is there maybe another config file, where this last line could be missing?
Maybe I can force somehow the installer, to reload the resolv conf, so it can fetch ne new line?
I appreciate your help.
Jan
4 years, 2 months
OVN Geneve tunnels not been established
by Konstantinos B
Hi all
We have a small installation based on OVIRT 4.3.
1 Cluster is based on Centos 7 and the other on OVIRT NG Node image.
The environment was stable till an upgrade took place a couple of months ago.
As such we had to re-install one of the Centos 7 node and start from scratch.
Even though the installation completed successfully and VMs are created, the following are not working as expected:
1. ovn geneve tunnels are not established with the other Centos 7 node in the cluster.
2. Centos 7 node is configured by ovirt engine however no geneve tunnel is established when "ovn-sbctl show" is issued on the engine.
3. no flows are shown on the engine on port 6642 for the ovs db.
Does anyone have any experience on how to troubleshoot OVN on ovirt?
Thank you
4 years, 2 months
"Add NFS storage domain", "status": "FAILED",
by info@worldhostess.com
I am new to oVirt and this is my first installation
https://ovirt.org/documentation/installing_ovirt_as_a_self-hosted_engine_us
ing_the_cockpit_web_interface/
Everything seems fine, but then when I tried step 5, "Hosted Engine
Deployment" it keep giving me this error. This information might be too
much, but I am not familiar to know what is important.
2020-09-08 06:53:58,991+0200 DEBUG var changed: host "localhost" var
"datacenter_name" type "<class
'ansible.utils.unsafe_proxy.AnsibleUnsafeText'>" value: ""Default""
2020-09-08 06:53:58,991+0200 INFO ansible ok {'status': 'OK',
'ansible_type': 'task', 'ansible_playbook':
'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml',
'ansible_host': 'localhost', 'ansible_task': 'Fetch Datacenter name',
'task_duration': 0}
2020-09-08 06:53:58,991+0200 DEBUG ansible on_any args
<ansible.executor.task_result.TaskResult object at 0x7f825fb425c0> kwargs
2020-09-08 06:53:59,242+0200 INFO ansible task start {'status': 'OK',
'ansible_type': 'task', 'ansible_playbook':
'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml',
'ansible_task': 'ovirt.hosted_engine_setup : Add NFS storage domain'}
2020-09-08 06:53:59,242+0200 DEBUG ansible on_any args TASK:
ovirt.hosted_engine_setup : Add NFS storage domain kwargs
is_conditional:False
2020-09-08 06:53:59,242+0200 DEBUG ansible on_any args localhostTASK:
ovirt.hosted_engine_setup : Add NFS storage domain kwargs
2020-09-08 06:56:04,962+0200 DEBUG var changed: host "localhost" var
"ansible_play_hosts" type "<class 'list'>" value: "[]"
2020-09-08 06:56:04,963+0200 DEBUG var changed: host "localhost" var
"ansible_play_batch" type "<class 'list'>" value: "[]"
2020-09-08 06:56:04,963+0200 DEBUG var changed: host "localhost" var
"play_hosts" type "<class 'list'>" value: "[]"
2020-09-08 06:56:04,963+0200 ERROR ansible failed {
"ansible_host": "localhost",
"ansible_playbook":
"/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml",
"ansible_result": {
"_ansible_no_log": false,
"changed": false,
"exception": "Traceback (most recent call last):\n File
\"/tmp/ansible_ovirt_storage_domain_payload_ir_v98cz/ansible_ovirt_storage_d
omain_payload.zip/ansible/modules/cloud/ovirt/ovirt_storage_domain.py\",
line 792, in main\n File
\"/tmp/ansible_ovirt_storage_domain_payload_ir_v98cz/ansible_ovirt_storage_d
omain_payload.zip/ansible/module_utils/ovirt.py\", line 623, in create\n
**kwargs\n File
\"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 26097, in
add\n return self._internal_add(storage_domain, headers, query, wait)\n
File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 232,
in _internal_add\n return future.wait() if wait else future\n File
\"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 55, in
wait\n return self._code(response)\n File
\"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in
callback\n self._check_fault(response)\n File
\"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 132, in
_check_fault\n self._raise_error(response, body)\n File
\"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in
_raise_error\n raise error\novirtsdk4.Error: Fault reason is \"Operation
Failed\". Fault detail is \"[Problem while trying to mount target]\". HTTP
response code is 400.\n",
"invocation": {
"module_args": {
"backup": null,
"comment": null,
"critical_space_action_blocker": null,
"data_center": "Default",
"description": null,
"destroy": null,
"discard_after_delete": null,
"domain_function": "data",
"fcp": null,
"fetch_nested": false,
"format": null,
"glusterfs": null,
"host": "host.worldhostess.co.za",
"id": null,
"iscsi": null,
"localfs": null,
"managed_block_storage": null,
"name": "hosted_storage",
"nested_attributes": [],
"nfs": {
"address": "storage01.worldhostess.co.za",
"mount_options": null,
"path": "/data/nfs",
"version": "auto"
},
"poll_interval": 3,
"posixfs": null,
"state": "unattached",
"timeout": 180,
"wait": true,
"warning_low_space": null,
"wipe_after_delete": null
}
},
"msg": "Fault reason is \"Operation Failed\". Fault detail is
\"[Problem while trying to mount target]\". HTTP response code is 400."
},
"ansible_task": "Add NFS storage domain",
"ansible_type": "task",
"status": "FAILED",
"task_duration": 125
}
2020-09-08 06:56:04,963+0200 DEBUG ansible on_any args
<ansible.executor.task_result.TaskResult object at 0x7f825f9c3a58> kwargs
ignore_errors:None
2020-09-08 06:56:04,964+0200 INFO ansible stats {
"ansible_playbook":
"/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml",
"ansible_playbook_duration": "02:16 Minutes",
"ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 14,
'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 2, 'rescued': 0,
'ignored': 0}}",
"ansible_type": "finish",
"status": "FAILED"
}
2020-09-08 06:56:04,964+0200 INFO SUMMARY:
Duration Task Name
-------- --------
[ < 1 sec ] Execute just a specific set of steps
[ 00:01 ] Force facts gathering
[ < 1 sec ] Check local VM dir stat
[ < 1 sec ] Obtain SSO token using username/password credentials
[ < 1 sec ] Fetch host facts
[ < 1 sec ] Fetch cluster ID
[ < 1 sec ] Fetch cluster facts
[ < 1 sec ] Fetch Datacenter facts
[ < 1 sec ] Fetch Datacenter ID
[ < 1 sec ] Fetch Datacenter name
[ FAILED ] Add NFS storage domain
2020-09-08 06:56:04,965+0200 DEBUG ansible on_any args
<ansible.executor.stats.AggregateStats object at 0x7f826252ef60> kwargs
Yours Sincerely,
Dr Henni Appelgryn PhD DBA MBA AEP MDP
WorldHostess
Skype: genie_za
WeChat: HenniAppelgryn
LinkedIn: https://www.linkedin.com/in/henni-appelgryn/
Hosting with a difference
4 years, 2 months
Failed to create a new VM in ovirt due to "The host did not satisfy internal filter Memory" error
by KISHOR K
Hi All,
I'm new to Ovirt and not having a perfect experience with Ovirt yet.
I ran into a strange issue today when I tried to create a new VM with 32GB in ovirt. VM creation basically due to an error, pasted below.
Cannot run VM. There is no host that satisfies current scheduling constraints. See below for details:, The host host-01 did not satisfy internal filter Memory., The host host-01 did not satisfy internal filter Memory.
After some troubleshooting, I found that there was enough available memory (around 120 GB free ) in the host to host new VM and there were around 10 VMs already running on this host.
But, I noticed that ovirt is actually scheduling and creating VM based on "Max free Memory for scheduling new VMs" value, which seems to be set/updated based on total allocated memory for all VMs running on the host and it's not really the then consumed memory by running VMs.
Can anyone help to explain if it is some kind of bug in ovirt or it expected behavior?
If it is expected behavior, is there any possibility to change it to make sure VM is created based on actual free memory?
Thanks a lot in advance for your support.
/Kishore
/Kishore
4 years, 2 months
Diskless oVirt Nodes
by harryo.dk@gmail.com
Hi, why does my ovirt hypervisor node crash when i pull the hypervisor boot drive?
The following thread claims that the oVirt node hypervisor should lodad into memory, so boot disk failure shouldn't have any impact on a running node.
https://lists.ovirt.org/pipermail/users/2014-December/029771.html
Only memory failure should have a impact, and if RAM fails there is a possibility that the hypervisor OS would have crashed regardless.
If you load the hypervisor OS into memory there will be moved/combined a point of failure (the boot disk) with another(the RAM).
RAM=memory
4 years, 2 months
Hyperconverged set up with Gluster and 6 Nodes
by Holger Petrick
Dear All,
in all documentation to set up oVirt in a Hyperconverged system with Gluster says that it must have 3 Enterprise nodes.
In our test environment we set up successfully with oVirt wizard an oVirt cluster with gluster + Arbiter and hosted engine.
Is it also possible to build a cluster with 6 nodes? The wizard gives only the option to add 3 nodes. Is this kind of design supported?
My idea is, first to set up the Gluster, then install the oVirt nodes and deploy the engine.
Would this work? Any documentation avaiable?
Thanks
Holger
4 years, 2 months
Hosted engine install failure: ipv6.gateway: gateway cannot be set if there are no addresses configured
by Sverker Abrahamsson
Hi,
I'm attempting to install hosted engine but getting this failure:
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "The
host has been set in non_operational status, deployment errors: code
505: Host h1-mgmt.limetransit.com installation failed. Failed to
configure management network on the host., code 1120: Failed to
configure management network on host h1-mgmt.limetransit.com due to
setup networks failure., code 9000: Failed to verify Power Management
configuration for Host h1-mgmt.limetransit.com., code 10802: VDSM
h1-mgmt.limetransit.com command HostSetupNetworksVDS failed: Internal
JSON-RPC error: {'reason': 'Unexpected failure of libnm when running the
mainloop: run execution'}, fix accordingly and re-deploy."}
Looking in vdsm.log I find this which I believe is the root cause:
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:25,897::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper)
call setupNetworks with ({'ovirtmgmt': {'vlan': '4000', 'netmask':
'255.255.255.0', 'ipv6autoconf': False,
'nic': 'enp4s0', 'bridged': 'true', 'ipaddr': '172.27.1.1',
'defaultRoute': True, 'dhcpv6': False, 'STP': 'no', 'mtu': 1500,
'switch': 'legacy'}}, {}, {'connectivityTimeout': 120,
'commitOnSuccess': True, 'connectivityCh
eck': 'true'}) {}
MainProcess|jsonrpc/0::INFO::2020-09-02
16:38:25,897::api::220::root::(setupNetworks) Setting up network
according to configuration: networks:{'ovirtmgmt': {'vlan': '4000',
'netmask': '255.255.255.0', 'ipv6autoconf': Fal
se, 'nic': 'enp4s0', 'bridged': 'true', 'ipaddr': '172.27.1.1',
'defaultRoute': True, 'dhcpv6': False, 'STP': 'no', 'mtu': 1500,
'switch': 'legacy'}}, bondings:{}, options:{'connectivityTimeout': 120,
'commitOnSuccess':
True, 'connectivityCheck': 'true'}
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:25,902::cmdutils::130::root::(exec_cmd) /sbin/tc qdisc show (cwd None)
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:25,906::cmdutils::138::root::(exec_cmd) SUCCESS: <err> = b''; <rc> = 0
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:25,945::vsctl::74::root::(commit) Executing commands:
/usr/bin/ovs-vsctl --timeout=5 --oneline --format=json -- list Bridge --
list Port -- list Interface
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:25,945::cmdutils::130::root::(exec_cmd) /usr/bin/ovs-vsctl
--timeout=5 --oneline --format=json -- list Bridge -- list Port -- list
Interface (cwd None)
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:25,952::cmdutils::138::root::(exec_cmd) SUCCESS: <err> = b''; <rc> = 0
MainProcess|jsonrpc/0::INFO::2020-09-02
16:38:25,957::netconfpersistence::58::root::(setNetwork) Adding network
ovirtmgmt({'vlan': 4000, 'netmask': '255.255.255.0', 'ipv6autoconf':
False, 'nic': 'enp4s0', 'bridged': True
, 'ipaddr': '172.27.1.1', 'defaultRoute': True, 'dhcpv6': False, 'mtu':
1500, 'switch': 'legacy', 'stp': False, 'bootproto': 'none',
'nameservers': ['213.133.98.98']})
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:25,958::commands::153::common.commands::(start) /usr/bin/taskset
--cpu-list 0-7 /usr/libexec/vdsm/hooks/before_network_setup/50_fcoe (cwd
None)
MainProcess|jsonrpc/0::INFO::2020-09-02
16:38:26,154::hooks::122::root::(_runHooksDir)
/usr/libexec/vdsm/hooks/before_network_setup/50_fcoe: rc=0 err=b''
MainProcess|jsonrpc/0::INFO::2020-09-02
16:38:26,155::configurator::195::root::(_setup_nmstate) Processing setup
through nmstate
MainProcess|jsonrpc/0::INFO::2020-09-02
16:38:26,175::configurator::197::root::(_setup_nmstate) Desired state:
{'interfaces': [{'name': 'enp4s0', 'state': 'up', 'mtu': 1500}, {'vlan':
{'id': 4000, 'base-iface': 'enp4s0'}
, 'name': 'enp4s0.4000', 'type': 'vlan', 'state': 'up', 'mtu': 1500,
'ipv4': {'enabled': False}, 'ipv6': {'enabled': False}}, {'name':
'ovirtmgmt', 'type': 'linux-bridge', 'state': 'up', 'mtu': 1500,
'bridge': {'port': [
{'name': 'enp4s0.4000'}], 'options': {'stp': {'enabled': False}}},
'ipv4': {'enabled': True, 'address': [{'ip': '172.27.1.1',
'prefix-length': 24}], 'dhcp': False}, 'ipv6': {'enabled': False}}],
'dns-resolver': {'config'
: {'server': ['213.133.98.98']}}}
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:26,217::checkpoint::121::root::(create) Checkpoint
/org/freedesktop/NetworkManager/Checkpoint/1 created for all devices: 60
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:26,218::netapplier::239::root::(_add_interfaces) Adding new
interfaces: ['ovirtmgmt']
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:26,221::netapplier::251::root::(_edit_interfaces) Editing
interfaces: ['enp4s0.4000', 'enp4s0']
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:26,224::nmclient::136::root::(execute_next_action) Executing NM
action: func=add_connection_async
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:26,231::connection::329::root::(_add_connection_callback)
Connection adding succeeded: dev=ovirtmgmt
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:26,232::nmclient::136::root::(execute_next_action) Executing NM
action: func=commit_changes_async
MainProcess|jsonrpc/0::ERROR::2020-09-02
16:38:26,233::nmclient::200::root::(quit) NM main-loop aborted:
Connection update failed: error=nm-connection-error-quark: ipv6.gateway:
gateway cannot be set if there are no addresses configured (7),
dev=enp4s0/<enum NM_DEVICE_STATE_ACTIVATED of type NM.DeviceState>
MainProcess|jsonrpc/0::DEBUG::2020-09-02
16:38:26,246::checkpoint::164::root::(rollback) Checkpoint
/org/freedesktop/NetworkManager/Checkpoint/1 rollback executed:
dbus.Dictionary({dbus.String('/org/freedesktop/NetworkManager/Devices/6'):
dbus.UInt32(0),
dbus.String('/org/freedesktop/NetworkManager/Devices/7'):
dbus.UInt32(0),
dbus.String('/org/freedesktop/NetworkManager/Devices/8'):
dbus.UInt32(0),
dbus.String('/org/freedesktop/NetworkManager/Devices/9'):
dbus.UInt32(0),
dbus.String('/org/freedesktop/NetworkManager/Devices/3'):
dbus.UInt32(0),
dbus.String('/org/freedesktop/NetworkManager/Devices/2'):
dbus.UInt32(0),
dbus.String('/org/freedesktop/NetworkManager/Devices/4'):
dbus.UInt32(0),
dbus.String('/org/freedesktop/NetworkManager/Devices/1'):
dbus.UInt32(0),
dbus.String('/org/freedesktop/NetworkManager/Devices/5'):
dbus.UInt32(0)}, signature=dbus.Signature('su'))
MainProcess|jsonrpc/0::ERROR::2020-09-02
16:38:31,251::supervdsm_server::97::SuperVdsm.ServerCallback::(wrapper)
Error in setupNetworks
Traceback (most recent call last):
File "/usr/lib/python3.6/site-packages/vdsm/supervdsm_server.py",
line 95, in wrapper
res = func(*args, **kwargs)
File "/usr/lib/python3.6/site-packages/vdsm/network/api.py", line
241, in setupNetworks
_setup_networks(networks, bondings, options, net_info)
File "/usr/lib/python3.6/site-packages/vdsm/network/api.py", line
266, in _setup_networks
networks, bondings, options, net_info, in_rollback
File
"/usr/lib/python3.6/site-packages/vdsm/network/netswitch/configurator.py",
line 154, in setup
_setup_nmstate(networks, bondings, options, in_rollback)
File
"/usr/lib/python3.6/site-packages/vdsm/network/netswitch/configurator.py",
line 199, in _setup_nmstate
nmstate.setup(desired_state, verify_change=not in_rollback)
File "/usr/lib/python3.6/site-packages/vdsm/network/nmstate.py", line
63, in setup
state_apply(desired_state, verify_change=verify_change)
File "/usr/lib/python3.6/site-packages/libnmstate/deprecation.py",
line 40, in wrapper
return func(*args, **kwargs)
File "/usr/lib/python3.6/site-packages/libnmstate/nm/nmclient.py",
line 96, in wrapped
ret = func(*args, **kwargs)
File "/usr/lib/python3.6/site-packages/libnmstate/netapplier.py",
line 73, in apply
state.State(desired_state), verify_change, commit, rollback_timeout
File "/usr/lib/python3.6/site-packages/libnmstate/netapplier.py",
line 163, in _apply_ifaces_state
con_profiles=ifaces_add_configs + ifaces_edit_configs,
File "/usr/lib64/python3.6/contextlib.py", line 88, in __exit__
next(self.gen)
File "/usr/lib/python3.6/site-packages/libnmstate/netapplier.py",
line 232, in _setup_providers
mainloop.run(timeout=MAINLOOP_TIMEOUT)
File "/usr/lib/python3.6/site-packages/libnmstate/nm/nmclient.py",
line 177, in run
f"Unexpected failure of libnm when running the mainloop: {err}"
libnmstate.error.NmstateLibnmError: Unexpected failure of libnm when
running the mainloop: run execution
My host have a main network interface and a vlan interface on the same
physical interface which is an internal management network which should
be used as management network. The vlan interface doesn't have any ipv6
interface, but nmclient fails on setting ipv6 gateway. This seems like a
bug?
/Sverker
_______________________________________________
discuss mailing list
discuss(a)openvswitch.org
https://mail.openvswitch.org/mailman/listinfo/ovs-discuss
4 years, 2 months
iSCSI multipath issue
by Nick Kas
Hello evryone,
setup ovirt 4.4.1 on CentOS 8.2 as an experiment, and I am trying to get an iSCSI domain working but have issues. The little experimental cluster has 3 hosts. There is an ovirtmgmt network on the default vlan, and two iSCSI network (172.27.0/1.X) with vlans 20/21. ovirtmgmt has all the functions (Data, display, migration etc), and the iSCSI networks nothing yet, and they are not set as required.
The SAN device is already serving a few iSCSI volumes to a vmware cluster, so I know things are fine on this end. It has two controllers, and four NICs per controller so a total of 8 NICs, half of the NICS per controller on 172.27.0.X and half on 172.27.1.X.
When I create the iSCSI domain, I login to only one of the targets, and add the Volume, all is good and I can use the disc fine.
However when I login to more than one of the targets, then I start having issues with the Volume. Even when I enabled multipath in the cluster, and I created a single multipath by selecting both of the 172.27.0/1.X networks, and all the targets, the end result was the same. The hosts have difficulty accessing the volume, they may even swing between 'non-operational' and 'up' if I transfer data to the volume. When I ssh to the hosts and i check things in the command line I also get inconsistent results between hosts, and blocks that appear with lsblk when I first setup iSCSI have dissapeared after I try to actively use the volume.
I am new to iSCSI so I am not sure how to debug this. I am not sure if my multipath configuration is correct or not. The documentation on this part was not very detailed. I also tried to remove the domain, and try to experiment with mounting the iSCSI volume from the command line, but I cannot even discover the target from the command line, which is very bizarre. The command
iscsiadm --mode discovery --target sendtargets --portal 172.27.0.55 --discover
returns the message 'iscsiadm: cannot make connection to 172.27.0.55: No route to host'. Yet through ovirt, and if I select only one target, everything work fine!
Any suggestions on how to start debugging this would really be appreciated.
4 years, 2 months
New Installation OVIRT NODE 4.4.1
by henni@africadragon.com
I am on step 5, but I am not sure what I am doing wrong. I get the following
error. Any help will be appreciated.
I have a subdomain, but I am not sure if it is needed.
I am not sure if I have to add things in /etc/hosts
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Obtain SSO token using
username/password credentials]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Fetch host facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Fetch cluster ID]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Fetch cluster facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Fetch Datacenter facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Fetch Datacenter ID]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Fetch Datacenter name]
[ INFO ] ok: [localhost]
[ INFO ] TASK [ovirt.hosted_engine_setup : Add NFS storage domain]
[ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail
is "[Problem while trying to mount target]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault
reason is \"Operation Failed\". Fault detail is \"[Problem while trying to
mount target]\". HTTP response code is 400."}
4 years, 2 months