Reg: Hosted-engine "vm not running on this host"
by syedquadeer@ctel.in
Dear Team,
We have setup of 3 node cluster with 3.4 version of ovirt, after node server
restart, hosted engine showing as "vm not running on this host" on 2 nodes
and it is running on only one node. Please find screen shot below,
--== Host 1 status ==--
Status up-to-date : True
Hostname : node2.ctelhyd.in
Host ID : 1
Engine status : {"reason": "vm not running on this
host", "health": "bad", "vm": "down", "detail": "unknown"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : 04c18d76
Host timestamp : 231158
--== Host 2 status ==--
Status up-to-date : True
Hostname : node3.ctelhyd.in
Host ID : 2
Engine status : {"reason": "vm not running on this
host", "health": "bad", "vm": "down", "detail": "unknown"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : f0b6a1b2
Host timestamp : 231171
--== Host 3 status ==--
Status up-to-date : True
Hostname : node1.ctelhyd.in
Host ID : 3
Engine status : {"health": "good", "vm": "up",
"detail": "up"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : 7ef5dd17
Host timestamp : 231223
Thanks & Regards,
Syed Abdul Qadeer.
7660022818.
6 years, 5 months
HE + Gluster : Engine corrupted?
by Hanson Turner
Hi Benny,
Who should I be reaching out to for help with a gluster based hosted
engine corruption?
--== Host 1 status ==--
conf_on_shared_storage : True
Status up-to-date : True
Hostname : ovirtnode1.abcxyzdomains.net
Host ID : 1
Engine status : {"reason": "failed liveliness
check", "health": "bad", "vm": "up", "detail": "Up"}
Score : 3400
stopped : False
Local maintenance : False
crc32 : 92254a68
local_conf_timestamp : 115910
Host timestamp : 115910
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=115910 (Mon Jun 18 09:43:20 2018)
host-id=1
score=3400
vm_conf_refresh_time=115910 (Mon Jun 18 09:43:20 2018)
conf_on_shared_storage=True
maintenance=False
state=GlobalMaintenance
stopped=False
My when I VNC into my HE, All I get is:
Probing EDD (edd=off to disable)... ok
So, that's why it's failing the liveliness check... I cannot get the
screen on HE to change short of ctl-alt-del which will reboot the HE.
I do have backups for the HE that are/were run on a nightly basis.
If the cluster was left alone, the HE vm would bounce from machine to
machine trying to boot. This is why the cluster is in maintenance mode.
One of the nodes was down for a period of time and brought back,
sometime through the night, which is when the automated backup kicks,
the HE started bouncing around. Got nearly 1000 emails.
This seems to be the same error (but may not be the same cause) as
listed here:
https://bugzilla.redhat.com/show_bug.cgi?id=1569827
Thanks,
Hanson
6 years, 5 months
oVirt Engine 4.1 with oVirt Node 4.2
by stefanos@prismatelecomtesting.com
Hi,
I would like to know if oVirt 4.1 can manage 4.2 hosts just like RHVM does.
precisely, can oVirt 4.1 manage:
- an oVirt Node 4.2 with a cluster compatibility of 4.1?
- a CentOS host with 4.2 repo with a cluster compatibility of 4.1?
Thank you,
--
Stefano Stagnaro
Prisma Telecom Testing S.r.l.
Via Petrocchi, 4
20127 Milano – Italy
Tel. 02 26113507 int 339
e-mail: stefanos at prismatelecomtesting.com
skype: stefano.stagnaro
6 years, 5 months
engine
by du_hongyu@yeah.net
Hi
I export OVA from my ovirt , then want to import the OVA to anther ovirt, but this is failed
2018-06-25 13:52:03,836+08 INFO [org.ovirt.engine.core.bll.exportimport.ConvertVmCallback] (EE-ManagedThreadFactory-engineScheduled-Thread-94) [5c66b700] Conversion of VM from external environment failed: Job u'7ef7bff2-5766-4f38-ac28-cdd9da4c407c' process failed exit-code: 1
Regards
Hongyu Du
6 years, 5 months
unable to create iSCSI storage domain
by Bernhard Dick
Hi,
I've a problem creating an iSCSI storage domain. My hosts are running
the current ovirt 4.2 engine-ng version. I can detect and login to the
iSCSI targets, but I cannot see any LUNs (on the LUNs > Targets page).
That happens with our storage and with a linux based iSCSI target which
I created for testing purposes.
When I logon to the ovirt hosts I see that they are connected with the
target LUNs (dmesg is telling that there are iscsi devices being found
and they are getting assigned to devices in /dev/sdX ). Writing and
reading from the devices (also accros hosts) works. Do you have some
advice how to troubleshoot this?
Regards
Bernhard
6 years, 5 months
snapshot going to locked state and stays with it
by Hari Prasanth Loganathan
Hi Team,
I took a snapshot using oVirt and it stays in the LOCKED state for half an
hour.
DateJun 25, 2018, 1:06:58 PMStatus*LOCKED*MemoryfalseDescription
Immediate2018625-13655Defined Memory1024MBPhysical Memory
Guaranteed1024MBNumber
of CPU Cores1 (1:1:1)
1) What could be the reason for this LOCKED state?
2) How can I recover from it?
Thanks,
Hari
6 years, 5 months
OVirt SHE deployment failed - 4.2.2 / 4.2.3
by jeanbaptiste@nfrance.com
Hello List 😊
I’m facing an issue regarding Self Hosted Engine appliance deployment on OVirt 4.2( tested on 4.2.2 and 4.2.3).
During setup tool, I’m unable to mount a remote storage NFS / iSCSI.
When I try to mount an NFS share:
[ INFO ] TASK [Remove host-deploy configuration file]
[ INFO ] changed: [localhost]
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: nfs
Please specify the nfs version you would like to use (auto, v3, v4, v4_1)[auto]: xxxxxx:/YYYYYY
[ ERROR ] Invalid value
Please specify the nfs version you would like to use (auto, v3, v4, v4_1)[auto]:
Please specify the full shared storage connection path to use (example: host:/path): xxxxxx:/YYYYYY
If needed, specify additional mount options for the connection to the hosted-engine storagedomain []:
[ INFO ] Creating Storage Domain
[ INFO ] TASK [Gathering Facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Check local VM dir stat]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Enforce local VM dir existence]
[ INFO ] skipping: [localhost]
[ INFO ] TASK [include_tasks]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Obtain SSO token using username/password credentials]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch host facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch cluster ID]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch cluster facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch Datacenter facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch Datacenter ID]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch Datacenter name]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Add NFS storage domain]
[ ERROR ] Verify permission settings on the specified storage path.]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[Permission settings on the specified path do not allow access to the storage.\nVerify permission settings on the specified storage path.]\". HTTP response code is 400."}
When I try to mount an iSCSI LUN:
Please specify the storage you would like to use (glusterfs, iscsi, fc, nfs)[nfs]: iscsi
Please specify the iSCSI portal IP address: XXXXXX
Please specify the iSCSI portal port [3260]:
Please specify the iSCSI discover user:
Please specify the iSCSI discover password:
Please specify the iSCSI portal login user:
Please specify the iSCSI portal login password:
[ INFO ] Discovering iSCSI targets
[ INFO ] TASK [Gathering Facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [include_tasks]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Obtain SSO token using username/password credentials]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Prepare iSCSI parameters]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch host facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [iSCSI discover with REST API]
[ INFO ] ok: [localhost]
The following targets have been found:
[1] iqn.1992-04.com.emc:cx.aaaaaaaaaaaaaaaaa
TPGT: 7, portals:
XXXXXXX:3260
[2] iqn.1992-04.com.emc:cx. aaaaaaaaaaaaaaaab
TPGT: 6, portals:
XXXXXXX:3260
[3] iqn.1992-04.com.emc:cx. aaaaaaaaaaaaaaaac
TPGT: 8, portals:
XXXXXX:3260
[4] iqn.1992-04.com.emc:cx. aaaaaaaaaaaaaaaad
TPGT: 5, portals:
XXXXXXX:3260
Please select a target (1, 2, 3, 4) [1]: 4
[ INFO ] Getting iSCSI LUNs list
[ INFO ] TASK [Gathering Facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [include_tasks]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Obtain SSO token using username/password credentials]
[ INFO ] ok: [localhost]
[ INFO ] TASK [iSCSI login]
[ INFO ] TASK [Get iSCSI LUNs]
[ INFO ] ok: [localhost]
The following luns have been found on the requested target:
[1] 36006016045004300897d2b5bdfea8478 200GiB DGC VRAID
status: free, paths: 1 active
Please select the destination LUN (1) [1]: 1
[ INFO ] iSCSI discard after delete is disabled
[ INFO ] Creating Storage Domain
[ INFO ] TASK [Gathering Facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Check local VM dir stat]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Enforce local VM dir existence]
[ INFO ] skipping: [localhost]
[ INFO ] TASK [include_tasks]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Obtain SSO token using username/password credentials]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch host facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch cluster ID]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch cluster facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch Datacenter facts]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch Datacenter ID]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Fetch Datacenter name]
[ INFO ] ok: [localhost]
[ INFO ] TASK [Add NFS storage domain]
[ INFO ] skipping: [localhost]
[ INFO ] TASK [Add glusterfs storage domain]
[ INFO ] skipping: [localhost]
[ INFO ] TASK [Add iSCSI storage domain]
[ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is "[Network error during communication with the Host.]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[Network error during communication with the Host.]\". HTTP response code is 400."}
Some unnecessary informations like IP or IQN were hidden
I’m pretty disturbed since iSCSI issue seem resoled since 4.2.2 release: https://bugzilla.redhat.com/show_bug.cgi?id=1529226
Did you have same issue to deploy SHE since 4.2* ? ovirt-hosted-engine-setup seems not working has expected
Regards,
Jean-Baptiste,
6 years, 5 months
rhgs 3.3.1 rhev 4.2 vdsm gluster host deploy fails with no real indication of an error, looks like a network error
by Dan Lavu
Hello,
Seeing some odd behavior deploying a host to a gluster cluster on rhev 4.2
running rhgs 3.3.1. The host failed to be added to the cluster and there
are no real errors in the host deploy logs. Upon further investigation it
looks like its failing network host checks. The interfaces gets added to
the host when trying to configure host networks, but fails even though
everything seems routable.
These are the main errors I can see that might be causing an issue.
*Jun 24 20:27:12 deadpool.glstr.runlevelone.lan daemonAdapter[17160]:
libvirt: Network Driver error : Network not found: no network with matching
name 'vdsm-ovirtmgmt'*
*Jun 24 20:27:12 deadpool.glstr.runlevelone.lan daemonAdapter[17160]:
libvirt: Network Driver error : Network not found: no network with matching
name 'vdsm-infiniband'*
Fails connectivity check.
*Jun 24 20:27:21 deadpool.glstr.runlevelone.lan vdsm[17316]: vdsm vds ERROR
connectivity check failed*
* Traceback
(most recent call last):*
* File
"/usr/share/vdsm/API.py", line 1650, in _rollback*
* yield
rollbackCtx*
* File
"/usr/share/vdsm/API.py", line 1502, in setupNetworks*
*
supervdsm.getProxy().setupNetworks(networks, bondings, options)*
* File
"/usr/share/vdsm/supervdsm.py", line 50, in __call__*
* return
callMethod()*
* File
"/usr/share/vdsm/supervdsm.py", line 48, in <lambda>*
* **kwargs)*
* File
"<string>", line 2, in setupNetworks*
* File
"/usr/lib64/python2.7/multiprocessing/managers.py", line 773, in
_callmethod*
* raise
convert_to_error(kind, result)*
*
ConfigNetworkError: (10, 'connectivity check failed')*
Fulls logs can be found here,
https://paste.fedoraproject.org/paste/9v7DL2uCl2EcRnduliChsQ
Any help is appreciated, also let me know if I need to file a bug.
Thanks,
Dan
6 years, 5 months
private networking question
by Matthew Wimpelberg
Hi,
I'm currently running oVirt with a self hosted engine configuration on an Intel NUC. All of my VMs are on my home 192.168.1.0/24 network, but I want to set up a completely isolated network like VirtualBox has (https://blogs.oracle.com/scoter/virtualbox-host-only). The NUC that I have only has one NIC so I think I'll have to set up a dummy interface of some kind. I tried that with the below example. I couldn't find a good guide or doc on this. Can someone please assist?
[root@host ~]# cat /etc/sysconfig/network-scripts/ifcfg-dummy0
DEVICE=dummy0
BOOTPROTO=none
ONBOOT=yes
NM_CONTROLLED=no
PROMISC=yes
3: dummy0: <BROADCAST,NOARP,UP,LOWER_UP> mtu 1500 qdisc noqueue master vnet0 state UNKNOWN group default qlen 1000
link/ether 52:54:00:7e:27:af brd ff:ff:ff:ff:ff:ff
inet6 fe80::5054:ff:fe7e:27af/64 scope link
valid_lft forever preferred_lft forever
Sent with [ProtonMail](https://protonmail.com) Secure Email.
6 years, 5 months
vGPU setup guide
by Callum Smith
Dear All,
IS this the most current and useful example of implenting vGPUs in oVirt? I had understood that 4.2 had NVIDIA GRID support as a flagship feature, but this appears to be 4.1.4? It seems a very reasonable and decent guide, just don't want to go down this route if there's alternatives now available in 4.2.x.
https://mpolednik.github.io/2017/09/13/vgpu-in-ovirt/
Regards,
Callum
--
Callum Smith
Research Computing Core
Wellcome Trust Centre for Human Genetics
University of Oxford
e. callum(a)well.ox.ac.uk<mailto:callum@well.ox.ac.uk>
6 years, 5 months