Problem with Ovirt
by Keith Forman
HiNeed help with setting up Ovirt.engine-setup run successfully on CentOS 7, with PostgreSQL 12.Now when running systemctl start ovirt-engine , it starts successfully, but in the frontend, it says 404 Not Found. In engine.log, I see the following 2 errors:Error initializing: Unable to determine the correct call signature - no procedure/function/signature for 'getallfrommacpools'andError in getting DB connection, database is inaccessible: Unable to determine the correct call signature - no procedure/function/signature for 'checkdbconnection'Any tips as to the way forward would be much appreciated
4 years, 3 months
Hyperconverged Ceph + Managed Block Storage
by Shantur Rathore
Hi all,
I am planning my new oVirt cluster on Apple hosts. These hosts can only
have one disk which I plan to partition and use for hyper converged setup.
As this is my first oVirt cluster I need help in understanding a few bits.
1. Is Hyper converged setup possible with Ceph using cinderlib?
2. Can this hyper converged setup be on oVirt Node Next hosts or only
Centos?
3. Can I install cinderlib on oVirt Node Next hosts?
4. Are there any pitfalls in such a setup?
Thanks for your help
Regards,
Shantur
4 years, 3 months
Create new user, but why cannot login ?
by tommy
I just create a new user:
[root@oeng ~]# ovirt-aaa-jdbc-tool user add cuitao
adding user cuitao...
user added successfully
[root@oeng ~]# ovirt-aaa-jdbc-tool user password-reset cuitao
Password:
Reenter password:
updating user cuitao...
user updated successfully
[root@oeng ~]# ovirt-aaa-jdbc-tool user edit cuitao
--password-valid-to="2221-01-15 05:23:41Z"
updating user cuitao...
user updated successfully
[root@oeng ~]# ovirt-aaa-jdbc-tool user show cuitao
-- User cuitao(300163db-8352-4fbd-86ac-d25014364f08) --
Namespace: *
Name: cuitao
ID: 300163db-8352-4fbd-86ac-d25014364f08
Display Name:
Email: sz_cuitao(a)163.com <mailto:sz_cuitao@163.com>
First Name: tommy
Last Name: cui
Department:
Title:
Description:
Account Disabled: false
Account Locked: false
Account Unlocked At: 1970-01-01 00:00:00Z
Account Valid From: 2021-01-15 05:23:41Z
Account Valid To: 2221-01-15 05:23:41Z
Account Without Password: false
Last successful Login At: 2021-01-15 05:54:49Z
Last unsuccessful Login At: 2021-01-15 05:32:12Z
Password Valid To: 2221-01-15 05:23:41Z
And I give VmCreator Role to the new account.
But why cannot login ?
4 years, 3 months
New oVirt Cluster and Managed block storage
by Shantur Rathore
Hi all,
I am planning my new oVirt cluster on Apple hosts. These hosts can only
have one disk which I plan to partition and use for hyper converged setup.
As this is my first oVirt cluster I need help in understanding a few bits.
1. Is Hyper converged setup possible with Ceph using cinderlib?
2. Can this hyper converged setup be on oVirt Node Next hosts or only
Centos?
3. Can I install cinderlib on oVirt Node Next hosts?
4. Are there any pitfalls in such a setup?
Thanks for your help
Regards,
Shantur
4 years, 3 months
potential split-brain after upgrading Gluster version and rebooting one of three storage nodes
by user-5138@yandex.com
Hello everyone,
I'm running an oVirt cluster (4.3.10.4-1.el7) on a bunch of physical nodes with Centos 7.9.2009 and the Hosted Engine is running as a virtual machine on one of these nodes. As for the storage, I'm running GlusterFS 6.7 on three separate physical storage nodes (also Centos 7). Gluster itself has three different volumes of the type "Replicate" or "Distributed-Replicate".
I recently updated both the system packages and the GlusterFS version to 6.10 on the first storage node (storage1) and now I'm seeing a potential split-brain situation for one of the three volumes when running "gluster volume heal info":
Brick storage1:/data/glusterfs/nvme/brick1/brick
Status: Connected
Number of entries: 0
Brick storage2:/data/glusterfs/nvme/brick1/brick
/c32d664d-69ba-4c3f-8ea1-240133963815/dom_md/ids
/
/.shard/.remove_me
Status: Connected
Number of entries: 3
Brick storage3:/data/glusterfs/nvme/brick1/brick
/c32d664d-69ba-4c3f-8ea1-240133963815/dom_md/ids
/
/.shard/.remove_me
Status: Connected
Number of entries: 3
I checked the hashes and the "dom_md/ids" file has a different md5 on every node. Using the heal on the volume command doesn't do anything and the entries remain. The heal info for the other two volumes shows no entries.
The affected gluster volume (type: replicate) is mounted as a Storage Domain using the path "storage1:/nvme" inside of oVirt and is used to store the root partitions of all virtual machines, which were running at the time of the upgrade and reboot of storage1. The volume has three bricks, with one brick being stored on each storage node. For the upgrade process I followed the steps shown at https://docs.gluster.org/en/latest/Upgrade-Guide/generic-upgrade-procedure/. I stopped and killed all gluster related services, upgraded both system and gluster packages and rebooted storage1.
Is this a split brain situation and how can I solve this? I would be very grateful for any help.
Please let me know if you require any additional information.
Best regards
4 years, 3 months
cockpit-ovirt-dashboard
by Gary Pedretty
So with the latest updates to Ovirt and CentOS 8 Stream it seems that you either can not install cockpit-ovirt-dashboard, or if you do, it downgrades cockpit-bridge and cockpit-system. You then end up not showing up to date on these hosts on Open Virtualization Manager and can only get back to a fully updated datacenter by doing a yum update --allowerasing which then removes cockpit-ovirt-dashboard.
My main concern is now that the command line hosted-engine commands have been removed, if the hosted engine is not running for some reason you have no visibility into what is going on since the regular cockpit host interface will not show the virtualization features. You can't put a host in maintenance or start the engine manually.
example....
[root@ravn-kvm-1 admin]# yum install cockpit-ovirt-dashboard
//////snip///////
Installed products updated.
Downgraded:
cockpit-bridge-217-1.el8.x86_64 cockpit-system-217-1.el8.noarch
Installed:
cockpit-dashboard-217-1.el8.noarch cockpit-ovirt-dashboard-0.14.17-1.el8.noarch ovirt-host-4.4.1-4.el8.x86_64 ovirt-hosted-engine-setup-2.4.9-1.el8.noarch
Complete!
[root@ravn-kvm-1 admin]# yum update
Last metadata expiration check: 3:08:22 ago on Tue 12 Jan 2021 01:00:15 PM AKST.
Error:
Problem 1: package ovirt-host-4.4.1-4.el8.x86_64 requires cockpit-dashboard, but none of the providers can be installed
- package cockpit-bridge-234-1.el8.x86_64 conflicts with cockpit-dashboard < 233 provided by cockpit-dashboard-217-1.el8.noarch
- cannot install the best update candidate for package ovirt-host-4.4.1-4.el8.x86_64
- cannot install the best update candidate for package cockpit-bridge-217-1.el8.x86_64
Problem 2: problem with installed package ovirt-host-4.4.1-4.el8.x86_64
- package ovirt-host-4.4.1-4.el8.x86_64 requires cockpit-dashboard, but none of the providers can be installed
- package cockpit-system-234-1.el8.noarch obsoletes cockpit-dashboard provided by cockpit-dashboard-217-1.el8.noarch
- cannot install the best update candidate for package cockpit-dashboard-217-1.el8.noarch
Problem 3: package ovirt-hosted-engine-setup-2.4.9-1.el8.noarch requires ovirt-host >= 4.4.0, but none of the providers can be installed
- package ovirt-host-4.4.1-4.el8.x86_64 requires cockpit-dashboard, but none of the providers can be installed
- package ovirt-host-4.4.1-1.el8.x86_64 requires cockpit-dashboard, but none of the providers can be installed
- package ovirt-host-4.4.1-2.el8.x86_64 requires cockpit-dashboard, but none of the providers can be installed
- package ovirt-host-4.4.1-3.el8.x86_64 requires cockpit-dashboard, but none of the providers can be installed
- package cockpit-system-234-1.el8.noarch obsoletes cockpit-dashboard provided by cockpit-dashboard-217-1.el8.noarch
- cannot install the best update candidate for package ovirt-hosted-engine-setup-2.4.9-1.el8.noarch
- cannot install the best update candidate for package cockpit-system-217-1.el8.noarch
(try to add '--allowerasing' to command line to replace conflicting packages or '--skip-broken' to skip uninstallable packages or '--nobest' to use not only best candidate packages)
_______________________________
Gary Pedretty
IT Manager
Ravn Alaska
Office: 907-266-8451
Mobile: 907-388-2247
Email: gary.pedretty(a)ravnalaska.com
"We call Alaska......Home!"
Ravn Alaska
CONFIDENTIALITY NOTICE:
The information in this email may be confidential and/or privileged. This email is intended to be reviewed by only the individual or organization named above. If you are not the intended recipient or an authorized representative of the intended recipient, you are hereby notified that any review, dissemination, forwarding or copying of the email and its attachments, if any, or the information contained herein is prohibited. If you have received this email in error, please immediately notify the sender by return email and delete this email from your system. Thank you.
4 years, 3 months
how does host's status changes ? Especially Unassigned or NonOperational or NonResponsible
by lifuqiong@sunyainfo.com
Hi all,
I'm confusing in Host's status in Ovirt Eninge.
When will the host's status become Unassigned or NonOperational or NonResponsible? and If the host's status change to these statuses, What will ovirt response to it?
After reading Ovirt Engine's source code, I find Only HostMonitoring.java and AutoRecoveryManager.java will change the status of Host.
for example, If the host's status is changed to NonOperational, The AutoRecoveryManager will traverse the NonOperational hosts and calling ActivateVdsCommand.java, which will only set the Host's status to Unassigned ? But I don't know What's the next step?
So where I can find the article or manual or other helpful information about this question?
Thank you .
Your sincerely
Mark
4 years, 3 months
Re: Constantly XFS in memory corruption inside VMs
by Strahil Nikolov
Damn...
You are using EFI boot. Does this happen only to EFI machines ?
Did you notice if only EL 8 is affected ?
Best Regards,
Strahil Nikolov
В неделя, 29 ноември 2020 г., 19:36:09 Гринуич+2, Vinícius Ferrão <ferrao(a)versatushpc.com.br> написа:
Yes!
I have a live VM right now that will de dead on a reboot:
[root@kontainerscomk ~]# cat /etc/*release
NAME="Red Hat Enterprise Linux"
VERSION="8.3 (Ootpa)"
ID="rhel"
ID_LIKE="fedora"
VERSION_ID="8.3"
PLATFORM_ID="platform:el8"
PRETTY_NAME="Red Hat Enterprise Linux 8.3 (Ootpa)"
ANSI_COLOR="0;31"
CPE_NAME="cpe:/o:redhat:enterprise_linux:8.3:GA"
HOME_URL="https://www.redhat.com/"
BUG_REPORT_URL="https://bugzilla.redhat.com/"
REDHAT_BUGZILLA_PRODUCT="Red Hat Enterprise Linux 8"
REDHAT_BUGZILLA_PRODUCT_VERSION=8.3
REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux"
REDHAT_SUPPORT_PRODUCT_VERSION="8.3"
Red Hat Enterprise Linux release 8.3 (Ootpa)
Red Hat Enterprise Linux release 8.3 (Ootpa)
[root@kontainerscomk ~]# sysctl -a | grep dirty
vm.dirty_background_bytes = 0
vm.dirty_background_ratio = 10
vm.dirty_bytes = 0
vm.dirty_expire_centisecs = 3000
vm.dirty_ratio = 30
vm.dirty_writeback_centisecs = 500
vm.dirtytime_expire_seconds = 43200
[root@kontainerscomk ~]# xfs_db -r /dev/dm-0
xfs_db: /dev/dm-0 is not a valid XFS filesystem (unexpected SB magic number 0xa82a0000)
Use -F to force a read attempt.
[root@kontainerscomk ~]# xfs_db -r /dev/dm-0 -F
xfs_db: /dev/dm-0 is not a valid XFS filesystem (unexpected SB magic number 0xa82a0000)
xfs_db: size check failed
xfs_db: V1 inodes unsupported. Please try an older xfsprogs.
[root@kontainerscomk ~]# cat /etc/fstab
#
# /etc/fstab
# Created by anaconda on Thu Nov 19 22:40:39 2020
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
/dev/mapper/rhel-root / xfs defaults 0 0
UUID=ad84d1ea-c9cc-4b22-8338-d1a6b2c7d27e /boot xfs defaults 0 0
UUID=4642-2FF6 /boot/efi vfat umask=0077,shortname=winnt 0 2
/dev/mapper/rhel-swap none swap defaults 0 0
Thanks,
-----Original Message-----
From: Strahil Nikolov <hunter86_bg(a)yahoo.com>
Sent: Sunday, November 29, 2020 2:33 PM
To: Vinícius Ferrão <ferrao(a)versatushpc.com.br>
Cc: users <users(a)ovirt.org>
Subject: Re: [ovirt-users] Re: Constantly XFS in memory corruption inside VMs
Can you check the output on the VM that was affected:
# cat /etc/*release
# sysctl -a | grep dirty
Best Regards,
Strahil Nikolov
В неделя, 29 ноември 2020 г., 19:07:48 Гринуич+2, Vinícius Ferrão via Users <users(a)ovirt.org> написа:
Hi Strahil.
I’m not using barrier options on mount. It’s the default settings from CentOS install.
I have some additional findings, there’s a big number of discarded packages on the switch on the hypervisor interfaces.
Discards are OK as far as I know, I hope TCP handles this and do the proper retransmissions, but I ask if this may be related or not. Our storage is over NFS. My general expertise is with iSCSI and I’ve never seen this kind of issue with iSCSI, not that I’m aware of.
In other clusters, I’ve seen a high number of discards with iSCSI on XenServer 7.2 but there’s no corruption on the VMs there...
Thanks,
Sent from my iPhone
> On 29 Nov 2020, at 04:00, Strahil Nikolov <hunter86_bg(a)yahoo.com> wrote:
>
> Are you using "nobarrier" mount options in the VM ?
>
> If yes, can you try to remove the "nobarrrier" option.
>
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В събота, 28 ноември 2020 г., 19:25:48 Гринуич+2, Vinícius Ferrão <ferrao(a)versatushpc.com.br> написа:
>
>
>
>
>
> Hi Strahil,
>
> I moved a running VM to other host, rebooted and no corruption was found. If there's any corruption it may be silent corruption... I've cases where the VM was new, just installed, run dnf -y update to get the updated packages, rebooted, and boom XFS corruption. So perhaps the motion process isn't the one to blame.
>
> But, in fact, I remember when moving a VM that it went down during the process and when I rebooted it was corrupted. But this may not seems related. It perhaps was already in a inconsistent state.
>
> Anyway, here's the mount options:
>
> Host1:
> 192.168.10.14:/mnt/pool0/ovirt/vm on
> /rhev/data-center/mnt/192.168.10.14:_mnt_pool0_ovirt_vm type nfs4
> (rw,relatime,vers=4.1,rsize=131072,wsize=131072,namlen=255,soft,noshar
> ecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=192.168.10.1,l
> ocal_lock=none,addr=192.168.10.14)
>
> Host2:
> 192.168.10.14:/mnt/pool0/ovirt/vm on
> /rhev/data-center/mnt/192.168.10.14:_mnt_pool0_ovirt_vm type nfs4
> (rw,relatime,vers=4.1,rsize=131072,wsize=131072,namlen=255,soft,noshar
> ecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=192.168.10.1,l
> ocal_lock=none,addr=192.168.10.14)
>
> The options are the default ones. I haven't changed anything when configuring this cluster.
>
> Thanks.
>
>
>
> -----Original Message-----
> From: Strahil Nikolov <hunter86_bg(a)yahoo.com>
> Sent: Saturday, November 28, 2020 1:54 PM
> To: users <users(a)ovirt.org>; Vinícius Ferrão
> <ferrao(a)versatushpc.com.br>
> Subject: Re: [ovirt-users] Constantly XFS in memory corruption inside
> VMs
>
> Can you try with a test vm, if this happens after a Virtual Machine migration ?
>
> What are your mount options for the storage domain ?
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В събота, 28 ноември 2020 г., 18:25:15 Гринуич+2, Vinícius Ferrão via Users <users(a)ovirt.org> написа:
>
>
>
>
>
>
>
>
> Hello,
>
>
>
> I’m trying to discover why an oVirt 4.4.3 Cluster with two hosts and NFS shared storage on TrueNAS 12.0 is constantly getting XFS corruption inside the VMs.
>
>
>
> For random reasons VM’s gets corrupted, sometimes halting it or just being silent corrupted and after a reboot the system is unable to boot due to “corruption of in-memory data detected”. Sometimes the corrupted data are “all zeroes”, sometimes there’s data there. In extreme cases the XFS superblock 0 get’s corrupted and the system cannot even detect a XFS partition anymore since the magic XFS key is corrupted on the first blocks of the virtual disk.
>
>
>
> This is happening for a month now. We had to rollback some backups, and I don’t trust anymore on the state of the VMs.
>
>
>
> Using xfs_db I can see that some VM’s have corrupted superblocks but the VM is up. One in specific, was with sb0 corrupted, so I knew when a reboot kicks in the machine will be gone, and that’s exactly what happened.
>
>
>
> Another day I was just installing a new CentOS 8 VM for random reasons, and after running dnf -y update and a reboot the VM was corrupted needing XFS repair. That was an extreme case.
>
>
>
> So, I’ve looked on the TrueNAS logs, and there’s apparently nothing wrong on the system. No errors logged on dmesg, nothing on /var/log/messages and no errors on the “zpools”, not even after scrub operations. On the switch, a Catalyst 2960X, we’ve been monitoring it and all it’s interfaces. There are no “up and down” and zero errors on all interfaces (we have a 4x Port LACP on the TrueNAS side and 2x Port LACP on each hosts), everything seems to be fine. The only metric that I was unable to get is “dropped packages”, but I’m don’t know if this can be an issue or not.
>
>
>
> Finally, on oVirt, I can’t find anything either. I looked on /var/log/messages and /var/log/sanlock.log but there’s nothing that I found suspicious.
>
>
>
> Is there’s anyone out there experiencing this? Our VM’s are mainly CentOS 7/8 with XFS, there’s 3 Windows VM’s that does not seems to be affected, everything else is affected.
>
>
>
> Thanks all.
>
>
>
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org Privacy
> Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VLYSE7HC
> FNWTWFZZTL2EJHV36OENHUGB/
_______________________________________________
Users mailing list -- users(a)ovirt.org
To unsubscribe send an email to users-leave(a)ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CZ5E55LJMA7...
4 years, 3 months
Using Ansible ovirt_vm
by Matthew.Stier@fujitsu.com
Over this weekend, I need to shutdown and restart 1500 VMs while I do work. Modifying an Ansible playbook I have, to accomplish that, seems easy enough.
Last weekend I used the web gui and edited each Virtual Machine and changed the 'cluster' to the one I wanted to assign it too. Now I would like to automate this with Ansible.
In between shutting them down and starting them up, I need to change the cluster registration of quite a few of the hosts, and I have concerns on how to do that.
The Ansible ovirt_vm documentation on this is a bit fuzzy. I'm assuming it for assigning the cluster, but in some examples it appears to be used as a filter. Which changes the meaning. Is it "start the VM in this cluster" or "start this VM if it is in this cluster".
The playbook I'm working from is straight forward and has three tasks.
First, an ovirt_auth task to log into the SHE.
Second, an ovirt_vm task which loops though the group of virtualmachines, defined in an 'ini' file.
Third, an ovirt_auth task to logout from the SHE.
For stopping, I assume I specify a name, and state to 'stopped' and that should initiate the shutdown.
For starting, I assume I should set the state to 'running'.
For changing the cluster a vm is assigned to, do I simply need to use the 'cluster' parameter with a value of new cluster? Are there other actions that need to be taken? The VM need to be un and re-registered for the change to take effect? I'd like to make the cluster change individual playbook, but could it be added to the start or stop playbook. (ie: stop and change cluster, start on new cluster)
Waiting for the wisdom of the community.
4 years, 3 months