Quoting Yaniv Kaul <ykaul@redhat.com>:
On Sun, Dec 24, 2017 at 4:34 AM, Jason Keltz <jas@cse.yorku.ca> wrote:
Can you specify what did you upgrade, and in which order? Engine, hosts?
On 12/23/2017 5:38 PM, Jason Keltz wrote:
Hi..
I took the plunge to 4.2, but I think maybe I should have waited a bit...
Cluster level, etc.?
I was running 4.1.8 everywhere. I upgraded engine (standalone) to 4.2, then the 4 hosts. I stopped ovirt-engine, added the new repo for 4.2, ran the yum update of ovirt setup, ran engine-setup and that process worked flawlessly. No errors. I had just upgraded to 4.1.8 a few days ago, so all my ovirt infrastructure was running latest ovirt and I also upgraded engine and hosts to latest CentOS and latest kernel with the last 4.1.8 update. I then upgraded cluster level. All the VMs were going to be upgraded as they were rebooted, and since it's the reboot that breaks console, and since a reinstall brings it back, I'm going to assume it's the switch from 4.1 to 4.2 cluster that breaks it. If I submit this as a bug then what log/logs would I submit?
Yes. VM type. Most of the VMs switched from desktop to server after the update.
You mean the VM type?
Initially, after upgrade to 4.2, the status of many of my hosts changed
from "server" to "desktop". That's okay - I can change them back.
I will do that later today.
Can you share the complete logs? Best if you could file a bug about it,My first VM, "archive", I had the ability to access console after thenetwork. I was going to try detatching the disk from the existing dist
upgrade. I rebooted archive, and I lost the ability (option is grayed
out). The VM boots, but I need access to the console.
My second VM is called "dist". That one, ovirt says is running, but I
can't access it, can't ping it, and there's no console either, so I
literally can't get to it. I can reboot it, and shut it down, but it would
be helpful to be able to access it. What to do?
I reinstalled "dist" because I needed the VM to be accessible on the
server, and attaching it to a new dist VM, but I ended up inadvertently
deleting the disk image. I can't believe that under "storage" you can't
detatch a disk from a VM - you can only delete the disk.
After reinstalling dist, I got back console, and network access! I tried
rebooting it several times, and console remains... so the loss of console
has something to do with switching from a 4.1 VM to 4.2.
I've very afraid to reboot my engine because it seems like when I reboot
hosts, I lose access to console.lost it (at least network access remains). Now that this situation is
I rebooted one more VM for which I had console access, and again, I've
repeatable, I'm going one of the ovirt gurus can send me the magical DB
command to fix it. Probably not a solution to reinstall my 37 VMs from
kickstart.. that would be a headache.
In addition, when I try to check for "host updates", I get an error that
it can't check for host updates. I ran a yum update on the hosts (after
upgrading repo to 4.2 and doing a yum update) and all I'm looking for it to
do is clear status, but it doesn't seem to work.
The error in engine.log when I try to update any of the hosts is:
2017-12-23 19:11:36,479-05 INFO [org.ovirt.engine.core.bll.hos
tdeploy.HostUpgradeCheckCommand] (default task-156)
[ae11a704-3b40-45d3-9850-932f6ed91ed9] Running command:
HostUpgradeCheckCommand internal: false. Entities affected : ID:
45f8b331-842e-48e7-9df8-56adddb93836 Type: VDSAction group
EDIT_HOST_CONFIGURATION with role type ADMIN
2017-12-23 19:11:36,496-05 INFO [org.ovirt.engine.core.dal.dbb
roker.auditloghandling.AuditLogDirector] (default task-156) [] EVENT_ID:
HOST_AVAILABLE_UPDATES_STARTED(884), Started to check for available
updates on host virt1.
2017-12-23 19:11:36,500-05 INFO [org.ovirt.engine.core.bll.hos
tdeploy.HostUpgradeCheckInternalCommand] (EE-ManagedThreadFactory-comma ndCoordinator-Thread-7)
[ae11a704-3b40-45d3-9850-932f6ed91ed9] Running command:
HostUpgradeCheckInternalCommand internal: true. Entities affected : ID:
45f8b331-842e-48e7-9df8-56adddb93836 Type: VDS
2017-12-23 19:11:36,504-05 INFO [org.ovirt.engine.core.common.utils.ansible.AnsibleExecutor]
(EE-ManagedThreadFactory-commandCoordinator-Thread-7)
[ae11a704-3b40-45d3-9850-932f6ed91ed9] Executing Ansible command:
ANSIBLE_STDOUT_CALLBACK=hostupgradeplugin [/usr/bin/ansible-playbook,
--check, --private-key=/etc/pki/ovirt-engine/keys/engine_id_rsa,
--inventory=/tmp/ansible-inventory1039100972039373314,
/usr/share/ovirt-engine/playbooks/ovirt-host-upgrade.yml] [Logfile: null]
2017-12-23 19:11:37,897-05 INFO [org.ovirt.engine.core.common.utils.ansible.AnsibleExecutor]
(EE-ManagedThreadFactory-commandCoordinator-Thread-7)
[ae11a704-3b40-45d3-9850-932f6ed91ed9] Ansible playbook command has
exited with value: 4
2017-12-23 19:11:37,897-05 ERROR [org.ovirt.engine.core.bll.host.HostUpgradeManager]
(EE-ManagedThreadFactory-commandCoordinator-Thread-7)
[ae11a704-3b40-45d3-9850-932f6ed91ed9] Failed to run check-update of host
'virt1-mgmt'.
2017-12-23 19:11:37,897-05 ERROR [org.ovirt.engine.core.bll.hostdeploy.HostUpdatesChecker]
(EE-ManagedThreadFactory-commandCoordinator-Thread-7)
[ae11a704-3b40-45d3-9850-932f6ed91ed9] Failed to check if updates are
available for host 'virt1' with error message 'Failed to run check-update
of host 'virt1-mgmt'.'
2017-12-23 19:11:37,904-05 ERROR [org.ovirt.engine.core.dal.dbb
roker.auditloghandling.AuditLogDirector] (EE-ManagedThreadFactory-comma ndCoordinator-Thread-7)
[ae11a704-3b40-45d3-9850-932f6ed91ed9] EVENT_ID:
HOST_AVAILABLE_UPDATES_FAILED(839), Failed to check for available updates
on host virt1 with message 'Failed to run check-update of host
'virt1-mgmt'.'.
with the logs attached.
Y.
Thanks!
Jason.
--
dfjklgdfgkljdfgklfdklgj
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users