--------------070306020106050808000908
Content-Type: text/plain; charset="utf-8"; format=flowed
Content-Transfer-Encoding: 7bit
I followed the instructions here
<
http://www.ovirt.org/OVirt_3.6_Release_Notes> on upgrading to Ovirt
3.6RC from 3.5.4 and have encountered a few problems.
My Ovirt 3.5.4 test environment consisted of:
1 Centos 7.1 host running hosted engine stored on a separate NFS server
1 Centos 7.1 ovirt engine vm
With some research I was able to solve two of the three issues I've
experienced. I'll list them here for acadamia - and perhaps they point
to a misstep on my behalf that is causing the third.
1. Upon a "successful" upgrade, the admin@local account was expired.
The problem is documented here
<
https://bugzilla.redhat.com/show_bug.cgi?id=1261382> and currently
caused by following the upgrade instructions as seen here
<
http://www.ovirt.org/OVirt_3.6_Release_Notes#Install_.2F_Upgrade_from_pre...;.
Solution was to do the following from the ovirt-engine vm (they may not
have all been necessary, it was late!):
a. ovirt-aaa-jdbc-tool
--db-config=/etc/ovirt-engine/aaa/internal.properties user
password-reset admin --force
b. ovirt-aaa-jdbc-tool
--db-config=/etc/ovirt-engine/aaa/internal.properties user
password-reset admin --password-valid-to="2019-01-01 12:00:00Z"
c. ovirt-aaa-jdbc-tool
--db-config=/etc/ovirt-engine/aaa/internal.properties user edit admin
--account-valid-from="2014-01-01 12:00:00Z"
--account-valid-to="2019-01-01 12:00:00Z"
d. ovirt-aaa-jdbc-tool
--db-config=/etc/ovirt-engine/aaa/internal.properties user unlock admin
2. Rebooting the Centos 7.1 host caused a loss of default gateway. The
engine does not allow you to modify the host because it is in use and
modifying /etc/sysconfig/network-scripts is undone by VDSM upon the next
reboot. I assume in the past this worked okay because I had a
GATEWAY=xxxx in /etc/sysconfig/network as a pre-Ovirt relic. Solution
here was to add gateway and defaultRoute fields using the vdsClient
command line utility:
a. vdsClient -s 0 setupNetworks
networks='{ovirtmgmt:{ipaddr:10.1.0.21,netmask:255.255.254.0,bonding:bond0,bridged:true,gateway:10.1.1.254,defaultRoute:True}}'
b. vdsClient -s 0 setSafeNetworkConfig
Now for the issue I can't solve. When I reboot the Centos 7.1 host I
get the following:
[root@ovirt-one /]# hosted-engine --vm-status
You must run deploy first
I then notice that the NFS share to the hosted engine is not mounted and
the ovirt-ha-agent.service has failed to start itself at boot.
[root@ovirt-one /]# systemctl status ovirt-ha-agent.service
ovirt-ha-agent.service - oVirt Hosted Engine High Availability
Monitoring Agent
Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-agent.service; enabled)
Active: failed (Result: exit-code) since Tue 2015-09-29 12:17:55
CDT; 9min ago
Process: 1424 ExecStop=/usr/lib/systemd/systemd-ovirt-ha-agent stop
(code=exited, status=0/SUCCESS)
Process: 1210 ExecStart=/usr/lib/systemd/systemd-ovirt-ha-agent start
(code=exited, status=0/SUCCESS)
Main PID: 1377 (code=exited, status=254)
CGroup: /system.slice/ovirt-ha-agent.service
Sep 29 12:17:55
ovirt-one.thaultanklines.com
systemd-ovirt-ha-agent[1210]: Starting ovirt-ha-agent: [ OK ]
Sep 29 12:17:55
ovirt-one.thaultanklines.com systemd[1]: Started oVirt
Hosted Engine High Availability Monitoring Agent.
Sep 29 12:17:55
ovirt-one.thaultanklines.com ovirt-ha-agent[1377]:
ovirt-ha-agent ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine
ERROR Service vdsmd is not running and the admin is responsible for
starting it. Shutting down.
Sep 29 12:17:55
ovirt-one.thaultanklines.com systemd[1]:
ovirt-ha-agent.service: main process exited, code=exited, status=254/n/a
Sep 29 12:17:55
ovirt-one.thaultanklines.com systemd[1]: Unit
ovirt-ha-agent.service entered failed state.
Manually starting ovirt-ha-agent.service works and it then correctly
mounts the hosted engine NFS share and all works and I can eventually
start the hosted engine. Why would the ovirt-ha-agent.service attempt
to start before VDSM was ready?
Snippet from /usr/lib/systemd/system/ovirt-ha-agent.service
[Unit]
Description=oVirt Hosted Engine High Availability Monitoring Agent
Wants=ovirt-ha-broker.service
Wants=vdsmd.service
Wants=sanlock.service
After=ovirt-ha-broker.service
Any help would be appreciated!
--------------070306020106050808000908
Content-Type: text/html; charset="utf-8"
Content-Transfer-Encoding: 8bit
<html>
<head>
<meta http-equiv="content-type" content="text/html;
charset=utf-8">
</head>
<body bgcolor="#FFFFFF" text="#000000">
I followed the instructions <a
href="http://www.ovirt.org/OVirt_3.6_Release_Notes">here<... on
upgrading to Ovirt 3.6RC from 3.5.4 and have encountered a few
problems.<br>
<br>
My Ovirt 3.5.4 test environment consisted of:<br>
<br>
1 Centos 7.1 host running hosted engine stored on a separate NFS
server<br>
1 Centos 7.1 ovirt engine vm<br>
<br>
With some research I was able to solve two of the three issues I've
experienced. I'll list them here for acadamia - and perhaps they
point to a misstep on my behalf that is causing the third.<br>
<br>
1. Upon a "successful" upgrade, the admin@local account was
expired. The problem is <a
href="https://bugzilla.redhat.com/show_bug.cgi?id=1261382">d...
here</a> and currently caused by following the upgrade
instructions <a
href="http://www.ovirt.org/OVirt_3.6_Release_Notes#Install_.2F_Upgra...
seen here</a>. Solution was to do the following from the
ovirt-engine vm (they may not have all been necessary, it was
late!):<br>
a. ovirt-aaa-jdbc-tool
--db-config=/etc/ovirt-engine/aaa/internal.properties user
password-reset admin --force<br>
b. ovirt-aaa-jdbc-tool
--db-config=/etc/ovirt-engine/aaa/internal.properties user
password-reset admin --password-valid-to="2019-01-01 12:00:00Z"<br>
c. ovirt-aaa-jdbc-tool
--db-config=/etc/ovirt-engine/aaa/internal.properties user edit
admin --account-valid-from="2014-01-01 12:00:00Z"
--account-valid-to="2019-01-01 12:00:00Z"<br>
d. ovirt-aaa-jdbc-tool
--db-config=/etc/ovirt-engine/aaa/internal.properties user unlock
admin<br>
<br>
2. Rebooting the Centos 7.1 host caused a loss of default gateway.
The engine does not allow you to modify the host because it is in
use and modifying /etc/sysconfig/network-scripts is undone by VDSM
upon the next reboot. I assume in the past this worked okay because
I had a GATEWAY=xxxx in /etc/sysconfig/network as a pre-Ovirt
relic. Solution here was to add gateway and defaultRoute fields
using the vdsClient command line utility:<br>
a. vdsClient -s 0 setupNetworks
networks='{ovirtmgmt:{ipaddr:10.1.0.21,netmask:255.255.254.0,bonding:bond0,bridged:true,gateway:10.1.1.254,defaultRoute:True}}'<br>
b. vdsClient -s 0 setSafeNetworkConfig<br>
<br>
Now for the issue I can't solve. When I reboot the Centos 7.1 host
I get the following:<br>
<br>
[root@ovirt-one /]# hosted-engine --vm-status<br>
You must run deploy first<br>
<br>
I then notice that the NFS share to the hosted engine is not mounted
and the ovirt-ha-agent.service has failed to start itself at boot.<br>
<br>
[root@ovirt-one /]# systemctl status ovirt-ha-agent.service<br>
ovirt-ha-agent.service - oVirt Hosted Engine High Availability
Monitoring Agent<br>
Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-agent.service;
enabled)<br>
Active: failed (Result: exit-code) since Tue 2015-09-29 12:17:55
CDT; 9min ago<br>
Process: 1424 ExecStop=/usr/lib/systemd/systemd-ovirt-ha-agent
stop (code=exited, status=0/SUCCESS)<br>
Process: 1210 ExecStart=/usr/lib/systemd/systemd-ovirt-ha-agent
start (code=exited, status=0/SUCCESS)<br>
Main PID: 1377 (code=exited, status=254)<br>
CGroup: /system.slice/ovirt-ha-agent.service<br>
<br>
Sep 29 12:17:55
ovirt-one.thaultanklines.com
systemd-ovirt-ha-agent[1210]: Starting ovirt-ha-agent: [ OK ]<br>
Sep 29 12:17:55
ovirt-one.thaultanklines.com systemd[1]: Started
oVirt Hosted Engine High Availability Monitoring Agent.<br>
Sep 29 12:17:55
ovirt-one.thaultanklines.com ovirt-ha-agent[1377]:
ovirt-ha-agent
ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine ERROR
Service vdsmd is not running and the admin is responsible for
starting it. Shutting down.<br>
Sep 29 12:17:55
ovirt-one.thaultanklines.com systemd[1]:
ovirt-ha-agent.service: main process exited, code=exited,
status=254/n/a<br>
Sep 29 12:17:55
ovirt-one.thaultanklines.com systemd[1]: Unit
ovirt-ha-agent.service entered failed state.<br>
<br>
Manually starting ovirt-ha-agent.service works and it then correctly
mounts the hosted engine NFS share and all works and I can
eventually start the hosted engine. Why would the
ovirt-ha-agent.service attempt to start before VDSM was ready?<br>
<br>
Snippet from /usr/lib/systemd/system/ovirt-ha-agent.service<br>
[Unit]<br>
Description=oVirt Hosted Engine High Availability Monitoring Agent<br>
Wants=ovirt-ha-broker.service<br>
Wants=vdsmd.service<br>
Wants=sanlock.service<br>
After=ovirt-ha-broker.service<br>
<br>
Any help would be appreciated!<br>
<br>
</body>
</html>
--------------070306020106050808000908--