This is a multi-part message in MIME format.
--------------070402000308020002020408
Content-Type: text/plain; charset=utf-8
Content-Transfer-Encoding: 7bit
On 03/10/2015 10:20 AM, Simone Tiraboschi wrote:
----- Original Message -----
> From: "Bob Doolittle" <bob(a)doolittle.us.com>
> To: "Simone Tiraboschi" <stirabos(a)redhat.com>
> Cc: "users-ovirt" <users(a)ovirt.org>
> Sent: Tuesday, March 10, 2015 2:40:13 PM
> Subject: Re: [ovirt-users] Error during hosted-engine-setup for 3.5.1 on F20 (The
VDSM host was found in a failed
> state)
>
>
> On 03/10/2015 04:58 AM, Simone Tiraboschi wrote:
>> ----- Original Message -----
>>> From: "Bob Doolittle" <bob(a)doolittle.us.com>
>>> To: "Simone Tiraboschi" <stirabos(a)redhat.com>
>>> Cc: "users-ovirt" <users(a)ovirt.org>
>>> Sent: Monday, March 9, 2015 11:48:03 PM
>>> Subject: Re: [ovirt-users] Error during hosted-engine-setup for 3.5.1 on
>>> F20 (The VDSM host was found in a failed
>>> state)
>>>
>>>
>>> On 03/09/2015 02:47 PM, Bob Doolittle wrote:
>>>> Resending with CC to list (and an update).
>>>>
>>>> On 03/09/2015 01:40 PM, Simone Tiraboschi wrote:
>>>>> ----- Original Message -----
>>>>>> From: "Bob Doolittle" <bob(a)doolittle.us.com>
>>>>>> To: "Simone Tiraboschi" <stirabos(a)redhat.com>
>>>>>> Cc: "users-ovirt" <users(a)ovirt.org>
>>>>>> Sent: Monday, March 9, 2015 6:26:30 PM
>>>>>> Subject: Re: [ovirt-users] Error during hosted-engine-setup for
3.5.1
>>>>>> on
>>>>>> F20 (Cannot add the host to cluster ... SSH
>>>>>> has failed)
>>>>>>
> ...
>>>>>> OK, I've started over. Simply removing the storage domain
was
>>>>>> insufficient,
>>>>>> the hosted-engine deploy failed when it found the HA and Broker
>>>>>> services
>>>>>> already configured. I decided to just start over fresh starting
with
>>>>>> re-installing the OS on my host.
>>>>>>
>>>>>> I can't deploy DNS at the moment, so I have to simply
replicate
>>>>>> /etc/hosts
>>>>>> files on my host/engine. I did that this time, but have run into
a new
>>>>>> problem:
>>>>>>
>>>>>> [ INFO ] Engine replied: DB Up!Welcome to Health Status!
>>>>>> Enter the name of the cluster to which you want to add
the
>>>>>> host
>>>>>> (Default) [Default]:
>>>>>> [ INFO ] Waiting for the host to become operational in the
engine.
>>>>>> This
>>>>>> may
>>>>>> take several minutes...
>>>>>> [ ERROR ] The VDSM host was found in a failed state. Please
check
>>>>>> engine
>>>>>> and
>>>>>> bootstrap installation logs.
>>>>>> [ ERROR ] Unable to add ovirt-vm to the manager
>>>>>> Please shutdown the VM allowing the system to launch it
as a
>>>>>> monitored service.
>>>>>> The system will wait until the VM is down.
>>>>>> [ ERROR ] Failed to execute stage 'Closing up': [Errno
111] Connection
>>>>>> refused
>>>>>> [ INFO ] Stage: Clean up
>>>>>> [ ERROR ] Failed to execute stage 'Clean up': [Errno 111]
Connection
>>>>>> refused
>>>>>>
>>>>>>
>>>>>> I've attached my engine log and the ovirt-hosted-engine-setup
log. I
>>>>>> think I
>>>>>> had an issue with resolving external hostnames, or else a
connectivity
>>>>>> issue
>>>>>> during the install.
>>>>> For some reason your engine wasn't able to deploy your hosts but
the SSH
>>>>> session this time was established.
>>>>> 2015-03-09 13:05:58,514 ERROR
>>>>> [org.ovirt.engine.core.bll.InstallVdsInternalCommand]
>>>>> (org.ovirt.thread.pool-8-thread-3) [3cf91626] Host installation
failed
>>>>> for host 217016bb-fdcd-4344-a0ca-4548262d10a8, ovirt-vm.:
>>>>> java.io.IOException: Command returned failure code 1 during SSH
session
>>>>> 'root(a)xion2.smartcity.net'
>>>>>
>>>>> Can you please attach host-deploy logs from the engine VM?
>>>> OK, attached.
>>>>
>>>> Like I said, it looks to me like a name-resolution issue during the yum
>>>> update on the engine. I think I've fixed that, but do you have a
better
>>>> suggestion for cleaning up and re-deploying other than installing the OS
>>>> on my host and starting all over again?
>>> I just finished starting over from scratch, starting with OS installation
>>> on
>>> my host/node, and wound up with a very similar problem - the engine
>>> couldn't
>>> reach the hosts during the yum operation. But this time the error was
>>> "Network is unreachable". Which is weird, because I can ssh into
the
>>> engine
>>> and ping many of those hosts, after the operation has failed.
>>>
>>> Here's my latest host-deploy log from the engine. I'd appreciate any
>>> clues.
>> It seams that now your host is able to resolve that addresses but it's not
>> able to connect over http.
>> On your hosts some of them resolves as IPv6 addresses; can you please try
>> to use curl to get one of the file that it wasn't able to fetch?
>> Can you please check your network configuration before and after
>> host-deploy?
> I can give you the network configuration after host-deploy, at least for the
> host/Node. The engine won't start for me this morning, after I shut down the
> host for the night.
>
> In order to give you the config before host-deploy (or, apparently for the
> engine), I'll have to re-install the OS on the host and start again from
> scratch. Obviously I'd rather not do that unless absolutely necessary.
>
> Here's the host config after the failed host-deploy:
>
> Host/Node:
>
> # ip route
> 169.254.0.0/16 dev ovirtmgmt scope link metric 1007
> 172.16.0.0/16 dev ovirtmgmt proto kernel scope link src 172.16.0.58
You are missing a default gateway and so the issue.
Are you sure that it was properly configured before trying to deploy that host?
It should have been, it was a fresh OS install. So I'm starting again, and keeping
careful records of my network config.
Here is my initial network config of my host/node, immediately following a new OS
install:
% ip route
default via 172.16.0.1 dev p3p1 proto static metric 1024
172.16.0.0/16 dev p3p1 proto kernel scope link src 172.16.0.58
% ip addr
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: p3p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP group
default qlen 1000
link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
inet 172.16.0.58/16 brd 172.16.255.255 scope global p3p1
valid_lft forever preferred_lft forever
inet6 fe80::baca:3aff:fe79:2212/64 scope link
valid_lft forever preferred_lft forever
3: wlp2s0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN group
default qlen 1000
link/ether 1c:3e:84:50:8d:c3 brd ff:ff:ff:ff:ff:ff
After the VM is first created, the host/node config is:
# ip route
default via 172.16.0.1 dev ovirtmgmt
169.254.0.0/16 dev ovirtmgmt scope link metric 1006
172.16.0.0/16 dev ovirtmgmt proto kernel scope link src 172.16.0.58
# ip addr
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: p3p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master
ovirtmgmt state UP group default qlen 1000
link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
inet6 fe80::baca:3aff:fe79:2212/64 scope link
valid_lft forever preferred_lft forever
3: wlp2s0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN group
default qlen 1000
link/ether 1c:3e:84:50:8d:c3 brd ff:ff:ff:ff:ff:ff
4: bond0: <NO-CARRIER,BROADCAST,MULTICAST,MASTER,UP> mtu 1500 qdisc noqueue state
DOWN group default
link/ether 92:cb:9d:97:18:36 brd ff:ff:ff:ff:ff:ff
5: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default
link/ether 9a:bc:29:52:82:38 brd ff:ff:ff:ff:ff:ff
6: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP
group default
link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
inet 172.16.0.58/16 brd 172.16.255.255 scope global ovirtmgmt
valid_lft forever preferred_lft forever
inet6 fe80::baca:3aff:fe79:2212/64 scope link
valid_lft forever preferred_lft forever
7: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master
ovirtmgmt state UNKNOWN group default qlen 500
link/ether fe:16:3e:16:a4:37 brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc16:3eff:fe16:a437/64 scope link
valid_lft forever preferred_lft forever
At this point, I was already seeing a problem on the host/node. I remembered that a newer
version of sos package is delivered from the ovirt repositories. So I tried to do a
"yum update" on my host, and got a similar problem:
% sudo yum update
[sudo] password for rad:
Loaded plugins: langpacks, refresh-packagekit
Resolving Dependencies
--> Running transaction check
---> Package sos.noarch 0:3.1-1.fc20 will be updated
---> Package sos.noarch 0:3.2-0.2.fc20.ovirt will be an update
--> Finished Dependency Resolution
Dependencies Resolved
================================================================================================================
Package Arch Version Repository
Size
================================================================================================================
Updating:
sos noarch 3.2-0.2.fc20.ovirt ovirt-3.5
292 k
Transaction Summary
================================================================================================================
Upgrade 1 Package
Total download size: 292 k
Is this ok [y/d/N]: y
Downloading packages:
No Presto metadata available for ovirt-3.5
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
http://www.gtlib.gatech.edu/pub/oVirt/pub/ovirt-3.5/rpm/fc20/noarch/sos-3...:
[Errno 14] curl#6 - "Could not resolve host:
www.gtlib.gatech.edu"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
ftp://ftp.gtlib.gatech.edu/pub/oVirt/pub/ovirt-3.5/rpm/fc20/noarch/sos-3....:
[Errno 14] curl#6 - "Could not resolve host: ftp.gtlib.gatech.edu"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
http://resources.ovirt.org/pub/ovirt-3.5/rpm/fc20/noarch/sos-3.2-0.2.fc20...:
[Errno 14] curl#6 - "Could not resolve host: resources.ovirt.org"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
http://ftp.snt.utwente.nl/pub/software/ovirt/ovirt-3.5/rpm/fc20/noarch/so...:
[Errno 14] curl#6 - "Could not resolve host: ftp.snt.utwente.nl"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
http://ftp.nluug.nl/os/Linux/virtual/ovirt/ovirt-3.5/rpm/fc20/noarch/sos-...:
[Errno 14] curl#6 - "Could not resolve host: ftp.nluug.nl"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
http://mirror.linux.duke.edu/ovirt/pub/ovirt-3.5/rpm/fc20/noarch/sos-3.2-...:
[Errno 14] curl#6 - "Could not resolve host: mirror.linux.duke.edu"
Trying other mirror.
Error downloading packages:
sos-3.2-0.2.fc20.ovirt.noarch: [Errno 256] No more mirrors to try.
This was similar to my previous failures. I took a look, and the problem was that
/etc/resolv.conf had no nameservers, and the
/etc/sysconfig/network-scripts/ifcfg-ovirtmgmt file contained no entries for DNS1 or
DOMAIN.
So, it appears that when hosted-engine set up my bridged network, it neglected to carry
over the DNS configuration necessary to the bridge.
Note that I am using *static* network configuration, rather than DHCP. During installation
of the OS I am setting up the network configuration as Manual. Perhaps the hosted-engine
script is not properly prepared to deal with that?
I went ahead and modified the ifcfg-ovirtmgmt network script (for the next service
restart/boot) and resolv.conf (I was afraid to restart the network in the middle of
hosted-engine execution since I don't know what might already be connected to the
engine). This time it got further, but ultimately it still failed at the very end:
[ INFO ] Waiting for the host to become operational in the engine. This may take several
minutes...
[ INFO ] Still waiting for VDSM host to become operational...
[ INFO ] The VDSM Host is now operational
Please shutdown the VM allowing the system to launch it as a monitored service.
The system will wait until the VM is down.
[ ERROR ] Failed to execute stage 'Closing up': Error acquiring VM status
[ INFO ] Stage: Clean up
[ INFO ] Generating answer file
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20150310140028.conf'
[ INFO ] Stage: Pre-termination
[ INFO ] Stage: Termination
At that point, neither the ovirt-ha-broker or ovirt-ha-agent services were running.
Note there was no significant pause after it said "The system will wait until the VM
is down".
After the script completed, I shut down the VM, and manually started the ha services, and
the VM came up. I could login to the Administration Portal, and finally see my
HostedEngine VM. :-)
I seem to be in a bad state however: The Data Center has *no* storage domains attached.
I'm not sure what else might need cleaning up. Any assistance appreciated.
-Bob
> # ip addr
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group
> default
> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> inet 127.0.0.1/8 scope host lo
> valid_lft forever preferred_lft forever
> inet6 ::1/128 scope host
> valid_lft forever preferred_lft forever
> 2: p3p2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master
> ovirtmgmt state UP group default qlen 1000
> link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
> inet6 fe80::baca:3aff:fe79:2212/64 scope link
> valid_lft forever preferred_lft forever
> 3: bond0: <NO-CARRIER,BROADCAST,MULTICAST,MASTER,UP> mtu 1500 qdisc noqueue
> state DOWN group default
> link/ether 56:56:f7:cf:73:27 brd ff:ff:ff:ff:ff:ff
> 4: wlp2s0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN
> group default qlen 1000
> link/ether 1c:3e:84:50:8d:c3 brd ff:ff:ff:ff:ff:ff
> 6: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
> default
> link/ether 22:a1:01:9e:30:71 brd ff:ff:ff:ff:ff:ff
> 7: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
> UP group default
> link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
> inet 172.16.0.58/16 brd 172.16.255.255 scope global ovirtmgmt
> valid_lft forever preferred_lft forever
> inet6 fe80::baca:3aff:fe79:2212/64 scope link
> valid_lft forever preferred_lft forever
>
>
> The only unusual thing about my setup that I can think of, from the network
> perspective, is that my physical host has a wireless interface, which I've
> not configured. Could it be confusing hosted-engine --deploy?
>
> -Bob
>
>
--------------070402000308020002020408
Content-Type: text/html; charset=utf-8
Content-Transfer-Encoding: 7bit
<html>
<head>
<meta content="text/html; charset=utf-8"
http-equiv="Content-Type">
</head>
<body bgcolor="#FFFFFF" text="#000000">
<br>
<div class="moz-cite-prefix">On 03/10/2015 10:20 AM, Simone
Tiraboschi wrote:<br>
</div>
<blockquote
cite="mid:330401686.19087599.1425997248583.JavaMail.zimbra@redhat.com"
type="cite">
<pre wrap="">
----- Original Message -----
</pre>
<blockquote type="cite">
<pre wrap="">From: "Bob Doolittle" <a
class="moz-txt-link-rfc2396E"
href="mailto:bob@doolittle.us.com"><bob@doolittle.us.com></a>
To: "Simone Tiraboschi" <a class="moz-txt-link-rfc2396E"
href="mailto:stirabos@redhat.com"><stirabos@redhat.com></a>
Cc: "users-ovirt" <a class="moz-txt-link-rfc2396E"
href="mailto:users@ovirt.org"><users@ovirt.org></a>
Sent: Tuesday, March 10, 2015 2:40:13 PM
Subject: Re: [ovirt-users] Error during hosted-engine-setup for 3.5.1 on F20 (The VDSM
host was found in a failed
state)
On 03/10/2015 04:58 AM, Simone Tiraboschi wrote:
</pre>
<blockquote type="cite">
<pre wrap="">
----- Original Message -----
</pre>
<blockquote type="cite">
<pre wrap="">From: "Bob Doolittle" <a
class="moz-txt-link-rfc2396E"
href="mailto:bob@doolittle.us.com"><bob@doolittle.us.com></a>
To: "Simone Tiraboschi" <a class="moz-txt-link-rfc2396E"
href="mailto:stirabos@redhat.com"><stirabos@redhat.com></a>
Cc: "users-ovirt" <a class="moz-txt-link-rfc2396E"
href="mailto:users@ovirt.org"><users@ovirt.org></a>
Sent: Monday, March 9, 2015 11:48:03 PM
Subject: Re: [ovirt-users] Error during hosted-engine-setup for 3.5.1 on
F20 (The VDSM host was found in a failed
state)
On 03/09/2015 02:47 PM, Bob Doolittle wrote:
</pre>
<blockquote type="cite">
<pre wrap="">Resending with CC to list (and an update).
On 03/09/2015 01:40 PM, Simone Tiraboschi wrote:
</pre>
<blockquote type="cite">
<pre wrap="">----- Original Message -----
</pre>
<blockquote type="cite">
<pre wrap="">From: "Bob Doolittle" <a
class="moz-txt-link-rfc2396E"
href="mailto:bob@doolittle.us.com"><bob@doolittle.us.com></a>
To: "Simone Tiraboschi" <a class="moz-txt-link-rfc2396E"
href="mailto:stirabos@redhat.com"><stirabos@redhat.com></a>
Cc: "users-ovirt" <a class="moz-txt-link-rfc2396E"
href="mailto:users@ovirt.org"><users@ovirt.org></a>
Sent: Monday, March 9, 2015 6:26:30 PM
Subject: Re: [ovirt-users] Error during hosted-engine-setup for 3.5.1
on
F20 (Cannot add the host to cluster ... SSH
has failed)
</pre>
</blockquote>
</blockquote>
</blockquote>
</blockquote>
</blockquote>
<pre wrap="">...
</pre>
<blockquote type="cite">
<blockquote type="cite">
<blockquote type="cite">
<blockquote type="cite">
<blockquote type="cite">
<pre wrap="">OK, I've started over. Simply removing
the storage domain was
insufficient,
the hosted-engine deploy failed when it found the HA and Broker
services
already configured. I decided to just start over fresh starting with
re-installing the OS on my host.
I can't deploy DNS at the moment, so I have to simply replicate
/etc/hosts
files on my host/engine. I did that this time, but have run into a new
problem:
[ INFO ] Engine replied: DB Up!Welcome to Health Status!
Enter the name of the cluster to which you want to add the
host
(Default) [Default]:
[ INFO ] Waiting for the host to become operational in the engine.
This
may
take several minutes...
[ ERROR ] The VDSM host was found in a failed state. Please check
engine
and
bootstrap installation logs.
[ ERROR ] Unable to add ovirt-vm to the manager
Please shutdown the VM allowing the system to launch it as a
monitored service.
The system will wait until the VM is down.
[ ERROR ] Failed to execute stage 'Closing up': [Errno 111] Connection
refused
[ INFO ] Stage: Clean up
[ ERROR ] Failed to execute stage 'Clean up': [Errno 111] Connection
refused
I've attached my engine log and the ovirt-hosted-engine-setup log. I
think I
had an issue with resolving external hostnames, or else a connectivity
issue
during the install.
</pre>
</blockquote>
<pre wrap="">For some reason your engine wasn't able
to deploy your hosts but the SSH
session this time was established.
2015-03-09 13:05:58,514 ERROR
[org.ovirt.engine.core.bll.InstallVdsInternalCommand]
(org.ovirt.thread.pool-8-thread-3) [3cf91626] Host installation failed
for host 217016bb-fdcd-4344-a0ca-4548262d10a8, ovirt-vm.:
java.io.IOException: Command returned failure code 1 during SSH session
'<a class="moz-txt-link-abbreviated"
href="mailto:root@xion2.smartcity.net">root@xion2.smartcity.net</a>'
Can you please attach host-deploy logs from the engine VM?
</pre>
</blockquote>
<pre wrap="">OK, attached.
Like I said, it looks to me like a name-resolution issue during the yum
update on the engine. I think I've fixed that, but do you have a better
suggestion for cleaning up and re-deploying other than installing the OS
on my host and starting all over again?
</pre>
</blockquote>
<pre wrap="">I just finished starting over from scratch,
starting with OS installation
on
my host/node, and wound up with a very similar problem - the engine
couldn't
reach the hosts during the yum operation. But this time the error was
"Network is unreachable". Which is weird, because I can ssh into the
engine
and ping many of those hosts, after the operation has failed.
Here's my latest host-deploy log from the engine. I'd appreciate any
clues.
</pre>
</blockquote>
<pre wrap="">It seams that now your host is able to resolve that
addresses but it's not
able to connect over http.
On your hosts some of them resolves as IPv6 addresses; can you please try
to use curl to get one of the file that it wasn't able to fetch?
Can you please check your network configuration before and after
host-deploy?
</pre>
</blockquote>
<pre wrap="">
I can give you the network configuration after host-deploy, at least for the
host/Node. The engine won't start for me this morning, after I shut down the
host for the night.
In order to give you the config before host-deploy (or, apparently for the
engine), I'll have to re-install the OS on the host and start again from
scratch. Obviously I'd rather not do that unless absolutely necessary.
Here's the host config after the failed host-deploy:
Host/Node:
# ip route
169.254.0.0/16 dev ovirtmgmt scope link metric 1007
172.16.0.0/16 dev ovirtmgmt proto kernel scope link src 172.16.0.58
</pre>
</blockquote>
<pre wrap="">
You are missing a default gateway and so the issue.
Are you sure that it was properly configured before trying to deploy that
host?</pre>
</blockquote>
<br>
It should have been, it was a fresh OS install. So I'm starting
again, and keeping careful records of my network config.<br>
<br>
Here is my initial network config of my host/node, immediately
following a new OS install:<br>
<br>
<pre>% ip route
default via 172.16.0.1 dev p3p1 proto static metric 1024
172.16.0.0/16 dev p3p1 proto kernel scope link src 172.16.0.58
% ip addr
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group
default
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: p3p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state
UP group default qlen 1000
link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
inet 172.16.0.58/16 brd 172.16.255.255 scope global p3p1
valid_lft forever preferred_lft forever
inet6 fe80::baca:3aff:fe79:2212/64 scope link
valid_lft forever preferred_lft forever
3: wlp2s0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN
group default qlen 1000
link/ether 1c:3e:84:50:8d:c3 brd ff:ff:ff:ff:ff:ff
</pre>
<br>
After the VM is first created, the host/node config is:<br>
<br>
<pre># ip route
default via 172.16.0.1 dev ovirtmgmt
169.254.0.0/16 dev ovirtmgmt scope link metric 1006
172.16.0.0/16 dev ovirtmgmt proto kernel scope link src 172.16.0.58
# ip addr
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group
default
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: p3p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master
ovirtmgmt state UP group default qlen 1000
link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
inet6 fe80::baca:3aff:fe79:2212/64 scope link
valid_lft forever preferred_lft forever
3: wlp2s0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN
group default qlen 1000
link/ether 1c:3e:84:50:8d:c3 brd ff:ff:ff:ff:ff:ff
4: bond0: <NO-CARRIER,BROADCAST,MULTICAST,MASTER,UP> mtu 1500 qdisc noqueue
state DOWN group default
link/ether 92:cb:9d:97:18:36 brd ff:ff:ff:ff:ff:ff
5: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
default
link/ether 9a:bc:29:52:82:38 brd ff:ff:ff:ff:ff:ff
6: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state
UP group default
link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
inet 172.16.0.58/16 brd 172.16.255.255 scope global ovirtmgmt
valid_lft forever preferred_lft forever
inet6 fe80::baca:3aff:fe79:2212/64 scope link
valid_lft forever preferred_lft forever
7: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master
ovirtmgmt state UNKNOWN group default qlen 500
link/ether fe:16:3e:16:a4:37 brd ff:ff:ff:ff:ff:ff
inet6 fe80::fc16:3eff:fe16:a437/64 scope link
valid_lft forever preferred_lft forever
</pre>
<br>
At this point, I was already seeing a problem on the host/node. I
remembered that a newer version of sos package is delivered from the
ovirt repositories. So I tried to do a "yum update" on my host, and
got a similar problem:<br>
<br>
<pre>% sudo yum update
[sudo] password for rad:
Loaded plugins: langpacks, refresh-packagekit
Resolving Dependencies
--> Running transaction check
---> Package sos.noarch 0:3.1-1.fc20 will be updated
---> Package sos.noarch 0:3.2-0.2.fc20.ovirt will be an update
--> Finished Dependency Resolution
Dependencies Resolved
================================================================================================================
Package Arch Version Repository
Size
================================================================================================================
Updating:
sos noarch 3.2-0.2.fc20.ovirt ovirt-3.5
292 k
Transaction Summary
================================================================================================================
Upgrade 1 Package
Total download size: 292 k
Is this ok [y/d/N]: y
Downloading packages:
No Presto metadata available for ovirt-3.5
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
<a class="moz-txt-link-freetext"
href="http://www.gtlib.gatech.edu/pub/oVirt/pub/ovirt-3.5/rpm/fc20/n...;:
[Errno 14] curl#6 - "Could not resolve host: <a
class="moz-txt-link-abbreviated"
href="http://www.gtlib.gatech.edu">www.gtlib.gatech.edu</a>"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
<a class="moz-txt-link-freetext"
href="ftp://ftp.gtlib.gatech.edu/pub/oVirt/pub/ovirt-3.5/rpm/fc20/no...;:
[Errno 14] curl#6 - "Could not resolve host: <a
class="moz-txt-link-abbreviated"
href="ftp://ftp.gtlib.gatech.edu">ftp.gtlib.gatech.edu</a>"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
<a class="moz-txt-link-freetext"
href="http://resources.ovirt.org/pub/ovirt-3.5/rpm/fc20/noarch/sos-3...;:
[Errno 14] curl#6 - "Could not resolve host: resources.ovirt.org"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
<a class="moz-txt-link-freetext"
href="http://ftp.snt.utwente.nl/pub/software/ovirt/ovirt-3.5/rpm/fc20/noarch/sos-3.2-0.2.fc20.ovirt.noarch.rpm">http://ftp.snt.utwente.nl/pub/software/ovirt/ovirt-3.5/rpm/fc20/noarch/sos-3.2-0.2.fc20.ovirt.noarch.rpm</a>:
[Errno 14] curl#6 - "Could not resolve host: <a
class="moz-txt-link-abbreviated"
href="ftp://ftp.snt.utwente.nl">ftp.snt.utwente.nl</a>"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
<a class="moz-txt-link-freetext"
href="http://ftp.nluug.nl/os/Linux/virtual/ovirt/ovirt-3.5/rpm/fc20/noarch/sos-3.2-0.2.fc20.ovirt.noarch.rpm">http://ftp.nluug.nl/os/Linux/virtual/ovirt/ovirt-3.5/rpm/fc20/noarch/sos-3.2-0.2.fc20.ovirt.noarch.rpm</a>:
[Errno 14] curl#6 - "Could not resolve host: <a
class="moz-txt-link-abbreviated"
href="ftp://ftp.nluug.nl">ftp.nluug.nl</a>"
Trying other mirror.
sos-3.2-0.2.fc20.ovirt.noarch. FAILED
<a class="moz-txt-link-freetext"
href="http://mirror.linux.duke.edu/ovirt/pub/ovirt-3.5/rpm/fc20/noar...;:
[Errno 14] curl#6 - "Could not resolve host: mirror.linux.duke.edu"
Trying other mirror.
Error downloading packages:
sos-3.2-0.2.fc20.ovirt.noarch: [Errno 256] No more mirrors to try.
</pre>
<br>
This was similar to my previous failures. I took a look, and the
problem was that /etc/resolv.conf had no nameservers, and the
/etc/sysconfig/network-scripts/ifcfg-ovirtmgmt file contained no
entries for DNS1 or DOMAIN.<br>
<br>
So, it appears that when hosted-engine set up my bridged network, it
neglected to carry over the DNS configuration necessary to the
bridge.<br>
<br>
Note that I am using *static* network configuration, rather than
DHCP. During installation of the OS I am setting up the network
configuration as Manual. Perhaps the hosted-engine script is not
properly prepared to deal with that?<br>
<br>
I went ahead and modified the ifcfg-ovirtmgmt network script (for
the next service restart/boot) and resolv.conf (I was afraid to
restart the network in the middle of hosted-engine execution since I
don't know what might already be connected to the engine). This time
it got further, but ultimately it still failed at the very end:<br>
<br>
<pre>[ INFO ] Waiting for the host to become operational in the engine. This
may take several minutes...
[ INFO ] Still waiting for VDSM host to become operational...
[ INFO ] The VDSM Host is now operational
Please shutdown the VM allowing the system to launch it as a monitored service.
The system will wait until the VM is down.
[ ERROR ] Failed to execute stage 'Closing up': Error acquiring VM status
[ INFO ] Stage: Clean up
[ INFO ] Generating answer file
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20150310140028.conf'
[ INFO ] Stage: Pre-termination
[ INFO ] Stage: Termination
</pre>
<br>
At that point, neither the ovirt-ha-broker or ovirt-ha-agent
services were running.<br>
<br>
Note there was no significant pause after it said "The system will
wait until the VM is down".<br>
<br>
After the script completed, I shut down the VM, and manually started
the ha services, and the VM came up. I could login to the
Administration Portal, and finally see my HostedEngine VM. :-)<br>
<br>
I seem to be in a bad state however: The Data Center has <b>no</b>
storage domains attached. I'm not sure what else might need cleaning
up. Any assistance appreciated.<br>
<br>
-Bob<br>
<br>
<br>
<blockquote
cite="mid:330401686.19087599.1425997248583.JavaMail.zimbra@redhat.com"
type="cite">
<blockquote type="cite">
<pre wrap=""># ip addr
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group
default
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: p3p2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master
ovirtmgmt state UP group default qlen 1000
link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
inet6 fe80::baca:3aff:fe79:2212/64 scope link
valid_lft forever preferred_lft forever
3: bond0: <NO-CARRIER,BROADCAST,MULTICAST,MASTER,UP> mtu 1500 qdisc noqueue
state DOWN group default
link/ether 56:56:f7:cf:73:27 brd ff:ff:ff:ff:ff:ff
4: wlp2s0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN
group default qlen 1000
link/ether 1c:3e:84:50:8d:c3 brd ff:ff:ff:ff:ff:ff
6: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
default
link/ether 22:a1:01:9e:30:71 brd ff:ff:ff:ff:ff:ff
7: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state
UP group default
link/ether b8:ca:3a:79:22:12 brd ff:ff:ff:ff:ff:ff
inet 172.16.0.58/16 brd 172.16.255.255 scope global ovirtmgmt
valid_lft forever preferred_lft forever
inet6 fe80::baca:3aff:fe79:2212/64 scope link
valid_lft forever preferred_lft forever
The only unusual thing about my setup that I can think of, from the network
perspective, is that my physical host has a wireless interface, which I've
not configured. Could it be confusing hosted-engine --deploy?
-Bob
</pre>
</blockquote>
</blockquote>
<br>
</body>
</html>
--------------070402000308020002020408--