Ok, so the issue seams here: now by default vdsm binds on :: and its
heuristc can end up using ipv6.
See this one:
Can you please try enabling ipv6 on your host or setting
management_ip = 0.0.0.0
under the [address] section in /etc/vdsm/vdsm.conf
and then restarting vdsm.
25.07.2016, 15:35, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
> On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>> # ss -plutn
>>
>> Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port
>>
>> udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6))
>> udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6))
>> udp UNCONN 0 0 127.0.0.1:323 *:* users:(("chronyd",pid=795,fd=1))
>> udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7))
>> udp UNCONN 0 0 127.0.0.1:25375 *:* users:(("snmpd",pid=1609,fd=8))
>> udp UNCONN 0 0 127.0.0.1:25376 *:* users:(("cmapeerd",pid=2056,fd=5))
>> udp UNCONN 0 0 127.0.0.1:25393 *:* users:(("cmanicd",pid=2278,fd=3))
>> udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9))
>> udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10))
>> tcp LISTEN 0 128 *:2381 *:*
users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4))
>> tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8))
>> tcp LISTEN 0 5 *:54322 *:* users:(("ovirt-imageio-d",pid=753,fd=3))
>> tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3))
>> tcp LISTEN 0 100 127.0.0.1:25 *:* users:(("master",pid=1948,fd=13))
>> tcp LISTEN 0 128 *:2301 *:*
users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
>> tcp LISTEN 0 30 *:16514 *:* users:(("libvirtd",pid=10688,fd=13))
>> tcp LISTEN 0 128 127.0.0.1:199 *:* users:(("snmpd",pid=1609,fd=9))
>> tcp LISTEN 0 128 :::111 :::* users:(("rpcbind",pid=827,fd=11))
>> tcp LISTEN 0 5 :::54321 :::* users:(("vdsm",pid=11077,fd=23))
>
> vdsm is properly bind over ipv6.
>
> Can you please check if you can connect to vdsm with:
> telnet
kom-ad01-vm31.holding.com 54321
> and with
> telnet ::1 54321
> ?
>
>> tcp LISTEN 0 30 :::16514 :::* users:(("libvirtd",pid=10688,fd=14))
>>
>> 25.07.2016, 15:11, "Simone Tiraboschi" <stirabos(a)redhat.com>:
>>> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>> Yes.
>>>>
>>>> # ping $(python -c 'import socket;
print(socket.gethostname())')
>>>>
>>>> PING
KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
>>>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64
time=0.030 ms
>>>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64
time=0.015 ms
>>>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64
time=0.011 ms
>>>> ^C
>>>> ---
KOM-AD01-VM31.holding.com ping statistics ---
>>>> 3 packets transmitted, 3 received, 0% packet loss, time 2001ms
>>>> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
>>>>
>>>> but...
>>>
>>> and the output of
>>> ss -plutn
>>>
>>>> # vdsClient -s 0 getVdsCaps
>>>>
>>>> Traceback (most recent call last):
>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in
<module>
>>>> code, message = commands[command][0](commandArgs)
>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in
do_getCap
>>>> return self.ExecAndExit(self.s.getVdsCapabilities())
>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in
__call__
>>>> return self.__send(self.__name, args)
>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in
__request
>>>> verbose=self.__verbose
>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in
request
>>>> return self.single_request(host, handler, request_body, verbose)
>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in
single_request
>>>> self.send_content(h, request_body)
>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in
send_content
>>>> connection.endheaders(request_body)
>>>> File "/usr/lib64/python2.7/httplib.py", line 975, in
endheaders
>>>> self._send_output(message_body)
>>>> File "/usr/lib64/python2.7/httplib.py", line 835, in
_send_output
>>>> self.send(msg)
>>>> File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>> self.connect()
>>>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py",
line 203, in connect
>>>> sock = socket.create_connection((self.host, self.port),
self.timeout)
>>>> File "/usr/lib64/python2.7/socket.py", line 571, in
create_connection
>>>> raise err
>>>> error: [Errno 101] Network is unreachable
>>>>
>>>> 25.07.2016, 14:58, "aleksey.maksimov(a)it-kb.ru"
<aleksey.maksimov(a)it-kb.ru>:
>>>>> Ok.
>>>>>
>>>>> 1) I stopped and disabled the service NetworkManager
>>>>> # systemctl stop NetworkManager
>>>>> # systemctl disable NetworkManager
>>>>>
>>>>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add
PEERDNS=no in ifcfg-* file.
>>>>>
>>>>> 3) Reboot server
>>>>>
>>>>> 4) Try deploy oVirt HE 4 and I get the same error
>>>>>
>>>>> [ INFO ] Creating Volume Group
>>>>> [ INFO ] Creating Storage Domain
>>>>> [ INFO ] Creating Storage Pool
>>>>> [ INFO ] Connecting Storage Pool
>>>>> [ INFO ] Verifying sanlock lockspace initialization
>>>>> [ ERROR ] Failed to execute stage 'Misc configuration':
[Errno 101] Network is unreachable
>>>>> [ INFO ] Stage: Clean up
>>>>> [ INFO ] Generating answer file
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
>>>>> [ INFO ] Stage: Pre-termination
>>>>> [ INFO ] Stage: Termination
>>>>> [ ERROR ] Hosted Engine deployment failed: this system is not
reliable, please check the issue, fix and redeploy
>>>>> Log file is located at
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
>>>>>
>>>>> What ideas further?
>>>>>
>>>>> 25.07.2016, 13:06, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
>>>>>> On Mon, Jul 25, 2016 at 11:54 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>> What am I supposed to do for successfully deploy ovirt 4
?
>>>>>>> Any ideas ?
>>>>>>
>>>>>> Can you please try to explicitly configure your DNS with
nameserver
>>>>>> under /etc/resolv.conf and remove DNS1 and DNS2 and set
PEERDNS=no for
>>>>>> the interface you are going to use?
>>>>>>
>>>>>>> 25.07.2016, 12:47, "aleksey.maksimov(a)it-kb.ru"
<aleksey.maksimov(a)it-kb.ru>:
>>>>>>>> "Aleksey, was your DNS configured with DNS1 and
DNS2 just on the interface you used to create the management bridge on?"
>>>>>>>>
>>>>>>>> Yes. Of course
>>>>>>>>
>>>>>>>> 25.07.2016, 12:27, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
>>>>>>>>> On Mon, Jul 25, 2016 at 10:56 AM, Piotr
Kliczewski
>>>>>>>>> <piotr.kliczewski(a)gmail.com> wrote:
>>>>>>>>>> This could be the issue here as well as for
BZ #1358530
>>>>>>>>>>
>>>>>>>>>> On Mon, Jul 25, 2016 at 10:53 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>> Could this be due to the fact that the
ovirt installer has changed network configuration files (ifcfg-*, resolv.conf) ?
>>>>>>>>>>> After the error in ovirt installation
process I see from resolv.conf disappeared on my DNS servers entry and now the server is
unable to resolve names.
>>>>>>>>>
>>>>>>>>> So it could be related to
https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>>>>>>>
>>>>>>>>> Aleksey, was your DNS configured with DNS1 and
DNS2 just on the
>>>>>>>>> interface you used to create the management
bridge on?
>>>>>>>>> Can you please try the workaround described
here
>>>>>>>>>
https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
>>>>>>>>>
>>>>>>>>>>> 25.07.2016, 11:26, "Simone
Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>> On Mon, Jul 25, 2016 at 10:22 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>> # vdsClient -s 0 getVdsCaps
>>>>>>>>>>>>>
>>>>>>>>>>>>> Traceback (most recent call
last):
>>>>>>>>>>>>> File
"/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>>>>>>>>>>>> code, message =
commands[command][0](commandArgs)
>>>>>>>>>>>>> File
"/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>>>>>>> return
self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>>>>>>>>>>> return
self.__send(self.__name, args)
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>>>>>>>>>>> verbose=self.__verbose
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>>>>>>>>>>> return
self.single_request(host, handler, request_body, verbose)
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>>>>>>>>>>>>> self.send_content(h,
request_body)
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>>>>>>>>>>>>>
connection.endheaders(request_body)
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>>>>>>>>>>>
self._send_output(message_body)
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>>>>>>>>>>> self.send(msg)
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>>>>>>>>>> self.connect()
>>>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect
>>>>>>>>>>>>> sock =
socket.create_connection((self.host, self.port), self.timeout)
>>>>>>>>>>>>> File
"/usr/lib64/python2.7/socket.py", line 571, in create_connection
>>>>>>>>>>>>> raise err
>>>>>>>>>>>>> error: [Errno 101] Network is
unreachable
>>>>>>>>>>>>
>>>>>>>>>>>> Yaniv, can you please take also a
look to this one?
>>>>>>>>>>>> it's exactly the opposite of
https://bugzilla.redhat.com/1358530
>>>>>>>>>>>> Here the jsonrpcclient works but
not the xmlrpc one.
>>>>>>>>>>>>
>>>>>>>>>>>>> 25.07.2016, 11:17,
"Simone Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>>>> On Mon, Jul 25, 2016 at
7:51 AM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>>>> Simone, there is
something interesting in the vdsm.log?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> For what I saw the issue
is not related to the storage but to the network.
>>>>>>>>>>>>>> ovirt-hosted-engine-setup
uses the jsonrpc client, instead the code
>>>>>>>>>>>>>> from
ovirt-hosted-engine-ha still uses the xmlrpc client somewhere and
>>>>>>>>>>>>>> this happens also when the
setup asks to create the lockspace volume.
>>>>>>>>>>>>>> It seams that in your case
the xmlrpc client could not connect vdsm on
>>>>>>>>>>>>>> the localhost.
>>>>>>>>>>>>>> It could be somehow
related to:
>>>>>>>>>>>>>>
https://bugzilla.redhat.com/1358530
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Can you please try
executing
>>>>>>>>>>>>>> sudo vdsClient -s 0
getVdsCaps
>>>>>>>>>>>>>> on that host?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> 22.07.2016, 19:36,
"aleksey.maksimov(a)it-kb.ru" <aleksey.maksimov(a)it-kb.ru>:
>>>>>>>>>>>>>>>> Simone, thanks
for link.
>>>>>>>>>>>>>>>> vdsm.log
attached
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 22.07.2016,
19:28, "Simone Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>>>>>>> On Fri, Jul
22, 2016 at 5:59 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>>>>>>> Thank
you for your response, Simone.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Log
attached.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> It seams it
comes from VDSM, can you please attach also vdsm.log?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I
don't use ovirt-engine-appliance because I have not found "how-to" for
ovirt-engine-appliance deployment in hosted engine configuration.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> yum install
ovirt-engine-appliance
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Then follow
the instruction here:
>>>>>>>>>>>>>>>>>
http://www.ovirt.org/develop/release-management/features/heapplianceflow/
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
22.07.2016, 17:09, "Simone Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>>>>>>>>> Hi
Aleksey,
>>>>>>>>>>>>>>>>>>> Can
you please attach hosted-engine-setup logs?
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On
Fri, Jul 22, 2016 at 3:46 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
Hello oVirt guru`s !
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
I have problem with initial deploy of ovirt 4.0 hosted engine.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
My environment :
>>>>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>>>>>
* Two servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected (with
>>>>>>>>>>>>>>>>>>>>
multipathd) to storage HP 3PAR 7200
>>>>>>>>>>>>>>>>>>>>
* On each server installed CentOS 7.2 Linux (3.10.0-327.22.2.el7.x86_64)
>>>>>>>>>>>>>>>>>>>>
* On 3PAR storage I created 2 LUNs for oVirt.
>>>>>>>>>>>>>>>>>>>>
- First LUN for oVirt Hosted Engine VM (60GB)
>>>>>>>>>>>>>>>>>>>>
- Second LUN for all other VMs (2TB)
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
# multipath -ll
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV
>>>>>>>>>>>>>>>>>>>>
size=60G features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>>>>
`-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>>>>
|- 2:0:1:1 sdd 8:48 active ready running
>>>>>>>>>>>>>>>>>>>>
|- 3:0:0:1 sdf 8:80 active ready running
>>>>>>>>>>>>>>>>>>>>
|- 2:0:0:1 sdb 8:16 active ready running
>>>>>>>>>>>>>>>>>>>>
`- 3:0:1:1 sdh 8:112 active ready running
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
3par-vv2 (360002ac000000000000000160000cec9) dm-1 3PARdata,VV
>>>>>>>>>>>>>>>>>>>>
size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>>>>
`-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>>>>
|- 2:0:0:0 sda 8:0 active ready running
>>>>>>>>>>>>>>>>>>>>
|- 3:0:0:0 sde 8:64 active ready running
>>>>>>>>>>>>>>>>>>>>
|- 2:0:1:0 sdc 8:32 active ready running
>>>>>>>>>>>>>>>>>>>>
`- 3:0:1:0 sdg 8:96 active ready running
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
My steps on first server (initial deploy of ovirt 4.0 hosted engine):
>>>>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
# systemctl stop NetworkManager
>>>>>>>>>>>>>>>>>>>>
# systemctl disable NetworkManager
>>>>>>>>>>>>>>>>>>>>
# yum -y install
http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
>>>>>>>>>>>>>>>>>>>>
# yum -y install epel-release
>>>>>>>>>>>>>>>>>>>>
# wget
>>>>>>>>>>>>>>>>>>>>
http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1...
>>>>>>>>>>>>>>>>>>>>
-P /tmp/
>>>>>>>>>>>>>>>>>>>>
# yum install ovirt-hosted-engine-setup
>>>>>>>>>>>>>>>>>>>>
# yum install screen
>>>>>>>>>>>>>>>>>>>>
# screen -RD
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
...in screen session :
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
# hosted-engine --deploy
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
...
>>>>>>>>>>>>>>>>>>>>
in configuration process I chose "fc" as storage type for oVirt hosted
>>>>>>>>>>>>>>>>>>>>
engine vm and select 60GB LUN...
>>>>>>>>>>>>>>>>>>>>
...
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
--== CONFIGURATION PREVIEW ==--
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
...
>>>>>>>>>>>>>>>>>>>>
Firewall manager : iptables
>>>>>>>>>>>>>>>>>>>>
Gateway address : 10.1.0.1
>>>>>>>>>>>>>>>>>>>>
Host name for web application : KOM-AD01-OVIRT1
>>>>>>>>>>>>>>>>>>>>
Storage Domain type : fc
>>>>>>>>>>>>>>>>>>>>
Host ID : 1
>>>>>>>>>>>>>>>>>>>>
LUN ID :
>>>>>>>>>>>>>>>>>>>>
360002ac0000000000000001b0000cec9
>>>>>>>>>>>>>>>>>>>>
Image size GB : 40
>>>>>>>>>>>>>>>>>>>>
Console type : vnc
>>>>>>>>>>>>>>>>>>>>
Memory size MB : 4096
>>>>>>>>>>>>>>>>>>>>
MAC address : 00:16:3e:77:1d:07
>>>>>>>>>>>>>>>>>>>>
Boot type : cdrom
>>>>>>>>>>>>>>>>>>>>
Number of CPUs : 2
>>>>>>>>>>>>>>>>>>>>
ISO image (cdrom boot/cloud-init) :
>>>>>>>>>>>>>>>>>>>>
/tmp/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Can
I ask why you prefer/need to manually create a VM installing from
>>>>>>>>>>>>>>>>>>> a
CD instead of using the ready-to-use ovirt-engine-appliance?
>>>>>>>>>>>>>>>>>>>
Using the appliance makes the setup process a lot shorted and more comfortable.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
CPU Type : model_Penryn
>>>>>>>>>>>>>>>>>>>>
...
>>>>>>>>>>>>>>>>>>>>
and get error after step "Verifying sanlock lockspace initialization"
>>>>>>>>>>>>>>>>>>>>
...
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>>>>>
[ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network
>>>>>>>>>>>>>>>>>>>>
is unreachable
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Clean up
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Generating answer file
>>>>>>>>>>>>>>>>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Pre-termination
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Termination
>>>>>>>>>>>>>>>>>>>>
[ ERROR ] Hosted Engine deployment failed: this system is not reliable,
>>>>>>>>>>>>>>>>>>>>
please check the issue, fix and redeploy
>>>>>>>>>>>>>>>>>>>>
Log file is located at
>>>>>>>>>>>>>>>>>>>>
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
Interestingly
>>>>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>>>>>
If I try to deploy hosted-engine v3.6, everything goes well in the same
>>>>>>>>>>>>>>>>>>>>
configuration !! :
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
....
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Transaction setup
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Package installation
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Configuring libvirt
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Configuring VDSM
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Starting vdsmd
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Waiting for VDSM hardware info
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Configuring the management bridge
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Volume Group
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Storage Domain
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Storage Pool
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Connecting Storage Pool
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Image for 'hosted-engine.lockspace' ...
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Image for 'hosted-engine.lockspace' created successfully
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Image for 'hosted-engine.metadata' ...
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Image for 'hosted-engine.metadata' created successfully
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating VM Image
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Destroying Storage Pool
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Start monitoring domain
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Configuring VM
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Updating hosted-engine configuration
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Transaction commit
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Closing up
>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating VM
>>>>>>>>>>>>>>>>>>>>
You can now connect to the VM with the following command:
>>>>>>>>>>>>>>>>>>>>
/bin/remote-viewer vnc://localhost:5900
>>>>>>>>>>>>>>>>>>>>
...
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
What could be the problem?
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>>>>>
Users mailing list
>>>>>>>>>>>>>>>>>>>>
Users(a)ovirt.org
>>>>>>>>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>> Users mailing list
>>>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list
>>>>>>>> Users(a)ovirt.org
>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> Users(a)ovirt.org
>>>>>
http://lists.ovirt.org/mailman/listinfo/users