Edi, danken, this is again the malfunctioning heuristics in the
client for
handling ipv6. Is there a bug on this issue ?
On Jul 27, 2016 8:57 AM, <aleksey.maksimov(a)it-kb.ru> wrote:
>
> I enabled ipv6 for "lo" and "ovirtmgmt" interfaces and deployment
process
> overt completed successfully.
>
> # cat /etc/sysctl.conf
>
> net.ipv6.conf.all.disable_ipv6 = 1
> net.ipv6.conf.default.disable_ipv6 = 1
> net.ipv6.conf.lo.disable_ipv6 = 0
> net.ipv6.conf.ovirtmgmt.disable_ipv6 = 0
>
> Thank you for your Help!
>
> 26.07.2016, 09:13, "Edward Haas" <ehaas(a)redhat.com>:
>
>
>
> On Mon, Jul 25, 2016 at 4:02 PM, Simone Tiraboschi <stirabos(a)redhat.com>
> wrote:
>
> On Mon, Jul 25, 2016 at 2:38 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
> > telnet
kom-ad01-vm31.holding.com 54321 = success connection
> >
> > telnet ::1 54321
> > Trying ::1...
> > telnet: connect to address ::1: Network is unreachable
> >
> > (ipv6 on my server disabled)
>
> Ok, so the issue seams here: now by default vdsm binds on :: and its
> heuristc can end up using ipv6.
> See this one:
https://bugzilla.redhat.com/show_bug.cgi?id=1350883
>
> Can you please try enabling ipv6 on your host or setting
> management_ip = 0.0.0.0
> under the [address] section in /etc/vdsm/vdsm.conf
> and then restarting vdsm.
>
>
>
> Could you please also add the 'ip addr' output? Just interested to see how
> IPv6 was
> disabled on the host.
> It will be even better if you could apply the patch
> (
https://gerrit.ovirt.org/#/c/60020) and check.
>
>
>
>
>
> > 25.07.2016, 15:35, "Simone Tiraboschi" <stirabos(a)redhat.com>:
> >> On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
> >>> # ss -plutn
> >>>
> >>> Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port
> >>>
> >>> udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6))
> >>> udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6))
> >>> udp UNCONN 0 0 127.0.0.1:323 *:*
users:(("chronyd",pid=795,fd=1))
> >>> udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7))
> >>> udp UNCONN 0 0 127.0.0.1:25375 *:*
users:(("snmpd",pid=1609,fd=8))
> >>> udp UNCONN 0 0 127.0.0.1:25376 *:*
users:(("cmapeerd",pid=2056,fd=5))
> >>> udp UNCONN 0 0 127.0.0.1:25393 *:*
users:(("cmanicd",pid=2278,fd=3))
> >>> udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9))
> >>> udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10))
> >>> tcp LISTEN 0 128 *:2381 *:*
> >>>
users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4))
> >>> tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8))
> >>> tcp LISTEN 0 5 *:54322 *:*
users:(("ovirt-imageio-d",pid=753,fd=3))
> >>> tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3))
> >>> tcp LISTEN 0 100 127.0.0.1:25 *:*
users:(("master",pid=1948,fd=13))
> >>> tcp LISTEN 0 128 *:2301 *:*
> >>>
users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
> >>> tcp LISTEN 0 30 *:16514 *:*
users:(("libvirtd",pid=10688,fd=13))
> >>> tcp LISTEN 0 128 127.0.0.1:199 *:*
users:(("snmpd",pid=1609,fd=9))
> >>> tcp LISTEN 0 128 :::111 :::*
users:(("rpcbind",pid=827,fd=11))
> >>> tcp LISTEN 0 5 :::54321 :::*
users:(("vdsm",pid=11077,fd=23))
> >>
> >> vdsm is properly bind over ipv6.
> >>
> >> Can you please check if you can connect to vdsm with:
> >> telnet
kom-ad01-vm31.holding.com 54321
> >> and with
> >> telnet ::1 54321
> >> ?
> >>
> >>> tcp LISTEN 0 30 :::16514 :::*
users:(("libvirtd",pid=10688,fd=14))
> >>>
> >>> 25.07.2016, 15:11, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
> >>>> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov(a)it-kb.ru>
wrote:
> >>>>> Yes.
> >>>>>
> >>>>> # ping $(python -c 'import socket;
print(socket.gethostname())')
> >>>>>
> >>>>> PING
KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of
data.
> >>>>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231):
icmp_seq=1
> >>>>> ttl=64 time=0.030 ms
> >>>>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231):
icmp_seq=2
> >>>>> ttl=64 time=0.015 ms
> >>>>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231):
icmp_seq=3
> >>>>> ttl=64 time=0.011 ms
> >>>>> ^C
> >>>>> ---
KOM-AD01-VM31.holding.com ping statistics ---
> >>>>> 3 packets transmitted, 3 received, 0% packet loss, time
2001ms
> >>>>> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
> >>>>>
> >>>>> but...
> >>>>
> >>>> and the output of
> >>>> ss -plutn
> >>>>
> >>>>> # vdsClient -s 0 getVdsCaps
> >>>>>
> >>>>> Traceback (most recent call last):
> >>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in
<module>
> >>>>> code, message = commands[command][0](commandArgs)
> >>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in
do_getCap
> >>>>> return self.ExecAndExit(self.s.getVdsCapabilities())
> >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line
1233, in __call__
> >>>>> return self.__send(self.__name, args)
> >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line
1587, in
> >>>>> __request
> >>>>> verbose=self.__verbose
> >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line
1273, in request
> >>>>> return self.single_request(host, handler, request_body,
> >>>>> verbose)
> >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line
1301, in
> >>>>> single_request
> >>>>> self.send_content(h, request_body)
> >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line
1448, in
> >>>>> send_content
> >>>>> connection.endheaders(request_body)
> >>>>> File "/usr/lib64/python2.7/httplib.py", line 975,
in endheaders
> >>>>> self._send_output(message_body)
> >>>>> File "/usr/lib64/python2.7/httplib.py", line 835,
in
> >>>>> _send_output
> >>>>> self.send(msg)
> >>>>> File "/usr/lib64/python2.7/httplib.py", line 797,
in send
> >>>>> self.connect()
> >>>>> File
"/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line
> >>>>> 203, in connect
> >>>>> sock = socket.create_connection((self.host, self.port),
> >>>>> self.timeout)
> >>>>> File "/usr/lib64/python2.7/socket.py", line 571,
in
> >>>>> create_connection
> >>>>> raise err
> >>>>> error: [Errno 101] Network is unreachable
> >>>>>
> >>>>> 25.07.2016, 14:58, "aleksey.maksimov(a)it-kb.ru"
> >>>>> <aleksey.maksimov(a)it-kb.ru>:
> >>>>>> Ok.
> >>>>>>
> >>>>>> 1) I stopped and disabled the service NetworkManager
> >>>>>> # systemctl stop NetworkManager
> >>>>>> # systemctl disable NetworkManager
> >>>>>>
> >>>>>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add
> >>>>>> PEERDNS=no in ifcfg-* file.
> >>>>>>
> >>>>>> 3) Reboot server
> >>>>>>
> >>>>>> 4) Try deploy oVirt HE 4 and I get the same error
> >>>>>>
> >>>>>> [ INFO ] Creating Volume Group
> >>>>>> [ INFO ] Creating Storage Domain
> >>>>>> [ INFO ] Creating Storage Pool
> >>>>>> [ INFO ] Connecting Storage Pool
> >>>>>> [ INFO ] Verifying sanlock lockspace initialization
> >>>>>> [ ERROR ] Failed to execute stage 'Misc
configuration': [Errno
> >>>>>> 101] Network is unreachable
> >>>>>> [ INFO ] Stage: Clean up
> >>>>>> [ INFO ] Generating answer file
> >>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
> >>>>>> [ INFO ] Stage: Pre-termination
> >>>>>> [ INFO ] Stage: Termination
> >>>>>> [ ERROR ] Hosted Engine deployment failed: this system is
not
> >>>>>> reliable, please check the issue, fix and redeploy
> >>>>>> Log file is located at
> >>>>>>
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
> >>>>>>
> >>>>>> What ideas further?
> >>>>>>
> >>>>>> 25.07.2016, 13:06, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
> >>>>>>> On Mon, Jul 25, 2016 at 11:54 AM,
<aleksey.maksimov(a)it-kb.ru>
> >>>>>>> wrote:
> >>>>>>>> What am I supposed to do for successfully deploy
ovirt 4 ?
> >>>>>>>> Any ideas ?
> >>>>>>>
> >>>>>>> Can you please try to explicitly configure your DNS
with
> >>>>>>> nameserver
> >>>>>>> under /etc/resolv.conf and remove DNS1 and DNS2 and
set
> >>>>>>> PEERDNS=no for
> >>>>>>> the interface you are going to use?
> >>>>>>>
> >>>>>>>> 25.07.2016, 12:47,
"aleksey.maksimov(a)it-kb.ru"
> >>>>>>>> <aleksey.maksimov(a)it-kb.ru>:
> >>>>>>>>> "Aleksey, was your DNS configured with
DNS1 and DNS2 just on
> >>>>>>>>> the interface you used to create the management
bridge on?"
> >>>>>>>>>
> >>>>>>>>> Yes. Of course
> >>>>>>>>>
> >>>>>>>>> 25.07.2016, 12:27, "Simone
Tiraboschi"
> >>>>>>>>> <stirabos(a)redhat.com>:
> >>>>>>>>>> On Mon, Jul 25, 2016 at 10:56 AM, Piotr
Kliczewski
> >>>>>>>>>> <piotr.kliczewski(a)gmail.com>
wrote:
> >>>>>>>>>>> This could be the issue here as
well as for BZ #1358530
> >>>>>>>>>>>
> >>>>>>>>>>> On Mon, Jul 25, 2016 at 10:53 AM,
> >>>>>>>>>>> <aleksey.maksimov(a)it-kb.ru>
wrote:
> >>>>>>>>>>>> Could this be due to the fact
that the ovirt installer
> >>>>>>>>>>>> has changed network configuration
files (ifcfg-*, resolv.conf) ?
> >>>>>>>>>>>> After the error in ovirt
installation process I see
> >>>>>>>>>>>> from resolv.conf disappeared on my
DNS servers entry and now the server is
> >>>>>>>>>>>> unable to resolve names.
> >>>>>>>>>>
> >>>>>>>>>> So it could be related to
> >>>>>>>>>>
https://bugzilla.redhat.com/show_bug.cgi?id=1160423
> >>>>>>>>>>
> >>>>>>>>>> Aleksey, was your DNS configured with
DNS1 and DNS2 just
> >>>>>>>>>> on the
> >>>>>>>>>> interface you used to create the
management bridge on?
> >>>>>>>>>> Can you please try the workaround
described here
> >>>>>>>>>>
https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
> >>>>>>>>>>
> >>>>>>>>>>>> 25.07.2016, 11:26,
"Simone Tiraboschi"
> >>>>>>>>>>>> <stirabos(a)redhat.com>:
> >>>>>>>>>>>>> On Mon, Jul 25, 2016 at
10:22 AM,
> >>>>>>>>>>>>>
<aleksey.maksimov(a)it-kb.ru> wrote:
> >>>>>>>>>>>>>> # vdsClient -s 0
getVdsCaps
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> Traceback (most
recent call last):
> >>>>>>>>>>>>>> File
"/usr/share/vdsm/vdsClient.py", line 2980, in
> >>>>>>>>>>>>>> <module>
> >>>>>>>>>>>>>> code, message =
> >>>>>>>>>>>>>>
commands[command][0](commandArgs)
> >>>>>>>>>>>>>> File
"/usr/share/vdsm/vdsClient.py", line 543, in
> >>>>>>>>>>>>>> do_getCap
> >>>>>>>>>>>>>> return
> >>>>>>>>>>>>>>
self.ExecAndExit(self.s.getVdsCapabilities())
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line
> >>>>>>>>>>>>>> 1233, in __call__
> >>>>>>>>>>>>>> return
self.__send(self.__name, args)
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line
> >>>>>>>>>>>>>> 1587, in __request
> >>>>>>>>>>>>>>
verbose=self.__verbose
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line
> >>>>>>>>>>>>>> 1273, in request
> >>>>>>>>>>>>>> return
self.single_request(host, handler,
> >>>>>>>>>>>>>> request_body, verbose)
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line
> >>>>>>>>>>>>>> 1301, in single_request
> >>>>>>>>>>>>>>
self.send_content(h, request_body)
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line
> >>>>>>>>>>>>>> 1448, in send_content
> >>>>>>>>>>>>>>
connection.endheaders(request_body)
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 975,
> >>>>>>>>>>>>>> in endheaders
> >>>>>>>>>>>>>>
self._send_output(message_body)
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 835,
> >>>>>>>>>>>>>> in _send_output
> >>>>>>>>>>>>>> self.send(msg)
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 797,
> >>>>>>>>>>>>>> in send
> >>>>>>>>>>>>>> self.connect()
> >>>>>>>>>>>>>> File
> >>>>>>>>>>>>>>
"/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect
> >>>>>>>>>>>>>> sock =
socket.create_connection((self.host,
> >>>>>>>>>>>>>> self.port), self.timeout)
> >>>>>>>>>>>>>> File
"/usr/lib64/python2.7/socket.py", line 571,
> >>>>>>>>>>>>>> in create_connection
> >>>>>>>>>>>>>> raise err
> >>>>>>>>>>>>>> error: [Errno 101]
Network is unreachable
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Yaniv, can you please take
also a look to this one?
> >>>>>>>>>>>>> it's exactly the
opposite of
> >>>>>>>>>>>>>
https://bugzilla.redhat.com/1358530
> >>>>>>>>>>>>> Here the jsonrpcclient
works but not the xmlrpc one.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>> 25.07.2016, 11:17,
"Simone Tiraboschi"
> >>>>>>>>>>>>>>
<stirabos(a)redhat.com>:
> >>>>>>>>>>>>>>> On Mon, Jul 25,
2016 at 7:51 AM,
> >>>>>>>>>>>>>>>
<aleksey.maksimov(a)it-kb.ru> wrote:
> >>>>>>>>>>>>>>>> Simone,
there is something interesting in the
> >>>>>>>>>>>>>>>> vdsm.log?
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> For what I saw
the issue is not related to the
> >>>>>>>>>>>>>>> storage but to the
network.
> >>>>>>>>>>>>>>>
ovirt-hosted-engine-setup uses the jsonrpc client,
> >>>>>>>>>>>>>>> instead the code
> >>>>>>>>>>>>>>> from
ovirt-hosted-engine-ha still uses the xmlrpc
> >>>>>>>>>>>>>>> client somewhere and
> >>>>>>>>>>>>>>> this happens also
when the setup asks to create the
> >>>>>>>>>>>>>>> lockspace volume.
> >>>>>>>>>>>>>>> It seams that in
your case the xmlrpc client could
> >>>>>>>>>>>>>>> not connect vdsm on
> >>>>>>>>>>>>>>> the localhost.
> >>>>>>>>>>>>>>> It could be
somehow related to:
> >>>>>>>>>>>>>>>
https://bugzilla.redhat.com/1358530
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> Can you please
try executing
> >>>>>>>>>>>>>>> sudo vdsClient
-s 0 getVdsCaps
> >>>>>>>>>>>>>>> on that host?
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> 22.07.2016,
19:36, "aleksey.maksimov(a)it-kb.ru"
> >>>>>>>>>>>>>>>>
<aleksey.maksimov(a)it-kb.ru>:
> >>>>>>>>>>>>>>>>> Simone,
thanks for link.
> >>>>>>>>>>>>>>>>> vdsm.log
attached
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
22.07.2016, 19:28, "Simone Tiraboschi"
> >>>>>>>>>>>>>>>>>
<stirabos(a)redhat.com>:
> >>>>>>>>>>>>>>>>>> On
Fri, Jul 22, 2016 at 5:59 PM,
> >>>>>>>>>>>>>>>>>>
<aleksey.maksimov(a)it-kb.ru> wrote:
> >>>>>>>>>>>>>>>>>>>
Thank you for your response, Simone.
> >>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>
Log attached.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> It
seams it comes from VDSM, can you please
> >>>>>>>>>>>>>>>>>> attach also
vdsm.log?
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>
I don't use ovirt-engine-appliance because I
> >>>>>>>>>>>>>>>>>>> have not
found "how-to" for ovirt-engine-appliance deployment in hosted
> >>>>>>>>>>>>>>>>>>> engine
configuration.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> yum
install ovirt-engine-appliance
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>
Then follow the instruction here:
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>
http://www.ovirt.org/develop/release-management/features/heapplianceflow/
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>
22.07.2016, 17:09, "Simone Tiraboschi"
> >>>>>>>>>>>>>>>>>>>
<stirabos(a)redhat.com>:
> >>>>>>>>>>>>>>>>>>>>
Hi Aleksey,
> >>>>>>>>>>>>>>>>>>>>
Can you please attach hosted-engine-setup
> >>>>>>>>>>>>>>>>>>>>
logs?
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>
On Fri, Jul 22, 2016 at 3:46 PM,
> >>>>>>>>>>>>>>>>>>>>
<aleksey.maksimov(a)it-kb.ru> wrote:
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
Hello oVirt guru`s !
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
I have problem with initial deploy of
> >>>>>>>>>>>>>>>>>>>>>
ovirt 4.0 hosted engine.
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
My environment :
> >>>>>>>>>>>>>>>>>>>>>
============================
> >>>>>>>>>>>>>>>>>>>>>
* Two servers HP ProLiant DL 360 G5 with
> >>>>>>>>>>>>>>>>>>>>>
Qlogic FC HBA connected (with
> >>>>>>>>>>>>>>>>>>>>>
multipathd) to storage HP 3PAR 7200
> >>>>>>>>>>>>>>>>>>>>>
* On each server installed CentOS 7.2
> >>>>>>>>>>>>>>>>>>>>>
Linux (3.10.0-327.22.2.el7.x86_64)
> >>>>>>>>>>>>>>>>>>>>>
* On 3PAR storage I created 2 LUNs for
> >>>>>>>>>>>>>>>>>>>>>
oVirt.
> >>>>>>>>>>>>>>>>>>>>>
- First LUN for oVirt Hosted Engine VM
> >>>>>>>>>>>>>>>>>>>>>
(60GB)
> >>>>>>>>>>>>>>>>>>>>>
- Second LUN for all other VMs (2TB)
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
# multipath -ll
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
3par-vv1
> >>>>>>>>>>>>>>>>>>>>>
(360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV
> >>>>>>>>>>>>>>>>>>>>>
size=60G features='1 queue_if_no_path'
> >>>>>>>>>>>>>>>>>>>>>
hwhandler='1 alua' wp=rw
> >>>>>>>>>>>>>>>>>>>>>
`-+- policy='round-robin 0' prio=50
> >>>>>>>>>>>>>>>>>>>>>
status=active
> >>>>>>>>>>>>>>>>>>>>>
|- 2:0:1:1 sdd 8:48 active ready
> >>>>>>>>>>>>>>>>>>>>>
running
> >>>>>>>>>>>>>>>>>>>>>
|- 3:0:0:1 sdf 8:80 active ready
> >>>>>>>>>>>>>>>>>>>>>
running
> >>>>>>>>>>>>>>>>>>>>>
|- 2:0:0:1 sdb 8:16 active ready
> >>>>>>>>>>>>>>>>>>>>>
running
> >>>>>>>>>>>>>>>>>>>>>
`- 3:0:1:1 sdh 8:112 active ready
> >>>>>>>>>>>>>>>>>>>>>
running
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
3par-vv2
> >>>>>>>>>>>>>>>>>>>>>
(360002ac000000000000000160000cec9) dm-1 3PARdata,VV
> >>>>>>>>>>>>>>>>>>>>>
size=2.0T features='1 queue_if_no_path'
> >>>>>>>>>>>>>>>>>>>>>
hwhandler='1 alua' wp=rw
> >>>>>>>>>>>>>>>>>>>>>
`-+- policy='round-robin 0' prio=50
> >>>>>>>>>>>>>>>>>>>>>
status=active
> >>>>>>>>>>>>>>>>>>>>>
|- 2:0:0:0 sda 8:0 active ready running
> >>>>>>>>>>>>>>>>>>>>>
|- 3:0:0:0 sde 8:64 active ready
> >>>>>>>>>>>>>>>>>>>>>
running
> >>>>>>>>>>>>>>>>>>>>>
|- 2:0:1:0 sdc 8:32 active ready
> >>>>>>>>>>>>>>>>>>>>>
running
> >>>>>>>>>>>>>>>>>>>>>
`- 3:0:1:0 sdg 8:96 active ready
> >>>>>>>>>>>>>>>>>>>>>
running
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
My steps on first server (initial deploy
> >>>>>>>>>>>>>>>>>>>>>
of ovirt 4.0 hosted engine):
> >>>>>>>>>>>>>>>>>>>>>
============================
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
# systemctl stop NetworkManager
> >>>>>>>>>>>>>>>>>>>>>
# systemctl disable NetworkManager
> >>>>>>>>>>>>>>>>>>>>>
# yum -y install
> >>>>>>>>>>>>>>>>>>>>>
http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
> >>>>>>>>>>>>>>>>>>>>>
# yum -y install epel-release
> >>>>>>>>>>>>>>>>>>>>>
# wget
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1...
> >>>>>>>>>>>>>>>>>>>>>
-P /tmp/
> >>>>>>>>>>>>>>>>>>>>>
# yum install ovirt-hosted-engine-setup
> >>>>>>>>>>>>>>>>>>>>>
# yum install screen
> >>>>>>>>>>>>>>>>>>>>>
# screen -RD
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
...in screen session :
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
# hosted-engine --deploy
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
...
> >>>>>>>>>>>>>>>>>>>>>
in configuration process I chose "fc" as
> >>>>>>>>>>>>>>>>>>>>>
storage type for oVirt hosted
> >>>>>>>>>>>>>>>>>>>>>
engine vm and select 60GB LUN...
> >>>>>>>>>>>>>>>>>>>>>
...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
--== CONFIGURATION PREVIEW ==--
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
...
> >>>>>>>>>>>>>>>>>>>>>
Firewall manager : iptables
> >>>>>>>>>>>>>>>>>>>>>
Gateway address : 10.1.0.1
> >>>>>>>>>>>>>>>>>>>>>
Host name for web application :
> >>>>>>>>>>>>>>>>>>>>>
KOM-AD01-OVIRT1
> >>>>>>>>>>>>>>>>>>>>>
Storage Domain type : fc
> >>>>>>>>>>>>>>>>>>>>>
Host ID : 1
> >>>>>>>>>>>>>>>>>>>>>
LUN ID :
> >>>>>>>>>>>>>>>>>>>>>
360002ac0000000000000001b0000cec9
> >>>>>>>>>>>>>>>>>>>>>
Image size GB : 40
> >>>>>>>>>>>>>>>>>>>>>
Console type : vnc
> >>>>>>>>>>>>>>>>>>>>>
Memory size MB : 4096
> >>>>>>>>>>>>>>>>>>>>>
MAC address : 00:16:3e:77:1d:07
> >>>>>>>>>>>>>>>>>>>>>
Boot type : cdrom
> >>>>>>>>>>>>>>>>>>>>>
Number of CPUs : 2
> >>>>>>>>>>>>>>>>>>>>>
ISO image (cdrom
> >>>>>>>>>>>>>>>>>>>>>
boot/cloud-init) :
> >>>>>>>>>>>>>>>>>>>>>
/tmp/CentOS-7-x86_64-NetInstall-1511.iso
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>
Can I ask why you prefer/need to manually
> >>>>>>>>>>>>>>>>>>>>
create a VM installing from
> >>>>>>>>>>>>>>>>>>>>
a CD instead of using the ready-to-use
> >>>>>>>>>>>>>>>>>>>>
ovirt-engine-appliance?
> >>>>>>>>>>>>>>>>>>>>
Using the appliance makes the setup process
> >>>>>>>>>>>>>>>>>>>> a
lot shorted and more comfortable.
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
CPU Type : model_Penryn
> >>>>>>>>>>>>>>>>>>>>>
...
> >>>>>>>>>>>>>>>>>>>>>
and get error after step "Verifying
> >>>>>>>>>>>>>>>>>>>>>
sanlock lockspace initialization"
> >>>>>>>>>>>>>>>>>>>>>
...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Verifying sanlock lockspace
> >>>>>>>>>>>>>>>>>>>>>
initialization
> >>>>>>>>>>>>>>>>>>>>>
[ ERROR ] Failed to execute stage 'Misc
> >>>>>>>>>>>>>>>>>>>>>
configuration': [Errno 101] Network
> >>>>>>>>>>>>>>>>>>>>>
is unreachable
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Clean up
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Generating answer file
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Pre-termination
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Termination
> >>>>>>>>>>>>>>>>>>>>>
[ ERROR ] Hosted Engine deployment
> >>>>>>>>>>>>>>>>>>>>>
failed: this system is not reliable,
> >>>>>>>>>>>>>>>>>>>>>
please check the issue, fix and redeploy
> >>>>>>>>>>>>>>>>>>>>>
Log file is located at
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
Interestingly
> >>>>>>>>>>>>>>>>>>>>>
============================
> >>>>>>>>>>>>>>>>>>>>>
If I try to deploy hosted-engine v3.6,
> >>>>>>>>>>>>>>>>>>>>>
everything goes well in the same
> >>>>>>>>>>>>>>>>>>>>>
configuration !! :
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
....
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Transaction setup
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Misc configuration
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Package installation
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Misc configuration
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Configuring libvirt
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Configuring VDSM
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Starting vdsmd
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Waiting for VDSM hardware info
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Configuring the management
> >>>>>>>>>>>>>>>>>>>>>
bridge
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Volume Group
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Storage Domain
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Storage Pool
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Connecting Storage Pool
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Verifying sanlock lockspace
> >>>>>>>>>>>>>>>>>>>>>
initialization
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Image for
> >>>>>>>>>>>>>>>>>>>>>
'hosted-engine.lockspace' ...
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Image for
> >>>>>>>>>>>>>>>>>>>>>
'hosted-engine.lockspace' created successfully
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating Image for
> >>>>>>>>>>>>>>>>>>>>>
'hosted-engine.metadata' ...
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Image for
> >>>>>>>>>>>>>>>>>>>>>
'hosted-engine.metadata' created successfully
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating VM Image
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Destroying Storage Pool
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Start monitoring domain
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Configuring VM
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Updating hosted-engine
> >>>>>>>>>>>>>>>>>>>>>
configuration
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Transaction commit
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Stage: Closing up
> >>>>>>>>>>>>>>>>>>>>>
[ INFO ] Creating VM
> >>>>>>>>>>>>>>>>>>>>>
You can now connect to the VM
> >>>>>>>>>>>>>>>>>>>>>
with the following command:
> >>>>>>>>>>>>>>>>>>>>>
/bin/remote-viewer
> >>>>>>>>>>>>>>>>>>>>>
vnc://localhost:5900
> >>>>>>>>>>>>>>>>>>>>>
...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
What could be the problem?
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
_______________________________________________
> >>>>>>>>>>>>>>>>>>>>>
Users mailing list
> >>>>>>>>>>>>>>>>>>>>>
Users(a)ovirt.org
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>>>>>>>
_______________________________________________
> >>>>>>>>>>>> Users mailing list
> >>>>>>>>>>>> Users(a)ovirt.org
> >>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>>>>
> >>>>>>>>>
_______________________________________________
> >>>>>>>>> Users mailing list
> >>>>>>>>> Users(a)ovirt.org
> >>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>
> >>>>>> _______________________________________________
> >>>>>> Users mailing list
> >>>>>> Users(a)ovirt.org
> >>>>>>
http://lists.ovirt.org/mailman/listinfo/users
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users
>
>
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users
>