[ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

Edward Haas ehaas at redhat.com
Tue Jul 26 02:13:05 EDT 2016


On Mon, Jul 25, 2016 at 4:02 PM, Simone Tiraboschi <stirabos at redhat.com>
wrote:

> On Mon, Jul 25, 2016 at 2:38 PM,  <aleksey.maksimov at it-kb.ru> wrote:
> > telnet kom-ad01-vm31.holding.com 54321 = success connection
> >
> > telnet ::1 54321
> > Trying ::1...
> > telnet: connect to address ::1: Network is unreachable
> >
> > (ipv6 on my server disabled)
>
> Ok, so the issue seams here: now by default vdsm binds on :: and its
> heuristc can end up using ipv6.
> See this one: https://bugzilla.redhat.com/show_bug.cgi?id=1350883
>
> Can you please try enabling ipv6 on your host or setting
>  management_ip = 0.0.0.0
> under the [address] section in /etc/vdsm/vdsm.conf
> and then restarting vdsm.
>
>
Could you please also add the 'ip addr' output? Just interested to see how
IPv6 was
disabled on the host.
It will be even better if you could apply the patch (
https://gerrit.ovirt.org/#/c/60020) and check.


>
>
> > 25.07.2016, 15:35, "Simone Tiraboschi" <stirabos at redhat.com>:
> >> On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksimov at it-kb.ru> wrote:
> >>>  # ss -plutn
> >>>
> >>>  Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port
> >>>
> >>>  udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6))
> >>>  udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6))
> >>>  udp UNCONN 0 0 127.0.0.1:323 *:* users:(("chronyd",pid=795,fd=1))
> >>>  udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7))
> >>>  udp UNCONN 0 0 127.0.0.1:25375 *:* users:(("snmpd",pid=1609,fd=8))
> >>>  udp UNCONN 0 0 127.0.0.1:25376 *:* users:(("cmapeerd",pid=2056,fd=5))
> >>>  udp UNCONN 0 0 127.0.0.1:25393 *:* users:(("cmanicd",pid=2278,fd=3))
> >>>  udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9))
> >>>  udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10))
> >>>  tcp LISTEN 0 128 *:2381 *:*
> users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4))
> >>>  tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8))
> >>>  tcp LISTEN 0 5 *:54322 *:* users:(("ovirt-imageio-d",pid=753,fd=3))
> >>>  tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3))
> >>>  tcp LISTEN 0 100 127.0.0.1:25 *:* users:(("master",pid=1948,fd=13))
> >>>  tcp LISTEN 0 128 *:2301 *:*
> users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
> >>>  tcp LISTEN 0 30 *:16514 *:* users:(("libvirtd",pid=10688,fd=13))
> >>>  tcp LISTEN 0 128 127.0.0.1:199 *:* users:(("snmpd",pid=1609,fd=9))
> >>>  tcp LISTEN 0 128 :::111 :::* users:(("rpcbind",pid=827,fd=11))
> >>>  tcp LISTEN 0 5 :::54321 :::* users:(("vdsm",pid=11077,fd=23))
> >>
> >> vdsm is properly bind over ipv6.
> >>
> >> Can you please check if you can connect to vdsm with:
> >>  telnet kom-ad01-vm31.holding.com 54321
> >> and with
> >>  telnet ::1 54321
> >> ?
> >>
> >>>  tcp LISTEN 0 30 :::16514 :::* users:(("libvirtd",pid=10688,fd=14))
> >>>
> >>>  25.07.2016, 15:11, "Simone Tiraboschi" <stirabos at redhat.com>:
> >>>>  On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov at it-kb.ru> wrote:
> >>>>>   Yes.
> >>>>>
> >>>>>   # ping $(python -c 'import socket; print(socket.gethostname())')
> >>>>>
> >>>>>   PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
> >>>>>   64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1
> ttl=64 time=0.030 ms
> >>>>>   64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2
> ttl=64 time=0.015 ms
> >>>>>   64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3
> ttl=64 time=0.011 ms
> >>>>>   ^C
> >>>>>   --- KOM-AD01-VM31.holding.com ping statistics ---
> >>>>>   3 packets transmitted, 3 received, 0% packet loss, time 2001ms
> >>>>>   rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
> >>>>>
> >>>>>   but...
> >>>>
> >>>>  and the output of
> >>>>   ss -plutn
> >>>>
> >>>>>   # vdsClient -s 0 getVdsCaps
> >>>>>
> >>>>>   Traceback (most recent call last):
> >>>>>     File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
> >>>>>       code, message = commands[command][0](commandArgs)
> >>>>>     File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
> >>>>>       return self.ExecAndExit(self.s.getVdsCapabilities())
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
> >>>>>       return self.__send(self.__name, args)
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
> >>>>>       verbose=self.__verbose
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
> >>>>>       return self.single_request(host, handler, request_body,
> verbose)
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in
> single_request
> >>>>>       self.send_content(h, request_body)
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in
> send_content
> >>>>>       connection.endheaders(request_body)
> >>>>>     File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
> >>>>>       self._send_output(message_body)
> >>>>>     File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
> >>>>>       self.send(msg)
> >>>>>     File "/usr/lib64/python2.7/httplib.py", line 797, in send
> >>>>>       self.connect()
> >>>>>     File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line
> 203, in connect
> >>>>>       sock = socket.create_connection((self.host, self.port),
> self.timeout)
> >>>>>     File "/usr/lib64/python2.7/socket.py", line 571, in
> create_connection
> >>>>>       raise err
> >>>>>   error: [Errno 101] Network is unreachable
> >>>>>
> >>>>>   25.07.2016, 14:58, "aleksey.maksimov at it-kb.ru" <
> aleksey.maksimov at it-kb.ru>:
> >>>>>>   Ok.
> >>>>>>
> >>>>>>   1) I stopped and disabled the service NetworkManager
> >>>>>>   # systemctl stop NetworkManager
> >>>>>>   # systemctl disable NetworkManager
> >>>>>>
> >>>>>>   2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add
> PEERDNS=no in ifcfg-* file.
> >>>>>>
> >>>>>>   3) Reboot server
> >>>>>>
> >>>>>>   4) Try deploy oVirt HE 4 and I get the same error
> >>>>>>
> >>>>>>   [ INFO ] Creating Volume Group
> >>>>>>   [ INFO ] Creating Storage Domain
> >>>>>>   [ INFO ] Creating Storage Pool
> >>>>>>   [ INFO ] Connecting Storage Pool
> >>>>>>   [ INFO ] Verifying sanlock lockspace initialization
> >>>>>>   [ ERROR ] Failed to execute stage 'Misc configuration': [Errno
> 101] Network is unreachable
> >>>>>>   [ INFO ] Stage: Clean up
> >>>>>>   [ INFO ] Generating answer file
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
> >>>>>>   [ INFO ] Stage: Pre-termination
> >>>>>>   [ INFO ] Stage: Termination
> >>>>>>   [ ERROR ] Hosted Engine deployment failed: this system is not
> reliable, please check the issue, fix and redeploy
> >>>>>>             Log file is located at
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
> >>>>>>
> >>>>>>   What ideas further?
> >>>>>>
> >>>>>>   25.07.2016, 13:06, "Simone Tiraboschi" <stirabos at redhat.com>:
> >>>>>>>    On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksimov at it-kb.ru>
> wrote:
> >>>>>>>>     What am I supposed to do for successfully deploy ovirt 4 ?
> >>>>>>>>     Any ideas ?
> >>>>>>>
> >>>>>>>    Can you please try to explicitly configure your DNS with
> nameserver
> >>>>>>>    under /etc/resolv.conf and remove DNS1 and DNS2 and set
> PEERDNS=no for
> >>>>>>>    the interface you are going to use?
> >>>>>>>
> >>>>>>>>     25.07.2016, 12:47, "aleksey.maksimov at it-kb.ru" <
> aleksey.maksimov at it-kb.ru>:
> >>>>>>>>>     "Aleksey, was your DNS configured with DNS1 and DNS2 just on
> the interface you used to create the management bridge on?"
> >>>>>>>>>
> >>>>>>>>>     Yes. Of course
> >>>>>>>>>
> >>>>>>>>>     25.07.2016, 12:27, "Simone Tiraboschi" <stirabos at redhat.com
> >:
> >>>>>>>>>>      On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
> >>>>>>>>>>      <piotr.kliczewski at gmail.com> wrote:
> >>>>>>>>>>>       This could be the issue here as well as for BZ #1358530
> >>>>>>>>>>>
> >>>>>>>>>>>       On Mon, Jul 25, 2016 at 10:53 AM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>       Could this be due to the fact that the ovirt installer
> has changed network configuration files (ifcfg-*, resolv.conf) ?
> >>>>>>>>>>>>       After the error in ovirt installation process I see
> from resolv.conf disappeared on my DNS servers entry and now the server is
> unable to resolve names.
> >>>>>>>>>>
> >>>>>>>>>>      So it could be related to
> https://bugzilla.redhat.com/show_bug.cgi?id=1160423
> >>>>>>>>>>
> >>>>>>>>>>      Aleksey, was your DNS configured with DNS1 and DNS2 just
> on the
> >>>>>>>>>>      interface you used to create the management bridge on?
> >>>>>>>>>>      Can you please try the workaround described here
> >>>>>>>>>>      https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
> >>>>>>>>>>
> >>>>>>>>>>>>       25.07.2016, 11:26, "Simone Tiraboschi" <
> stirabos at redhat.com>:
> >>>>>>>>>>>>>       On Mon, Jul 25, 2016 at 10:22 AM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>>>        # vdsClient -s 0 getVdsCaps
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>        Traceback (most recent call last):
> >>>>>>>>>>>>>>          File "/usr/share/vdsm/vdsClient.py", line 2980, in
> <module>
> >>>>>>>>>>>>>>            code, message = commands[command][0](commandArgs)
> >>>>>>>>>>>>>>          File "/usr/share/vdsm/vdsClient.py", line 543, in
> do_getCap
> >>>>>>>>>>>>>>            return
> self.ExecAndExit(self.s.getVdsCapabilities())
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1233, in __call__
> >>>>>>>>>>>>>>            return self.__send(self.__name, args)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1587, in __request
> >>>>>>>>>>>>>>            verbose=self.__verbose
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1273, in request
> >>>>>>>>>>>>>>            return self.single_request(host, handler,
> request_body, verbose)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1301, in single_request
> >>>>>>>>>>>>>>            self.send_content(h, request_body)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1448, in send_content
> >>>>>>>>>>>>>>            connection.endheaders(request_body)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/httplib.py", line 975,
> in endheaders
> >>>>>>>>>>>>>>            self._send_output(message_body)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/httplib.py", line 835,
> in _send_output
> >>>>>>>>>>>>>>            self.send(msg)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/httplib.py", line 797,
> in send
> >>>>>>>>>>>>>>            self.connect()
> >>>>>>>>>>>>>>          File
> "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect
> >>>>>>>>>>>>>>            sock = socket.create_connection((self.host,
> self.port), self.timeout)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/socket.py", line 571,
> in create_connection
> >>>>>>>>>>>>>>            raise err
> >>>>>>>>>>>>>>        error: [Errno 101] Network is unreachable
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>       Yaniv, can you please take also a look to this one?
> >>>>>>>>>>>>>       it's exactly the opposite of
> https://bugzilla.redhat.com/1358530
> >>>>>>>>>>>>>       Here the jsonrpcclient works but not the xmlrpc one.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>>        25.07.2016, 11:17, "Simone Tiraboschi" <
> stirabos at redhat.com>:
> >>>>>>>>>>>>>>>        On Mon, Jul 25, 2016 at 7:51 AM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>>>>>         Simone, there is something interesting in the
> vdsm.log?
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>        For what I saw the issue is not related to the
> storage but to the network.
> >>>>>>>>>>>>>>>        ovirt-hosted-engine-setup uses the jsonrpc client,
> instead the code
> >>>>>>>>>>>>>>>        from ovirt-hosted-engine-ha still uses the xmlrpc
> client somewhere and
> >>>>>>>>>>>>>>>        this happens also when the setup asks to create the
> lockspace volume.
> >>>>>>>>>>>>>>>        It seams that in your case the xmlrpc client could
> not connect vdsm on
> >>>>>>>>>>>>>>>        the localhost.
> >>>>>>>>>>>>>>>        It could be somehow related to:
> >>>>>>>>>>>>>>>        https://bugzilla.redhat.com/1358530
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>        Can you please try executing
> >>>>>>>>>>>>>>>         sudo vdsClient -s 0 getVdsCaps
> >>>>>>>>>>>>>>>        on that host?
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>         22.07.2016, 19:36, "aleksey.maksimov at it-kb.ru" <
> aleksey.maksimov at it-kb.ru>:
> >>>>>>>>>>>>>>>>>         Simone, thanks for link.
> >>>>>>>>>>>>>>>>>         vdsm.log attached
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>         22.07.2016, 19:28, "Simone Tiraboschi" <
> stirabos at redhat.com>:
> >>>>>>>>>>>>>>>>>>          On Fri, Jul 22, 2016 at 5:59 PM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>>>>>>>>           Thank you for your response, Simone.
> >>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>           Log attached.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>          It seams it comes from VDSM, can you please
> attach also vdsm.log?
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>           I don't use ovirt-engine-appliance because I
> have not found "how-to" for ovirt-engine-appliance deployment in hosted
> engine configuration.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>          yum install ovirt-engine-appliance
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>          Then follow the instruction here:
> >>>>>>>>>>>>>>>>>>
> http://www.ovirt.org/develop/release-management/features/heapplianceflow/
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>           22.07.2016, 17:09, "Simone Tiraboschi" <
> stirabos at redhat.com>:
> >>>>>>>>>>>>>>>>>>>>           Hi Aleksey,
> >>>>>>>>>>>>>>>>>>>>           Can you please attach hosted-engine-setup
> logs?
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>           On Fri, Jul 22, 2016 at 3:46 PM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            Hello oVirt guru`s !
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            I have problem with initial deploy of
> ovirt 4.0 hosted engine.
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            My environment :
> >>>>>>>>>>>>>>>>>>>>>            ============================
> >>>>>>>>>>>>>>>>>>>>>            * Two servers HP ProLiant DL 360 G5 with
> Qlogic FC HBA connected (with
> >>>>>>>>>>>>>>>>>>>>>            multipathd) to storage HP 3PAR 7200
> >>>>>>>>>>>>>>>>>>>>>            * On each server installed CentOS 7.2
> Linux (3.10.0-327.22.2.el7.x86_64)
> >>>>>>>>>>>>>>>>>>>>>            * On 3PAR storage I created 2 LUNs for
> oVirt.
> >>>>>>>>>>>>>>>>>>>>>            - First LUN for oVirt Hosted Engine VM
> (60GB)
> >>>>>>>>>>>>>>>>>>>>>            - Second LUN for all other VMs (2TB)
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            # multipath -ll
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            3par-vv1
> (360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV
> >>>>>>>>>>>>>>>>>>>>>            size=60G features='1 queue_if_no_path'
> hwhandler='1 alua' wp=rw
> >>>>>>>>>>>>>>>>>>>>>            `-+- policy='round-robin 0' prio=50
> status=active
> >>>>>>>>>>>>>>>>>>>>>              |- 2:0:1:1 sdd 8:48 active ready running
> >>>>>>>>>>>>>>>>>>>>>              |- 3:0:0:1 sdf 8:80 active ready running
> >>>>>>>>>>>>>>>>>>>>>              |- 2:0:0:1 sdb 8:16 active ready running
> >>>>>>>>>>>>>>>>>>>>>              `- 3:0:1:1 sdh 8:112 active ready
> running
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            3par-vv2
> (360002ac000000000000000160000cec9) dm-1 3PARdata,VV
> >>>>>>>>>>>>>>>>>>>>>            size=2.0T features='1 queue_if_no_path'
> hwhandler='1 alua' wp=rw
> >>>>>>>>>>>>>>>>>>>>>            `-+- policy='round-robin 0' prio=50
> status=active
> >>>>>>>>>>>>>>>>>>>>>              |- 2:0:0:0 sda 8:0 active ready running
> >>>>>>>>>>>>>>>>>>>>>              |- 3:0:0:0 sde 8:64 active ready running
> >>>>>>>>>>>>>>>>>>>>>              |- 2:0:1:0 sdc 8:32 active ready running
> >>>>>>>>>>>>>>>>>>>>>              `- 3:0:1:0 sdg 8:96 active ready running
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            My steps on first server (initial deploy
> of ovirt 4.0 hosted engine):
> >>>>>>>>>>>>>>>>>>>>>            ============================
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            # systemctl stop NetworkManager
> >>>>>>>>>>>>>>>>>>>>>            # systemctl disable NetworkManager
> >>>>>>>>>>>>>>>>>>>>>            # yum -y install
> http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
> >>>>>>>>>>>>>>>>>>>>>            # yum -y install epel-release
> >>>>>>>>>>>>>>>>>>>>>            # wget
> >>>>>>>>>>>>>>>>>>>>>
> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso
> >>>>>>>>>>>>>>>>>>>>>            -P /tmp/
> >>>>>>>>>>>>>>>>>>>>>            # yum install ovirt-hosted-engine-setup
> >>>>>>>>>>>>>>>>>>>>>            # yum install screen
> >>>>>>>>>>>>>>>>>>>>>            # screen -RD
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            ...in screen session :
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            # hosted-engine --deploy
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>            in configuration process I chose "fc" as
> storage type for oVirt hosted
> >>>>>>>>>>>>>>>>>>>>>            engine vm and select 60GB LUN...
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            --== CONFIGURATION PREVIEW ==--
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>                      Firewall manager : iptables
> >>>>>>>>>>>>>>>>>>>>>                      Gateway address : 10.1.0.1
> >>>>>>>>>>>>>>>>>>>>>                      Host name for web application :
> KOM-AD01-OVIRT1
> >>>>>>>>>>>>>>>>>>>>>                      Storage Domain type : fc
> >>>>>>>>>>>>>>>>>>>>>                      Host ID : 1
> >>>>>>>>>>>>>>>>>>>>>                      LUN ID :
> >>>>>>>>>>>>>>>>>>>>>            360002ac0000000000000001b0000cec9
> >>>>>>>>>>>>>>>>>>>>>                      Image size GB : 40
> >>>>>>>>>>>>>>>>>>>>>                      Console type : vnc
> >>>>>>>>>>>>>>>>>>>>>                      Memory size MB : 4096
> >>>>>>>>>>>>>>>>>>>>>                      MAC address : 00:16:3e:77:1d:07
> >>>>>>>>>>>>>>>>>>>>>                      Boot type : cdrom
> >>>>>>>>>>>>>>>>>>>>>                      Number of CPUs : 2
> >>>>>>>>>>>>>>>>>>>>>                      ISO image (cdrom
> boot/cloud-init) :
> >>>>>>>>>>>>>>>>>>>>>            /tmp/CentOS-7-x86_64-NetInstall-1511.iso
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>           Can I ask why you prefer/need to manually
> create a VM installing from
> >>>>>>>>>>>>>>>>>>>>           a CD instead of using the ready-to-use
> ovirt-engine-appliance?
> >>>>>>>>>>>>>>>>>>>>           Using the appliance makes the setup process
> a lot shorted and more comfortable.
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>                      CPU Type : model_Penryn
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>            and get error after step "Verifying
> sanlock lockspace initialization"
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Verifying sanlock lockspace
> initialization
> >>>>>>>>>>>>>>>>>>>>>            [ ERROR ] Failed to execute stage 'Misc
> configuration': [Errno 101] Network
> >>>>>>>>>>>>>>>>>>>>>            is unreachable
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Clean up
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Generating answer file
> >>>>>>>>>>>>>>>>>>>>>
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Pre-termination
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Termination
> >>>>>>>>>>>>>>>>>>>>>            [ ERROR ] Hosted Engine deployment
> failed: this system is not reliable,
> >>>>>>>>>>>>>>>>>>>>>            please check the issue, fix and redeploy
> >>>>>>>>>>>>>>>>>>>>>                      Log file is located at
> >>>>>>>>>>>>>>>>>>>>>
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            Interestingly
> >>>>>>>>>>>>>>>>>>>>>            ============================
> >>>>>>>>>>>>>>>>>>>>>            If I try to deploy hosted-engine v3.6,
> everything goes well in the same
> >>>>>>>>>>>>>>>>>>>>>            configuration !! :
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            ....
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Transaction setup
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Misc configuration
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Package installation
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Misc configuration
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Configuring libvirt
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Configuring VDSM
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Starting vdsmd
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Waiting for VDSM hardware info
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Configuring the management bridge
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Volume Group
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Storage Domain
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Storage Pool
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Connecting Storage Pool
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Verifying sanlock lockspace
> initialization
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Image for
> 'hosted-engine.lockspace' ...
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Image for
> 'hosted-engine.lockspace' created successfully
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Image for
> 'hosted-engine.metadata' ...
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Image for
> 'hosted-engine.metadata' created successfully
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating VM Image
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Destroying Storage Pool
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Start monitoring domain
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Configuring VM
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Updating hosted-engine
> configuration
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Transaction commit
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Closing up
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating VM
> >>>>>>>>>>>>>>>>>>>>>                      You can now connect to the VM
> with the following command:
> >>>>>>>>>>>>>>>>>>>>>                            /bin/remote-viewer
> vnc://localhost:5900
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            What could be the problem?
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
> _______________________________________________
> >>>>>>>>>>>>>>>>>>>>>            Users mailing list
> >>>>>>>>>>>>>>>>>>>>>            Users at ovirt.org
> >>>>>>>>>>>>>>>>>>>>>
> http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>>>>>>>       _______________________________________________
> >>>>>>>>>>>>       Users mailing list
> >>>>>>>>>>>>       Users at ovirt.org
> >>>>>>>>>>>>       http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>>>>
> >>>>>>>>>     _______________________________________________
> >>>>>>>>>     Users mailing list
> >>>>>>>>>     Users at ovirt.org
> >>>>>>>>>     http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>
> >>>>>>   _______________________________________________
> >>>>>>   Users mailing list
> >>>>>>   Users at ovirt.org
> >>>>>>   http://lists.ovirt.org/mailman/listinfo/users
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20160726/8d2dbe38/attachment-0001.html>


More information about the Users mailing list