[ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

Roy Golan rgolan at redhat.com
Wed Jul 27 06:12:44 UTC 2016


Edi, danken, this is again the malfunctioning heuristics in the client for
handling ipv6. Is there a bug on this issue ?

On Jul 27, 2016 8:57 AM, <aleksey.maksimov at it-kb.ru> wrote:

> I enabled ipv6 for "lo" and "ovirtmgmt" interfaces and deployment process
> overt completed successfully.
>
> # cat /etc/sysctl.conf
>
> net.ipv6.conf.all.disable_ipv6 = 1
> net.ipv6.conf.default.disable_ipv6 = 1
> net.ipv6.conf.lo.disable_ipv6 = 0
> net.ipv6.conf.ovirtmgmt.disable_ipv6 = 0
>
> Thank you for your Help!
>
> 26.07.2016, 09:13, "Edward Haas" <ehaas at redhat.com>:
>
>
>
> On Mon, Jul 25, 2016 at 4:02 PM, Simone Tiraboschi <stirabos at redhat.com>
> wrote:
>
> On Mon, Jul 25, 2016 at 2:38 PM,  <aleksey.maksimov at it-kb.ru> wrote:
> > telnet kom-ad01-vm31.holding.com 54321 = success connection
> >
> > telnet ::1 54321
> > Trying ::1...
> > telnet: connect to address ::1: Network is unreachable
> >
> > (ipv6 on my server disabled)
>
> Ok, so the issue seams here: now by default vdsm binds on :: and its
> heuristc can end up using ipv6.
> See this one: https://bugzilla.redhat.com/show_bug.cgi?id=1350883
>
> Can you please try enabling ipv6 on your host or setting
>  management_ip = 0.0.0.0
> under the [address] section in /etc/vdsm/vdsm.conf
> and then restarting vdsm.
>
>
>
> Could you please also add the 'ip addr' output? Just interested to see how
> IPv6 was
> disabled on the host.
> It will be even better if you could apply the patch (
> https://gerrit.ovirt.org/#/c/60020) and check.
>
>
>
>
>
> > 25.07.2016, 15:35, "Simone Tiraboschi" <stirabos at redhat.com>:
> >> On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksimov at it-kb.ru> wrote:
> >>>  # ss -plutn
> >>>
> >>>  Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port
> >>>
> >>>  udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6))
> >>>  udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6))
> >>>  udp UNCONN 0 0 127.0.0.1:323 *:* users:(("chronyd",pid=795,fd=1))
> >>>  udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7))
> >>>  udp UNCONN 0 0 127.0.0.1:25375 *:* users:(("snmpd",pid=1609,fd=8))
> >>>  udp UNCONN 0 0 127.0.0.1:25376 *:* users:(("cmapeerd",pid=2056,fd=5))
> >>>  udp UNCONN 0 0 127.0.0.1:25393 *:* users:(("cmanicd",pid=2278,fd=3))
> >>>  udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9))
> >>>  udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10))
> >>>  tcp LISTEN 0 128 *:2381 *:*
> users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4))
> >>>  tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8))
> >>>  tcp LISTEN 0 5 *:54322 *:* users:(("ovirt-imageio-d",pid=753,fd=3))
> >>>  tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3))
> >>>  tcp LISTEN 0 100 127.0.0.1:25 *:* users:(("master",pid=1948,fd=13))
> >>>  tcp LISTEN 0 128 *:2301 *:*
> users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
> >>>  tcp LISTEN 0 30 *:16514 *:* users:(("libvirtd",pid=10688,fd=13))
> >>>  tcp LISTEN 0 128 127.0.0.1:199 *:* users:(("snmpd",pid=1609,fd=9))
> >>>  tcp LISTEN 0 128 :::111 :::* users:(("rpcbind",pid=827,fd=11))
> >>>  tcp LISTEN 0 5 :::54321 :::* users:(("vdsm",pid=11077,fd=23))
> >>
> >> vdsm is properly bind over ipv6.
> >>
> >> Can you please check if you can connect to vdsm with:
> >>  telnet kom-ad01-vm31.holding.com 54321
> >> and with
> >>  telnet ::1 54321
> >> ?
> >>
> >>>  tcp LISTEN 0 30 :::16514 :::* users:(("libvirtd",pid=10688,fd=14))
> >>>
> >>>  25.07.2016, 15:11, "Simone Tiraboschi" <stirabos at redhat.com>:
> >>>>  On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov at it-kb.ru> wrote:
> >>>>>   Yes.
> >>>>>
> >>>>>   # ping $(python -c 'import socket; print(socket.gethostname())')
> >>>>>
> >>>>>   PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
> >>>>>   64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1
> ttl=64 time=0.030 ms
> >>>>>   64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2
> ttl=64 time=0.015 ms
> >>>>>   64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3
> ttl=64 time=0.011 ms
> >>>>>   ^C
> >>>>>   --- KOM-AD01-VM31.holding.com ping statistics ---
> >>>>>   3 packets transmitted, 3 received, 0% packet loss, time 2001ms
> >>>>>   rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
> >>>>>
> >>>>>   but...
> >>>>
> >>>>  and the output of
> >>>>   ss -plutn
> >>>>
> >>>>>   # vdsClient -s 0 getVdsCaps
> >>>>>
> >>>>>   Traceback (most recent call last):
> >>>>>     File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
> >>>>>       code, message = commands[command][0](commandArgs)
> >>>>>     File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
> >>>>>       return self.ExecAndExit(self.s.getVdsCapabilities())
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
> >>>>>       return self.__send(self.__name, args)
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
> >>>>>       verbose=self.__verbose
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
> >>>>>       return self.single_request(host, handler, request_body,
> verbose)
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in
> single_request
> >>>>>       self.send_content(h, request_body)
> >>>>>     File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in
> send_content
> >>>>>       connection.endheaders(request_body)
> >>>>>     File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
> >>>>>       self._send_output(message_body)
> >>>>>     File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
> >>>>>       self.send(msg)
> >>>>>     File "/usr/lib64/python2.7/httplib.py", line 797, in send
> >>>>>       self.connect()
> >>>>>     File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line
> 203, in connect
> >>>>>       sock = socket.create_connection((self.host, self.port),
> self.timeout)
> >>>>>     File "/usr/lib64/python2.7/socket.py", line 571, in
> create_connection
> >>>>>       raise err
> >>>>>   error: [Errno 101] Network is unreachable
> >>>>>
> >>>>>   25.07.2016, 14:58, "aleksey.maksimov at it-kb.ru" <
> aleksey.maksimov at it-kb.ru>:
> >>>>>>   Ok.
> >>>>>>
> >>>>>>   1) I stopped and disabled the service NetworkManager
> >>>>>>   # systemctl stop NetworkManager
> >>>>>>   # systemctl disable NetworkManager
> >>>>>>
> >>>>>>   2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add
> PEERDNS=no in ifcfg-* file.
> >>>>>>
> >>>>>>   3) Reboot server
> >>>>>>
> >>>>>>   4) Try deploy oVirt HE 4 and I get the same error
> >>>>>>
> >>>>>>   [ INFO ] Creating Volume Group
> >>>>>>   [ INFO ] Creating Storage Domain
> >>>>>>   [ INFO ] Creating Storage Pool
> >>>>>>   [ INFO ] Connecting Storage Pool
> >>>>>>   [ INFO ] Verifying sanlock lockspace initialization
> >>>>>>   [ ERROR ] Failed to execute stage 'Misc configuration': [Errno
> 101] Network is unreachable
> >>>>>>   [ INFO ] Stage: Clean up
> >>>>>>   [ INFO ] Generating answer file
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
> >>>>>>   [ INFO ] Stage: Pre-termination
> >>>>>>   [ INFO ] Stage: Termination
> >>>>>>   [ ERROR ] Hosted Engine deployment failed: this system is not
> reliable, please check the issue, fix and redeploy
> >>>>>>             Log file is located at
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-
> 20160725142534-t81kwf.log
> >>>>>>
> >>>>>>   What ideas further?
> >>>>>>
> >>>>>>   25.07.2016, 13:06, "Simone Tiraboschi" <stirabos at redhat.com>:
> >>>>>>>    On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksimov at it-kb.ru>
> wrote:
> >>>>>>>>     What am I supposed to do for successfully deploy ovirt 4 ?
> >>>>>>>>     Any ideas ?
> >>>>>>>
> >>>>>>>    Can you please try to explicitly configure your DNS with
> nameserver
> >>>>>>>    under /etc/resolv.conf and remove DNS1 and DNS2 and set
> PEERDNS=no for
> >>>>>>>    the interface you are going to use?
> >>>>>>>
> >>>>>>>>     25.07.2016, 12:47, "aleksey.maksimov at it-kb.ru" <
> aleksey.maksimov at it-kb.ru>:
> >>>>>>>>>     "Aleksey, was your DNS configured with DNS1 and DNS2 just on
> the interface you used to create the management bridge on?"
> >>>>>>>>>
> >>>>>>>>>     Yes. Of course
> >>>>>>>>>
> >>>>>>>>>     25.07.2016, 12:27, "Simone Tiraboschi" <stirabos at redhat.com
> >:
> >>>>>>>>>>      On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
> >>>>>>>>>>      <piotr.kliczewski at gmail.com> wrote:
> >>>>>>>>>>>       This could be the issue here as well as for BZ #1358530
> >>>>>>>>>>>
> >>>>>>>>>>>       On Mon, Jul 25, 2016 at 10:53 AM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>       Could this be due to the fact that the ovirt installer
> has changed network configuration files (ifcfg-*, resolv.conf) ?
> >>>>>>>>>>>>       After the error in ovirt installation process I see
> from resolv.conf disappeared on my DNS servers entry and now the server is
> unable to resolve names.
> >>>>>>>>>>
> >>>>>>>>>>      So it could be related to
> https://bugzilla.redhat.com/show_bug.cgi?id=1160423
> >>>>>>>>>>
> >>>>>>>>>>      Aleksey, was your DNS configured with DNS1 and DNS2 just
> on the
> >>>>>>>>>>      interface you used to create the management bridge on?
> >>>>>>>>>>      Can you please try the workaround described here
> >>>>>>>>>>      https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
> >>>>>>>>>>
> >>>>>>>>>>>>       25.07.2016, 11:26, "Simone Tiraboschi" <
> stirabos at redhat.com>:
> >>>>>>>>>>>>>       On Mon, Jul 25, 2016 at 10:22 AM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>>>        # vdsClient -s 0 getVdsCaps
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>        Traceback (most recent call last):
> >>>>>>>>>>>>>>          File "/usr/share/vdsm/vdsClient.py", line 2980, in
> <module>
> >>>>>>>>>>>>>>            code, message = commands[command][0](commandArgs)
> >>>>>>>>>>>>>>          File "/usr/share/vdsm/vdsClient.py", line 543, in
> do_getCap
> >>>>>>>>>>>>>>            return
> self.ExecAndExit(self.s.getVdsCapabilities())
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1233, in __call__
> >>>>>>>>>>>>>>            return self.__send(self.__name, args)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1587, in __request
> >>>>>>>>>>>>>>            verbose=self.__verbose
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1273, in request
> >>>>>>>>>>>>>>            return self.single_request(host, handler,
> request_body, verbose)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1301, in single_request
> >>>>>>>>>>>>>>            self.send_content(h, request_body)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/xmlrpclib.py", line
> 1448, in send_content
> >>>>>>>>>>>>>>            connection.endheaders(request_body)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/httplib.py", line 975,
> in endheaders
> >>>>>>>>>>>>>>            self._send_output(message_body)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/httplib.py", line 835,
> in _send_output
> >>>>>>>>>>>>>>            self.send(msg)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/httplib.py", line 797,
> in send
> >>>>>>>>>>>>>>            self.connect()
> >>>>>>>>>>>>>>          File
> "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect
> >>>>>>>>>>>>>>            sock = socket.create_connection((self.host,
> self.port), self.timeout)
> >>>>>>>>>>>>>>          File "/usr/lib64/python2.7/socket.py", line 571,
> in create_connection
> >>>>>>>>>>>>>>            raise err
> >>>>>>>>>>>>>>        error: [Errno 101] Network is unreachable
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>       Yaniv, can you please take also a look to this one?
> >>>>>>>>>>>>>       it's exactly the opposite of
> https://bugzilla.redhat.com/1358530
> >>>>>>>>>>>>>       Here the jsonrpcclient works but not the xmlrpc one.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>>        25.07.2016, 11:17, "Simone Tiraboschi" <
> stirabos at redhat.com>:
> >>>>>>>>>>>>>>>        On Mon, Jul 25, 2016 at 7:51 AM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>>>>>         Simone, there is something interesting in the
> vdsm.log?
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>        For what I saw the issue is not related to the
> storage but to the network.
> >>>>>>>>>>>>>>>        ovirt-hosted-engine-setup uses the jsonrpc client,
> instead the code
> >>>>>>>>>>>>>>>        from ovirt-hosted-engine-ha still uses the xmlrpc
> client somewhere and
> >>>>>>>>>>>>>>>        this happens also when the setup asks to create the
> lockspace volume.
> >>>>>>>>>>>>>>>        It seams that in your case the xmlrpc client could
> not connect vdsm on
> >>>>>>>>>>>>>>>        the localhost.
> >>>>>>>>>>>>>>>        It could be somehow related to:
> >>>>>>>>>>>>>>>        https://bugzilla.redhat.com/1358530
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>        Can you please try executing
> >>>>>>>>>>>>>>>         sudo vdsClient -s 0 getVdsCaps
> >>>>>>>>>>>>>>>        on that host?
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>         22.07.2016, 19:36, "aleksey.maksimov at it-kb.ru" <
> aleksey.maksimov at it-kb.ru>:
> >>>>>>>>>>>>>>>>>         Simone, thanks for link.
> >>>>>>>>>>>>>>>>>         vdsm.log attached
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>         22.07.2016, 19:28, "Simone Tiraboschi" <
> stirabos at redhat.com>:
> >>>>>>>>>>>>>>>>>>          On Fri, Jul 22, 2016 at 5:59 PM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>>>>>>>>           Thank you for your response, Simone.
> >>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>           Log attached.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>          It seams it comes from VDSM, can you please
> attach also vdsm.log?
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>           I don't use ovirt-engine-appliance because I
> have not found "how-to" for ovirt-engine-appliance deployment in hosted
> engine configuration.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>          yum install ovirt-engine-appliance
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>          Then follow the instruction here:
> >>>>>>>>>>>>>>>>>>
> http://www.ovirt.org/develop/release-management/features/heapplianceflow/
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>           22.07.2016, 17:09, "Simone Tiraboschi" <
> stirabos at redhat.com>:
> >>>>>>>>>>>>>>>>>>>>           Hi Aleksey,
> >>>>>>>>>>>>>>>>>>>>           Can you please attach hosted-engine-setup
> logs?
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>           On Fri, Jul 22, 2016 at 3:46 PM, <
> aleksey.maksimov at it-kb.ru> wrote:
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            Hello oVirt guru`s !
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            I have problem with initial deploy of
> ovirt 4.0 hosted engine.
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            My environment :
> >>>>>>>>>>>>>>>>>>>>>            ============================
> >>>>>>>>>>>>>>>>>>>>>            * Two servers HP ProLiant DL 360 G5 with
> Qlogic FC HBA connected (with
> >>>>>>>>>>>>>>>>>>>>>            multipathd) to storage HP 3PAR 7200
> >>>>>>>>>>>>>>>>>>>>>            * On each server installed CentOS 7.2
> Linux (3.10.0-327.22.2.el7.x86_64)
> >>>>>>>>>>>>>>>>>>>>>            * On 3PAR storage I created 2 LUNs for
> oVirt.
> >>>>>>>>>>>>>>>>>>>>>            - First LUN for oVirt Hosted Engine VM
> (60GB)
> >>>>>>>>>>>>>>>>>>>>>            - Second LUN for all other VMs (2TB)
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            # multipath -ll
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            3par-vv1 (360002ac0000000000000001b0000cec9)
> dm-0 3PARdata,VV
> >>>>>>>>>>>>>>>>>>>>>            size=60G features='1 queue_if_no_path'
> hwhandler='1 alua' wp=rw
> >>>>>>>>>>>>>>>>>>>>>            `-+- policy='round-robin 0' prio=50
> status=active
> >>>>>>>>>>>>>>>>>>>>>              |- 2:0:1:1 sdd 8:48 active ready running
> >>>>>>>>>>>>>>>>>>>>>              |- 3:0:0:1 sdf 8:80 active ready running
> >>>>>>>>>>>>>>>>>>>>>              |- 2:0:0:1 sdb 8:16 active ready running
> >>>>>>>>>>>>>>>>>>>>>              `- 3:0:1:1 sdh 8:112 active ready
> running
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            3par-vv2 (360002ac000000000000000160000cec9)
> dm-1 3PARdata,VV
> >>>>>>>>>>>>>>>>>>>>>            size=2.0T features='1 queue_if_no_path'
> hwhandler='1 alua' wp=rw
> >>>>>>>>>>>>>>>>>>>>>            `-+- policy='round-robin 0' prio=50
> status=active
> >>>>>>>>>>>>>>>>>>>>>              |- 2:0:0:0 sda 8:0 active ready running
> >>>>>>>>>>>>>>>>>>>>>              |- 3:0:0:0 sde 8:64 active ready running
> >>>>>>>>>>>>>>>>>>>>>              |- 2:0:1:0 sdc 8:32 active ready running
> >>>>>>>>>>>>>>>>>>>>>              `- 3:0:1:0 sdg 8:96 active ready running
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            My steps on first server (initial deploy
> of ovirt 4.0 hosted engine):
> >>>>>>>>>>>>>>>>>>>>>            ============================
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            # systemctl stop NetworkManager
> >>>>>>>>>>>>>>>>>>>>>            # systemctl disable NetworkManager
> >>>>>>>>>>>>>>>>>>>>>            # yum -y install
> http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
> >>>>>>>>>>>>>>>>>>>>>            # yum -y install epel-release
> >>>>>>>>>>>>>>>>>>>>>            # wget
> >>>>>>>>>>>>>>>>>>>>>
> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso
> >>>>>>>>>>>>>>>>>>>>>            -P /tmp/
> >>>>>>>>>>>>>>>>>>>>>            # yum install ovirt-hosted-engine-setup
> >>>>>>>>>>>>>>>>>>>>>            # yum install screen
> >>>>>>>>>>>>>>>>>>>>>            # screen -RD
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            ...in screen session :
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            # hosted-engine --deploy
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>            in configuration process I chose "fc" as
> storage type for oVirt hosted
> >>>>>>>>>>>>>>>>>>>>>            engine vm and select 60GB LUN...
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            --== CONFIGURATION PREVIEW ==--
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>                      Firewall manager : iptables
> >>>>>>>>>>>>>>>>>>>>>                      Gateway address : 10.1.0.1
> >>>>>>>>>>>>>>>>>>>>>                      Host name for web application :
> KOM-AD01-OVIRT1
> >>>>>>>>>>>>>>>>>>>>>                      Storage Domain type : fc
> >>>>>>>>>>>>>>>>>>>>>                      Host ID : 1
> >>>>>>>>>>>>>>>>>>>>>                      LUN ID :
> >>>>>>>>>>>>>>>>>>>>>            360002ac0000000000000001b0000cec9
> >>>>>>>>>>>>>>>>>>>>>                      Image size GB : 40
> >>>>>>>>>>>>>>>>>>>>>                      Console type : vnc
> >>>>>>>>>>>>>>>>>>>>>                      Memory size MB : 4096
> >>>>>>>>>>>>>>>>>>>>>                      MAC address : 00:16:3e:77:1d:07
> >>>>>>>>>>>>>>>>>>>>>                      Boot type : cdrom
> >>>>>>>>>>>>>>>>>>>>>                      Number of CPUs : 2
> >>>>>>>>>>>>>>>>>>>>>                      ISO image (cdrom
> boot/cloud-init) :
> >>>>>>>>>>>>>>>>>>>>>            /tmp/CentOS-7-x86_64-NetInstall-1511.iso
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>           Can I ask why you prefer/need to manually
> create a VM installing from
> >>>>>>>>>>>>>>>>>>>>           a CD instead of using the ready-to-use
> ovirt-engine-appliance?
> >>>>>>>>>>>>>>>>>>>>           Using the appliance makes the setup process
> a lot shorted and more comfortable.
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>                      CPU Type : model_Penryn
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>            and get error after step "Verifying
> sanlock lockspace initialization"
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Verifying sanlock lockspace
> initialization
> >>>>>>>>>>>>>>>>>>>>>            [ ERROR ] Failed to execute stage 'Misc
> configuration': [Errno 101] Network
> >>>>>>>>>>>>>>>>>>>>>            is unreachable
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Clean up
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Generating answer file
> >>>>>>>>>>>>>>>>>>>>>
> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Pre-termination
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Termination
> >>>>>>>>>>>>>>>>>>>>>            [ ERROR ] Hosted Engine deployment
> failed: this system is not reliable,
> >>>>>>>>>>>>>>>>>>>>>            please check the issue, fix and redeploy
> >>>>>>>>>>>>>>>>>>>>>                      Log file is located at
> >>>>>>>>>>>>>>>>>>>>>
> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-
> 20160722123404-t26vw0.log
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            Interestingly
> >>>>>>>>>>>>>>>>>>>>>            ============================
> >>>>>>>>>>>>>>>>>>>>>            If I try to deploy hosted-engine v3.6,
> everything goes well in the same
> >>>>>>>>>>>>>>>>>>>>>            configuration !! :
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            ....
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Transaction setup
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Misc configuration
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Package installation
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Misc configuration
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Configuring libvirt
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Configuring VDSM
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Starting vdsmd
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Waiting for VDSM hardware info
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Configuring the management bridge
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Volume Group
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Storage Domain
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Storage Pool
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Connecting Storage Pool
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Verifying sanlock lockspace
> initialization
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Image for
> 'hosted-engine.lockspace' ...
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Image for
> 'hosted-engine.lockspace' created successfully
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating Image for
> 'hosted-engine.metadata' ...
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Image for
> 'hosted-engine.metadata' created successfully
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating VM Image
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Destroying Storage Pool
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Start monitoring domain
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Configuring VM
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Updating hosted-engine
> configuration
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Transaction commit
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Stage: Closing up
> >>>>>>>>>>>>>>>>>>>>>            [ INFO ] Creating VM
> >>>>>>>>>>>>>>>>>>>>>                      You can now connect to the VM
> with the following command:
> >>>>>>>>>>>>>>>>>>>>>                            /bin/remote-viewer
> vnc://localhost:5900
> >>>>>>>>>>>>>>>>>>>>>            ...
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>            What could be the problem?
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>>
> _______________________________________________
> >>>>>>>>>>>>>>>>>>>>>            Users mailing list
> >>>>>>>>>>>>>>>>>>>>>            Users at ovirt.org
> >>>>>>>>>>>>>>>>>>>>>
> http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>>>>>>>       _______________________________________________
> >>>>>>>>>>>>       Users mailing list
> >>>>>>>>>>>>       Users at ovirt.org
> >>>>>>>>>>>>       http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>>>>
> >>>>>>>>>     _______________________________________________
> >>>>>>>>>     Users mailing list
> >>>>>>>>>     Users at ovirt.org
> >>>>>>>>>     http://lists.ovirt.org/mailman/listinfo/users
> >>>>>>
> >>>>>>   _______________________________________________
> >>>>>>   Users mailing list
> >>>>>>   Users at ovirt.org
> >>>>>>   http://lists.ovirt.org/mailman/listinfo/users
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20160727/8e439e5a/attachment-0001.html>


More information about the Users mailing list