[ovirt-users] oVirt 4 Hosted Engine deploy on fc storage - [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable

Simone Tiraboschi stirabos at redhat.com
Mon Jul 25 12:10:32 UTC 2016


On Mon, Jul 25, 2016 at 2:03 PM,  <aleksey.maksimov at it-kb.ru> wrote:
> Yes.
>
> # ping $(python -c 'import socket; print(socket.gethostname())')
>
> PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 time=0.030 ms
> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 time=0.015 ms
> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 time=0.011 ms
> ^C
> --- KOM-AD01-VM31.holding.com ping statistics ---
> 3 packets transmitted, 3 received, 0% packet loss, time 2001ms
> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
>
>
> but...

and the output of
 ss -plutn

> # vdsClient -s 0 getVdsCaps
>
> Traceback (most recent call last):
>   File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>     code, message = commands[command][0](commandArgs)
>   File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>     return self.ExecAndExit(self.s.getVdsCapabilities())
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>     return self.__send(self.__name, args)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>     verbose=self.__verbose
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>     return self.single_request(host, handler, request_body, verbose)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>     self.send_content(h, request_body)
>   File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>     connection.endheaders(request_body)
>   File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>     self._send_output(message_body)
>   File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>     self.send(msg)
>   File "/usr/lib64/python2.7/httplib.py", line 797, in send
>     self.connect()
>   File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect
>     sock = socket.create_connection((self.host, self.port), self.timeout)
>   File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
>     raise err
> error: [Errno 101] Network is unreachable
>
>
>
>
> 25.07.2016, 14:58, "aleksey.maksimov at it-kb.ru" <aleksey.maksimov at it-kb.ru>:
>> Ok.
>>
>> 1) I stopped and disabled the service NetworkManager
>> # systemctl stop NetworkManager
>> # systemctl disable NetworkManager
>>
>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in ifcfg-* file.
>>
>> 3) Reboot server
>>
>> 4) Try deploy oVirt HE 4 and I get the same error
>>
>> [ INFO ] Creating Volume Group
>> [ INFO ] Creating Storage Domain
>> [ INFO ] Creating Storage Pool
>> [ INFO ] Connecting Storage Pool
>> [ INFO ] Verifying sanlock lockspace initialization
>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable
>> [ INFO ] Stage: Clean up
>> [ INFO ] Generating answer file '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
>> [ INFO ] Stage: Pre-termination
>> [ INFO ] Stage: Termination
>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, please check the issue, fix and redeploy
>>           Log file is located at /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
>>
>> What ideas further?
>>
>> 25.07.2016, 13:06, "Simone Tiraboschi" <stirabos at redhat.com>:
>>>  On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksimov at it-kb.ru> wrote:
>>>>   What am I supposed to do for successfully deploy ovirt 4 ?
>>>>   Any ideas ?
>>>
>>>  Can you please try to explicitly configure your DNS with nameserver
>>>  under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
>>>  the interface you are going to use?
>>>
>>>>   25.07.2016, 12:47, "aleksey.maksimov at it-kb.ru" <aleksey.maksimov at it-kb.ru>:
>>>>>   "Aleksey, was your DNS configured with DNS1 and DNS2 just on the interface you used to create the management bridge on?"
>>>>>
>>>>>   Yes. Of course
>>>>>
>>>>>   25.07.2016, 12:27, "Simone Tiraboschi" <stirabos at redhat.com>:
>>>>>>    On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>>>>    <piotr.kliczewski at gmail.com> wrote:
>>>>>>>     This could be the issue here as well as for BZ #1358530
>>>>>>>
>>>>>>>     On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksimov at it-kb.ru> wrote:
>>>>>>>>     Could this be due to the fact that the ovirt installer has changed network configuration files (ifcfg-*, resolv.conf) ?
>>>>>>>>     After the error in ovirt installation process I see from resolv.conf disappeared on my DNS servers entry and now the server is unable to resolve names.
>>>>>>
>>>>>>    So it could be related to https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>>>>
>>>>>>    Aleksey, was your DNS configured with DNS1 and DNS2 just on the
>>>>>>    interface you used to create the management bridge on?
>>>>>>    Can you please try the workaround described here
>>>>>>    https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
>>>>>>
>>>>>>>>     25.07.2016, 11:26, "Simone Tiraboschi" <stirabos at redhat.com>:
>>>>>>>>>     On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksimov at it-kb.ru> wrote:
>>>>>>>>>>      # vdsClient -s 0 getVdsCaps
>>>>>>>>>>
>>>>>>>>>>      Traceback (most recent call last):
>>>>>>>>>>        File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>>>>>>>>>          code, message = commands[command][0](commandArgs)
>>>>>>>>>>        File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>>>>          return self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>>>>        File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>>>>>>>>          return self.__send(self.__name, args)
>>>>>>>>>>        File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>>>>>>>>          verbose=self.__verbose
>>>>>>>>>>        File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>>>>>>>>          return self.single_request(host, handler, request_body, verbose)
>>>>>>>>>>        File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>>>>>>>>>>          self.send_content(h, request_body)
>>>>>>>>>>        File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>>>>>>>>>>          connection.endheaders(request_body)
>>>>>>>>>>        File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>>>>>>>>          self._send_output(message_body)
>>>>>>>>>>        File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>>>>>>>>          self.send(msg)
>>>>>>>>>>        File "/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>>>>>>>          self.connect()
>>>>>>>>>>        File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect
>>>>>>>>>>          sock = socket.create_connection((self.host, self.port), self.timeout)
>>>>>>>>>>        File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
>>>>>>>>>>          raise err
>>>>>>>>>>      error: [Errno 101] Network is unreachable
>>>>>>>>>
>>>>>>>>>     Yaniv, can you please take also a look to this one?
>>>>>>>>>     it's exactly the opposite of https://bugzilla.redhat.com/1358530
>>>>>>>>>     Here the jsonrpcclient works but not the xmlrpc one.
>>>>>>>>>
>>>>>>>>>>      25.07.2016, 11:17, "Simone Tiraboschi" <stirabos at redhat.com>:
>>>>>>>>>>>      On Mon, Jul 25, 2016 at 7:51 AM, <aleksey.maksimov at it-kb.ru> wrote:
>>>>>>>>>>>>       Simone, there is something interesting in the vdsm.log?
>>>>>>>>>>>
>>>>>>>>>>>      For what I saw the issue is not related to the storage but to the network.
>>>>>>>>>>>      ovirt-hosted-engine-setup uses the jsonrpc client, instead the code
>>>>>>>>>>>      from ovirt-hosted-engine-ha still uses the xmlrpc client somewhere and
>>>>>>>>>>>      this happens also when the setup asks to create the lockspace volume.
>>>>>>>>>>>      It seams that in your case the xmlrpc client could not connect vdsm on
>>>>>>>>>>>      the localhost.
>>>>>>>>>>>      It could be somehow related to:
>>>>>>>>>>>      https://bugzilla.redhat.com/1358530
>>>>>>>>>>>
>>>>>>>>>>>      Can you please try executing
>>>>>>>>>>>       sudo vdsClient -s 0 getVdsCaps
>>>>>>>>>>>      on that host?
>>>>>>>>>>>
>>>>>>>>>>>>       22.07.2016, 19:36, "aleksey.maksimov at it-kb.ru" <aleksey.maksimov at it-kb.ru>:
>>>>>>>>>>>>>       Simone, thanks for link.
>>>>>>>>>>>>>       vdsm.log attached
>>>>>>>>>>>>>
>>>>>>>>>>>>>       22.07.2016, 19:28, "Simone Tiraboschi" <stirabos at redhat.com>:
>>>>>>>>>>>>>>        On Fri, Jul 22, 2016 at 5:59 PM, <aleksey.maksimov at it-kb.ru> wrote:
>>>>>>>>>>>>>>>         Thank you for your response, Simone.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         Log attached.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        It seams it comes from VDSM, can you please attach also vdsm.log?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         I don't use ovirt-engine-appliance because I have not found "how-to" for ovirt-engine-appliance deployment in hosted engine configuration.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        yum install ovirt-engine-appliance
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>        Then follow the instruction here:
>>>>>>>>>>>>>>        http://www.ovirt.org/develop/release-management/features/heapplianceflow/
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>         22.07.2016, 17:09, "Simone Tiraboschi" <stirabos at redhat.com>:
>>>>>>>>>>>>>>>>         Hi Aleksey,
>>>>>>>>>>>>>>>>         Can you please attach hosted-engine-setup logs?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>         On Fri, Jul 22, 2016 at 3:46 PM, <aleksey.maksimov at it-kb.ru> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          Hello oVirt guru`s !
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          I have problem with initial deploy of ovirt 4.0 hosted engine.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          My environment :
>>>>>>>>>>>>>>>>>          ============================
>>>>>>>>>>>>>>>>>          * Two servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected (with
>>>>>>>>>>>>>>>>>          multipathd) to storage HP 3PAR 7200
>>>>>>>>>>>>>>>>>          * On each server installed CentOS 7.2 Linux (3.10.0-327.22.2.el7.x86_64)
>>>>>>>>>>>>>>>>>          * On 3PAR storage I created 2 LUNs for oVirt.
>>>>>>>>>>>>>>>>>          - First LUN for oVirt Hosted Engine VM (60GB)
>>>>>>>>>>>>>>>>>          - Second LUN for all other VMs (2TB)
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          # multipath -ll
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV
>>>>>>>>>>>>>>>>>          size=60G features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>          `-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>            |- 2:0:1:1 sdd 8:48 active ready running
>>>>>>>>>>>>>>>>>            |- 3:0:0:1 sdf 8:80 active ready running
>>>>>>>>>>>>>>>>>            |- 2:0:0:1 sdb 8:16 active ready running
>>>>>>>>>>>>>>>>>            `- 3:0:1:1 sdh 8:112 active ready running
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          3par-vv2 (360002ac000000000000000160000cec9) dm-1 3PARdata,VV
>>>>>>>>>>>>>>>>>          size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>          `-+- policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>            |- 2:0:0:0 sda 8:0 active ready running
>>>>>>>>>>>>>>>>>            |- 3:0:0:0 sde 8:64 active ready running
>>>>>>>>>>>>>>>>>            |- 2:0:1:0 sdc 8:32 active ready running
>>>>>>>>>>>>>>>>>            `- 3:0:1:0 sdg 8:96 active ready running
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          My steps on first server (initial deploy of ovirt 4.0 hosted engine):
>>>>>>>>>>>>>>>>>          ============================
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          # systemctl stop NetworkManager
>>>>>>>>>>>>>>>>>          # systemctl disable NetworkManager
>>>>>>>>>>>>>>>>>          # yum -y install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
>>>>>>>>>>>>>>>>>          # yum -y install epel-release
>>>>>>>>>>>>>>>>>          # wget
>>>>>>>>>>>>>>>>>          http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>>          -P /tmp/
>>>>>>>>>>>>>>>>>          # yum install ovirt-hosted-engine-setup
>>>>>>>>>>>>>>>>>          # yum install screen
>>>>>>>>>>>>>>>>>          # screen -RD
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          ...in screen session :
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          # hosted-engine --deploy
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          ...
>>>>>>>>>>>>>>>>>          in configuration process I chose "fc" as storage type for oVirt hosted
>>>>>>>>>>>>>>>>>          engine vm and select 60GB LUN...
>>>>>>>>>>>>>>>>>          ...
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          --== CONFIGURATION PREVIEW ==--
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          ...
>>>>>>>>>>>>>>>>>                    Firewall manager : iptables
>>>>>>>>>>>>>>>>>                    Gateway address : 10.1.0.1
>>>>>>>>>>>>>>>>>                    Host name for web application : KOM-AD01-OVIRT1
>>>>>>>>>>>>>>>>>                    Storage Domain type : fc
>>>>>>>>>>>>>>>>>                    Host ID : 1
>>>>>>>>>>>>>>>>>                    LUN ID :
>>>>>>>>>>>>>>>>>          360002ac0000000000000001b0000cec9
>>>>>>>>>>>>>>>>>                    Image size GB : 40
>>>>>>>>>>>>>>>>>                    Console type : vnc
>>>>>>>>>>>>>>>>>                    Memory size MB : 4096
>>>>>>>>>>>>>>>>>                    MAC address : 00:16:3e:77:1d:07
>>>>>>>>>>>>>>>>>                    Boot type : cdrom
>>>>>>>>>>>>>>>>>                    Number of CPUs : 2
>>>>>>>>>>>>>>>>>                    ISO image (cdrom boot/cloud-init) :
>>>>>>>>>>>>>>>>>          /tmp/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>         Can I ask why you prefer/need to manually create a VM installing from
>>>>>>>>>>>>>>>>         a CD instead of using the ready-to-use ovirt-engine-appliance?
>>>>>>>>>>>>>>>>         Using the appliance makes the setup process a lot shorted and more comfortable.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>                    CPU Type : model_Penryn
>>>>>>>>>>>>>>>>>          ...
>>>>>>>>>>>>>>>>>          and get error after step "Verifying sanlock lockspace initialization"
>>>>>>>>>>>>>>>>>          ...
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          [ INFO ] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>>          [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network
>>>>>>>>>>>>>>>>>          is unreachable
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Clean up
>>>>>>>>>>>>>>>>>          [ INFO ] Generating answer file
>>>>>>>>>>>>>>>>>          '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Pre-termination
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Termination
>>>>>>>>>>>>>>>>>          [ ERROR ] Hosted Engine deployment failed: this system is not reliable,
>>>>>>>>>>>>>>>>>          please check the issue, fix and redeploy
>>>>>>>>>>>>>>>>>                    Log file is located at
>>>>>>>>>>>>>>>>>          /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          Interestingly
>>>>>>>>>>>>>>>>>          ============================
>>>>>>>>>>>>>>>>>          If I try to deploy hosted-engine v3.6, everything goes well in the same
>>>>>>>>>>>>>>>>>          configuration !! :
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          ....
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Transaction setup
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Package installation
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Misc configuration
>>>>>>>>>>>>>>>>>          [ INFO ] Configuring libvirt
>>>>>>>>>>>>>>>>>          [ INFO ] Configuring VDSM
>>>>>>>>>>>>>>>>>          [ INFO ] Starting vdsmd
>>>>>>>>>>>>>>>>>          [ INFO ] Waiting for VDSM hardware info
>>>>>>>>>>>>>>>>>          [ INFO ] Configuring the management bridge
>>>>>>>>>>>>>>>>>          [ INFO ] Creating Volume Group
>>>>>>>>>>>>>>>>>          [ INFO ] Creating Storage Domain
>>>>>>>>>>>>>>>>>          [ INFO ] Creating Storage Pool
>>>>>>>>>>>>>>>>>          [ INFO ] Connecting Storage Pool
>>>>>>>>>>>>>>>>>          [ INFO ] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>>          [ INFO ] Creating Image for 'hosted-engine.lockspace' ...
>>>>>>>>>>>>>>>>>          [ INFO ] Image for 'hosted-engine.lockspace' created successfully
>>>>>>>>>>>>>>>>>          [ INFO ] Creating Image for 'hosted-engine.metadata' ...
>>>>>>>>>>>>>>>>>          [ INFO ] Image for 'hosted-engine.metadata' created successfully
>>>>>>>>>>>>>>>>>          [ INFO ] Creating VM Image
>>>>>>>>>>>>>>>>>          [ INFO ] Destroying Storage Pool
>>>>>>>>>>>>>>>>>          [ INFO ] Start monitoring domain
>>>>>>>>>>>>>>>>>          [ INFO ] Configuring VM
>>>>>>>>>>>>>>>>>          [ INFO ] Updating hosted-engine configuration
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Transaction commit
>>>>>>>>>>>>>>>>>          [ INFO ] Stage: Closing up
>>>>>>>>>>>>>>>>>          [ INFO ] Creating VM
>>>>>>>>>>>>>>>>>                    You can now connect to the VM with the following command:
>>>>>>>>>>>>>>>>>                          /bin/remote-viewer vnc://localhost:5900
>>>>>>>>>>>>>>>>>          ...
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          What could be the problem?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>          _______________________________________________
>>>>>>>>>>>>>>>>>          Users mailing list
>>>>>>>>>>>>>>>>>          Users at ovirt.org
>>>>>>>>>>>>>>>>>          http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>     _______________________________________________
>>>>>>>>     Users mailing list
>>>>>>>>     Users at ovirt.org
>>>>>>>>     http://lists.ovirt.org/mailman/listinfo/users
>>>>>
>>>>>   _______________________________________________
>>>>>   Users mailing list
>>>>>   Users at ovirt.org
>>>>>   http://lists.ovirt.org/mailman/listinfo/users
>>
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users



More information about the Users mailing list