
# ss -plutn Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6)) udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6)) udp UNCONN 0 0 127.0.0.1:323 *:* users:(("chronyd",pid=795,fd=1)) udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7)) udp UNCONN 0 0 127.0.0.1:25375 *:* users:(("snmpd",pid=1609,fd=8)) udp UNCONN 0 0 127.0.0.1:25376 *:* users:(("cmapeerd",pid=2056,fd=5)) udp UNCONN 0 0 127.0.0.1:25393 *:* users:(("cmanicd",pid=2278,fd=3)) udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9)) udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10)) tcp LISTEN 0 128 *:2381 *:* users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4)) tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8)) tcp LISTEN 0 5 *:54322 *:* users:(("ovirt-imageio-d",pid=753,fd=3)) tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3)) tcp LISTEN 0 100 127.0.0.1:25 *:* users:(("master",pid=1948,fd=13)) tcp LISTEN 0 128 *:2301 *:* users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3)) tcp LISTEN 0 30 *:16514 *:* users:(("libvirtd",pid=10688,fd=13)) tcp LISTEN 0 128 127.0.0.1:199 *:* users:(("snmpd",pid=1609,fd=9)) tcp LISTEN 0 128 :::111 :::* users:(("rpcbind",pid=827,fd=11)) tcp LISTEN 0 5 :::54321 :::* users:(("vdsm",pid=11077,fd=23)) tcp LISTEN 0 30 :::16514 :::* users:(("libvirtd",pid=10688,fd=14)) 25.07.2016, 15:11, "Simone Tiraboschi" <stirabos@redhat.com>:
On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov@it-kb.ru> wrote:
Yes.
# ping $(python -c 'import socket; print(socket.gethostname())')
PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data. 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 time=0.030 ms 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 time=0.015 ms 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 time=0.011 ms ^C --- KOM-AD01-VM31.holding.com ping statistics --- 3 packets transmitted, 3 received, 0% packet loss, time 2001ms rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
but...
and the output of ss -plutn
# vdsClient -s 0 getVdsCaps
Traceback (most recent call last): File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> code, message = commands[command][0](commandArgs) File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap return self.ExecAndExit(self.s.getVdsCapabilities()) File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ return self.__send(self.__name, args) File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request verbose=self.__verbose File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request return self.single_request(host, handler, request_body, verbose) File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request self.send_content(h, request_body) File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content connection.endheaders(request_body) File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders self._send_output(message_body) File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output self.send(msg) File "/usr/lib64/python2.7/httplib.py", line 797, in send self.connect() File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect sock = socket.create_connection((self.host, self.port), self.timeout) File "/usr/lib64/python2.7/socket.py", line 571, in create_connection raise err error: [Errno 101] Network is unreachable
25.07.2016, 14:58, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>:
Ok.
1) I stopped and disabled the service NetworkManager # systemctl stop NetworkManager # systemctl disable NetworkManager
2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in ifcfg-* file.
3) Reboot server
4) Try deploy oVirt HE 4 and I get the same error
[ INFO ] Creating Volume Group [ INFO ] Creating Storage Domain [ INFO ] Creating Storage Pool [ INFO ] Connecting Storage Pool [ INFO ] Verifying sanlock lockspace initialization [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable [ INFO ] Stage: Clean up [ INFO ] Generating answer file '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf' [ INFO ] Stage: Pre-termination [ INFO ] Stage: Termination [ ERROR ] Hosted Engine deployment failed: this system is not reliable, please check the issue, fix and redeploy Log file is located at /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
What ideas further?
25.07.2016, 13:06, "Simone Tiraboschi" <stirabos@redhat.com>:
On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksimov@it-kb.ru> wrote:
What am I supposed to do for successfully deploy ovirt 4 ? Any ideas ?
Can you please try to explicitly configure your DNS with nameserver under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for the interface you are going to use?
25.07.2016, 12:47, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>:
"Aleksey, was your DNS configured with DNS1 and DNS2 just on the interface you used to create the management bridge on?"
Yes. Of course
25.07.2016, 12:27, "Simone Tiraboschi" <stirabos@redhat.com>: > On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski > <piotr.kliczewski@gmail.com> wrote: >> This could be the issue here as well as for BZ #1358530 >> >> On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksimov@it-kb.ru> wrote: >>> Could this be due to the fact that the ovirt installer has changed network configuration files (ifcfg-*, resolv.conf) ? >>> After the error in ovirt installation process I see from resolv.conf disappeared on my DNS servers entry and now the server is unable to resolve names. > > So it could be related to https://bugzilla.redhat.com/show_bug.cgi?id=1160423 > > Aleksey, was your DNS configured with DNS1 and DNS2 just on the > interface you used to create the management bridge on? > Can you please try the workaround described here > https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ? > >>> 25.07.2016, 11:26, "Simone Tiraboschi" <stirabos@redhat.com>: >>>> On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksimov@it-kb.ru> wrote: >>>>> # vdsClient -s 0 getVdsCaps >>>>> >>>>> Traceback (most recent call last): >>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> >>>>> code, message = commands[command][0](commandArgs) >>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap >>>>> return self.ExecAndExit(self.s.getVdsCapabilities()) >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ >>>>> return self.__send(self.__name, args) >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request >>>>> verbose=self.__verbose >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request >>>>> return self.single_request(host, handler, request_body, verbose) >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request >>>>> self.send_content(h, request_body) >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content >>>>> connection.endheaders(request_body) >>>>> File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders >>>>> self._send_output(message_body) >>>>> File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output >>>>> self.send(msg) >>>>> File "/usr/lib64/python2.7/httplib.py", line 797, in send >>>>> self.connect() >>>>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect >>>>> sock = socket.create_connection((self.host, self.port), self.timeout) >>>>> File "/usr/lib64/python2.7/socket.py", line 571, in create_connection >>>>> raise err >>>>> error: [Errno 101] Network is unreachable >>>> >>>> Yaniv, can you please take also a look to this one? >>>> it's exactly the opposite of https://bugzilla.redhat.com/1358530 >>>> Here the jsonrpcclient works but not the xmlrpc one. >>>> >>>>> 25.07.2016, 11:17, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>> On Mon, Jul 25, 2016 at 7:51 AM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>> Simone, there is something interesting in the vdsm.log? >>>>>> >>>>>> For what I saw the issue is not related to the storage but to the network. >>>>>> ovirt-hosted-engine-setup uses the jsonrpc client, instead the code >>>>>> from ovirt-hosted-engine-ha still uses the xmlrpc client somewhere and >>>>>> this happens also when the setup asks to create the lockspace volume. >>>>>> It seams that in your case the xmlrpc client could not connect vdsm on >>>>>> the localhost. >>>>>> It could be somehow related to: >>>>>> https://bugzilla.redhat.com/1358530 >>>>>> >>>>>> Can you please try executing >>>>>> sudo vdsClient -s 0 getVdsCaps >>>>>> on that host? >>>>>> >>>>>>> 22.07.2016, 19:36, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>: >>>>>>>> Simone, thanks for link. >>>>>>>> vdsm.log attached >>>>>>>> >>>>>>>> 22.07.2016, 19:28, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>>>> On Fri, Jul 22, 2016 at 5:59 PM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>>>> Thank you for your response, Simone. >>>>>>>>>> >>>>>>>>>> Log attached. >>>>>>>>> >>>>>>>>> It seams it comes from VDSM, can you please attach also vdsm.log? >>>>>>>>> >>>>>>>>>> I don't use ovirt-engine-appliance because I have not found "how-to" for ovirt-engine-appliance deployment in hosted engine configuration. >>>>>>>>> >>>>>>>>> yum install ovirt-engine-appliance >>>>>>>>> >>>>>>>>> Then follow the instruction here: >>>>>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/ >>>>>>>>> >>>>>>>>>> 22.07.2016, 17:09, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>>>>>> Hi Aleksey, >>>>>>>>>>> Can you please attach hosted-engine-setup logs? >>>>>>>>>>> >>>>>>>>>>> On Fri, Jul 22, 2016 at 3:46 PM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hello oVirt guru`s ! >>>>>>>>>>>> >>>>>>>>>>>> I have problem with initial deploy of ovirt 4.0 hosted engine. >>>>>>>>>>>> >>>>>>>>>>>> My environment : >>>>>>>>>>>> ============================ >>>>>>>>>>>> * Two servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected (with >>>>>>>>>>>> multipathd) to storage HP 3PAR 7200 >>>>>>>>>>>> * On each server installed CentOS 7.2 Linux (3.10.0-327.22.2.el7.x86_64) >>>>>>>>>>>> * On 3PAR storage I created 2 LUNs for oVirt. >>>>>>>>>>>> - First LUN for oVirt Hosted Engine VM (60GB) >>>>>>>>>>>> - Second LUN for all other VMs (2TB) >>>>>>>>>>>> >>>>>>>>>>>> # multipath -ll >>>>>>>>>>>> >>>>>>>>>>>> 3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV >>>>>>>>>>>> size=60G features='1 queue_if_no_path' hwhandler='1 alua' wp=rw >>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 status=active >>>>>>>>>>>> |- 2:0:1:1 sdd 8:48 active ready running >>>>>>>>>>>> |- 3:0:0:1 sdf 8:80 active ready running >>>>>>>>>>>> |- 2:0:0:1 sdb 8:16 active ready running >>>>>>>>>>>> `- 3:0:1:1 sdh 8:112 active ready running >>>>>>>>>>>> >>>>>>>>>>>> 3par-vv2 (360002ac000000000000000160000cec9) dm-1 3PARdata,VV >>>>>>>>>>>> size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw >>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 status=active >>>>>>>>>>>> |- 2:0:0:0 sda 8:0 active ready running >>>>>>>>>>>> |- 3:0:0:0 sde 8:64 active ready running >>>>>>>>>>>> |- 2:0:1:0 sdc 8:32 active ready running >>>>>>>>>>>> `- 3:0:1:0 sdg 8:96 active ready running >>>>>>>>>>>> >>>>>>>>>>>> My steps on first server (initial deploy of ovirt 4.0 hosted engine): >>>>>>>>>>>> ============================ >>>>>>>>>>>> >>>>>>>>>>>> # systemctl stop NetworkManager >>>>>>>>>>>> # systemctl disable NetworkManager >>>>>>>>>>>> # yum -y install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm >>>>>>>>>>>> # yum -y install epel-release >>>>>>>>>>>> # wget >>>>>>>>>>>> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511... >>>>>>>>>>>> -P /tmp/ >>>>>>>>>>>> # yum install ovirt-hosted-engine-setup >>>>>>>>>>>> # yum install screen >>>>>>>>>>>> # screen -RD >>>>>>>>>>>> >>>>>>>>>>>> ...in screen session : >>>>>>>>>>>> >>>>>>>>>>>> # hosted-engine --deploy >>>>>>>>>>>> >>>>>>>>>>>> ... >>>>>>>>>>>> in configuration process I chose "fc" as storage type for oVirt hosted >>>>>>>>>>>> engine vm and select 60GB LUN... >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> --== CONFIGURATION PREVIEW ==-- >>>>>>>>>>>> >>>>>>>>>>>> ... >>>>>>>>>>>> Firewall manager : iptables >>>>>>>>>>>> Gateway address : 10.1.0.1 >>>>>>>>>>>> Host name for web application : KOM-AD01-OVIRT1 >>>>>>>>>>>> Storage Domain type : fc >>>>>>>>>>>> Host ID : 1 >>>>>>>>>>>> LUN ID : >>>>>>>>>>>> 360002ac0000000000000001b0000cec9 >>>>>>>>>>>> Image size GB : 40 >>>>>>>>>>>> Console type : vnc >>>>>>>>>>>> Memory size MB : 4096 >>>>>>>>>>>> MAC address : 00:16:3e:77:1d:07 >>>>>>>>>>>> Boot type : cdrom >>>>>>>>>>>> Number of CPUs : 2 >>>>>>>>>>>> ISO image (cdrom boot/cloud-init) : >>>>>>>>>>>> /tmp/CentOS-7-x86_64-NetInstall-1511.iso >>>>>>>>>>> >>>>>>>>>>> Can I ask why you prefer/need to manually create a VM installing from >>>>>>>>>>> a CD instead of using the ready-to-use ovirt-engine-appliance? >>>>>>>>>>> Using the appliance makes the setup process a lot shorted and more comfortable. >>>>>>>>>>> >>>>>>>>>>>> CPU Type : model_Penryn >>>>>>>>>>>> ... >>>>>>>>>>>> and get error after step "Verifying sanlock lockspace initialization" >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace initialization >>>>>>>>>>>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network >>>>>>>>>>>> is unreachable >>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf' >>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, >>>>>>>>>>>> please check the issue, fix and redeploy >>>>>>>>>>>> Log file is located at >>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log >>>>>>>>>>>> >>>>>>>>>>>> Interestingly >>>>>>>>>>>> ============================ >>>>>>>>>>>> If I try to deploy hosted-engine v3.6, everything goes well in the same >>>>>>>>>>>> configuration !! : >>>>>>>>>>>> >>>>>>>>>>>> .... >>>>>>>>>>>> [ INFO ] Stage: Transaction setup >>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>> [ INFO ] Stage: Package installation >>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>> [ INFO ] Configuring libvirt >>>>>>>>>>>> [ INFO ] Configuring VDSM >>>>>>>>>>>> [ INFO ] Starting vdsmd >>>>>>>>>>>> [ INFO ] Waiting for VDSM hardware info >>>>>>>>>>>> [ INFO ] Configuring the management bridge >>>>>>>>>>>> [ INFO ] Creating Volume Group >>>>>>>>>>>> [ INFO ] Creating Storage Domain >>>>>>>>>>>> [ INFO ] Creating Storage Pool >>>>>>>>>>>> [ INFO ] Connecting Storage Pool >>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace initialization >>>>>>>>>>>> [ INFO ] Creating Image for 'hosted-engine.lockspace' ... >>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.lockspace' created successfully >>>>>>>>>>>> [ INFO ] Creating Image for 'hosted-engine.metadata' ... >>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.metadata' created successfully >>>>>>>>>>>> [ INFO ] Creating VM Image >>>>>>>>>>>> [ INFO ] Destroying Storage Pool >>>>>>>>>>>> [ INFO ] Start monitoring domain >>>>>>>>>>>> [ INFO ] Configuring VM >>>>>>>>>>>> [ INFO ] Updating hosted-engine configuration >>>>>>>>>>>> [ INFO ] Stage: Transaction commit >>>>>>>>>>>> [ INFO ] Stage: Closing up >>>>>>>>>>>> [ INFO ] Creating VM >>>>>>>>>>>> You can now connect to the VM with the following command: >>>>>>>>>>>> /bin/remote-viewer vnc://localhost:5900 >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> What could be the problem? >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Users mailing list >>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>> _______________________________________________ >>> Users mailing list >>> Users@ovirt.org >>> http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users