
On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksimov@it-kb.ru> wrote:
# ss -plutn
Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port
udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6)) udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6)) udp UNCONN 0 0 127.0.0.1:323 *:* users:(("chronyd",pid=795,fd=1)) udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7)) udp UNCONN 0 0 127.0.0.1:25375 *:* users:(("snmpd",pid=1609,fd=8)) udp UNCONN 0 0 127.0.0.1:25376 *:* users:(("cmapeerd",pid=2056,fd=5)) udp UNCONN 0 0 127.0.0.1:25393 *:* users:(("cmanicd",pid=2278,fd=3)) udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9)) udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10)) tcp LISTEN 0 128 *:2381 *:* users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4)) tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8)) tcp LISTEN 0 5 *:54322 *:* users:(("ovirt-imageio-d",pid=753,fd=3)) tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3)) tcp LISTEN 0 100 127.0.0.1:25 *:* users:(("master",pid=1948,fd=13)) tcp LISTEN 0 128 *:2301 *:* users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3)) tcp LISTEN 0 30 *:16514 *:* users:(("libvirtd",pid=10688,fd=13)) tcp LISTEN 0 128 127.0.0.1:199 *:* users:(("snmpd",pid=1609,fd=9)) tcp LISTEN 0 128 :::111 :::* users:(("rpcbind",pid=827,fd=11)) tcp LISTEN 0 5 :::54321 :::* users:(("vdsm",pid=11077,fd=23))
vdsm is properly bind over ipv6. Can you please check if you can connect to vdsm with: telnet kom-ad01-vm31.holding.com 54321 and with telnet ::1 54321 ?
tcp LISTEN 0 30 :::16514 :::* users:(("libvirtd",pid=10688,fd=14))
25.07.2016, 15:11, "Simone Tiraboschi" <stirabos@redhat.com>:
On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov@it-kb.ru> wrote:
Yes.
# ping $(python -c 'import socket; print(socket.gethostname())')
PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data. 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 time=0.030 ms 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 time=0.015 ms 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 time=0.011 ms ^C --- KOM-AD01-VM31.holding.com ping statistics --- 3 packets transmitted, 3 received, 0% packet loss, time 2001ms rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
but...
and the output of ss -plutn
# vdsClient -s 0 getVdsCaps
Traceback (most recent call last): File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> code, message = commands[command][0](commandArgs) File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap return self.ExecAndExit(self.s.getVdsCapabilities()) File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ return self.__send(self.__name, args) File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request verbose=self.__verbose File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request return self.single_request(host, handler, request_body, verbose) File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request self.send_content(h, request_body) File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content connection.endheaders(request_body) File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders self._send_output(message_body) File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output self.send(msg) File "/usr/lib64/python2.7/httplib.py", line 797, in send self.connect() File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect sock = socket.create_connection((self.host, self.port), self.timeout) File "/usr/lib64/python2.7/socket.py", line 571, in create_connection raise err error: [Errno 101] Network is unreachable
25.07.2016, 14:58, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>:
Ok.
1) I stopped and disabled the service NetworkManager # systemctl stop NetworkManager # systemctl disable NetworkManager
2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in ifcfg-* file.
3) Reboot server
4) Try deploy oVirt HE 4 and I get the same error
[ INFO ] Creating Volume Group [ INFO ] Creating Storage Domain [ INFO ] Creating Storage Pool [ INFO ] Connecting Storage Pool [ INFO ] Verifying sanlock lockspace initialization [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable [ INFO ] Stage: Clean up [ INFO ] Generating answer file '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf' [ INFO ] Stage: Pre-termination [ INFO ] Stage: Termination [ ERROR ] Hosted Engine deployment failed: this system is not reliable, please check the issue, fix and redeploy Log file is located at /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
What ideas further?
25.07.2016, 13:06, "Simone Tiraboschi" <stirabos@redhat.com>:
On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksimov@it-kb.ru> wrote:
What am I supposed to do for successfully deploy ovirt 4 ? Any ideas ?
Can you please try to explicitly configure your DNS with nameserver under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for the interface you are going to use?
25.07.2016, 12:47, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>: > "Aleksey, was your DNS configured with DNS1 and DNS2 just on the interface you used to create the management bridge on?" > > Yes. Of course > > 25.07.2016, 12:27, "Simone Tiraboschi" <stirabos@redhat.com>: >> On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski >> <piotr.kliczewski@gmail.com> wrote: >>> This could be the issue here as well as for BZ #1358530 >>> >>> On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksimov@it-kb.ru> wrote: >>>> Could this be due to the fact that the ovirt installer has changed network configuration files (ifcfg-*, resolv.conf) ? >>>> After the error in ovirt installation process I see from resolv.conf disappeared on my DNS servers entry and now the server is unable to resolve names. >> >> So it could be related to https://bugzilla.redhat.com/show_bug.cgi?id=1160423 >> >> Aleksey, was your DNS configured with DNS1 and DNS2 just on the >> interface you used to create the management bridge on? >> Can you please try the workaround described here >> https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ? >> >>>> 25.07.2016, 11:26, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>> On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>> # vdsClient -s 0 getVdsCaps >>>>>> >>>>>> Traceback (most recent call last): >>>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> >>>>>> code, message = commands[command][0](commandArgs) >>>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap >>>>>> return self.ExecAndExit(self.s.getVdsCapabilities()) >>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ >>>>>> return self.__send(self.__name, args) >>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request >>>>>> verbose=self.__verbose >>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request >>>>>> return self.single_request(host, handler, request_body, verbose) >>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request >>>>>> self.send_content(h, request_body) >>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content >>>>>> connection.endheaders(request_body) >>>>>> File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders >>>>>> self._send_output(message_body) >>>>>> File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output >>>>>> self.send(msg) >>>>>> File "/usr/lib64/python2.7/httplib.py", line 797, in send >>>>>> self.connect() >>>>>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect >>>>>> sock = socket.create_connection((self.host, self.port), self.timeout) >>>>>> File "/usr/lib64/python2.7/socket.py", line 571, in create_connection >>>>>> raise err >>>>>> error: [Errno 101] Network is unreachable >>>>> >>>>> Yaniv, can you please take also a look to this one? >>>>> it's exactly the opposite of https://bugzilla.redhat.com/1358530 >>>>> Here the jsonrpcclient works but not the xmlrpc one. >>>>> >>>>>> 25.07.2016, 11:17, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>> On Mon, Jul 25, 2016 at 7:51 AM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>> Simone, there is something interesting in the vdsm.log? >>>>>>> >>>>>>> For what I saw the issue is not related to the storage but to the network. >>>>>>> ovirt-hosted-engine-setup uses the jsonrpc client, instead the code >>>>>>> from ovirt-hosted-engine-ha still uses the xmlrpc client somewhere and >>>>>>> this happens also when the setup asks to create the lockspace volume. >>>>>>> It seams that in your case the xmlrpc client could not connect vdsm on >>>>>>> the localhost. >>>>>>> It could be somehow related to: >>>>>>> https://bugzilla.redhat.com/1358530 >>>>>>> >>>>>>> Can you please try executing >>>>>>> sudo vdsClient -s 0 getVdsCaps >>>>>>> on that host? >>>>>>> >>>>>>>> 22.07.2016, 19:36, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>: >>>>>>>>> Simone, thanks for link. >>>>>>>>> vdsm.log attached >>>>>>>>> >>>>>>>>> 22.07.2016, 19:28, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>>>>> On Fri, Jul 22, 2016 at 5:59 PM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>>>>> Thank you for your response, Simone. >>>>>>>>>>> >>>>>>>>>>> Log attached. >>>>>>>>>> >>>>>>>>>> It seams it comes from VDSM, can you please attach also vdsm.log? >>>>>>>>>> >>>>>>>>>>> I don't use ovirt-engine-appliance because I have not found "how-to" for ovirt-engine-appliance deployment in hosted engine configuration. >>>>>>>>>> >>>>>>>>>> yum install ovirt-engine-appliance >>>>>>>>>> >>>>>>>>>> Then follow the instruction here: >>>>>>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/ >>>>>>>>>> >>>>>>>>>>> 22.07.2016, 17:09, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>>>>>>> Hi Aleksey, >>>>>>>>>>>> Can you please attach hosted-engine-setup logs? >>>>>>>>>>>> >>>>>>>>>>>> On Fri, Jul 22, 2016 at 3:46 PM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hello oVirt guru`s ! >>>>>>>>>>>>> >>>>>>>>>>>>> I have problem with initial deploy of ovirt 4.0 hosted engine. >>>>>>>>>>>>> >>>>>>>>>>>>> My environment : >>>>>>>>>>>>> ============================ >>>>>>>>>>>>> * Two servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected (with >>>>>>>>>>>>> multipathd) to storage HP 3PAR 7200 >>>>>>>>>>>>> * On each server installed CentOS 7.2 Linux (3.10.0-327.22.2.el7.x86_64) >>>>>>>>>>>>> * On 3PAR storage I created 2 LUNs for oVirt. >>>>>>>>>>>>> - First LUN for oVirt Hosted Engine VM (60GB) >>>>>>>>>>>>> - Second LUN for all other VMs (2TB) >>>>>>>>>>>>> >>>>>>>>>>>>> # multipath -ll >>>>>>>>>>>>> >>>>>>>>>>>>> 3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV >>>>>>>>>>>>> size=60G features='1 queue_if_no_path' hwhandler='1 alua' wp=rw >>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 status=active >>>>>>>>>>>>> |- 2:0:1:1 sdd 8:48 active ready running >>>>>>>>>>>>> |- 3:0:0:1 sdf 8:80 active ready running >>>>>>>>>>>>> |- 2:0:0:1 sdb 8:16 active ready running >>>>>>>>>>>>> `- 3:0:1:1 sdh 8:112 active ready running >>>>>>>>>>>>> >>>>>>>>>>>>> 3par-vv2 (360002ac000000000000000160000cec9) dm-1 3PARdata,VV >>>>>>>>>>>>> size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw >>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 status=active >>>>>>>>>>>>> |- 2:0:0:0 sda 8:0 active ready running >>>>>>>>>>>>> |- 3:0:0:0 sde 8:64 active ready running >>>>>>>>>>>>> |- 2:0:1:0 sdc 8:32 active ready running >>>>>>>>>>>>> `- 3:0:1:0 sdg 8:96 active ready running >>>>>>>>>>>>> >>>>>>>>>>>>> My steps on first server (initial deploy of ovirt 4.0 hosted engine): >>>>>>>>>>>>> ============================ >>>>>>>>>>>>> >>>>>>>>>>>>> # systemctl stop NetworkManager >>>>>>>>>>>>> # systemctl disable NetworkManager >>>>>>>>>>>>> # yum -y install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm >>>>>>>>>>>>> # yum -y install epel-release >>>>>>>>>>>>> # wget >>>>>>>>>>>>> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511... >>>>>>>>>>>>> -P /tmp/ >>>>>>>>>>>>> # yum install ovirt-hosted-engine-setup >>>>>>>>>>>>> # yum install screen >>>>>>>>>>>>> # screen -RD >>>>>>>>>>>>> >>>>>>>>>>>>> ...in screen session : >>>>>>>>>>>>> >>>>>>>>>>>>> # hosted-engine --deploy >>>>>>>>>>>>> >>>>>>>>>>>>> ... >>>>>>>>>>>>> in configuration process I chose "fc" as storage type for oVirt hosted >>>>>>>>>>>>> engine vm and select 60GB LUN... >>>>>>>>>>>>> ... >>>>>>>>>>>>> >>>>>>>>>>>>> --== CONFIGURATION PREVIEW ==-- >>>>>>>>>>>>> >>>>>>>>>>>>> ... >>>>>>>>>>>>> Firewall manager : iptables >>>>>>>>>>>>> Gateway address : 10.1.0.1 >>>>>>>>>>>>> Host name for web application : KOM-AD01-OVIRT1 >>>>>>>>>>>>> Storage Domain type : fc >>>>>>>>>>>>> Host ID : 1 >>>>>>>>>>>>> LUN ID : >>>>>>>>>>>>> 360002ac0000000000000001b0000cec9 >>>>>>>>>>>>> Image size GB : 40 >>>>>>>>>>>>> Console type : vnc >>>>>>>>>>>>> Memory size MB : 4096 >>>>>>>>>>>>> MAC address : 00:16:3e:77:1d:07 >>>>>>>>>>>>> Boot type : cdrom >>>>>>>>>>>>> Number of CPUs : 2 >>>>>>>>>>>>> ISO image (cdrom boot/cloud-init) : >>>>>>>>>>>>> /tmp/CentOS-7-x86_64-NetInstall-1511.iso >>>>>>>>>>>> >>>>>>>>>>>> Can I ask why you prefer/need to manually create a VM installing from >>>>>>>>>>>> a CD instead of using the ready-to-use ovirt-engine-appliance? >>>>>>>>>>>> Using the appliance makes the setup process a lot shorted and more comfortable. >>>>>>>>>>>> >>>>>>>>>>>>> CPU Type : model_Penryn >>>>>>>>>>>>> ... >>>>>>>>>>>>> and get error after step "Verifying sanlock lockspace initialization" >>>>>>>>>>>>> ... >>>>>>>>>>>>> >>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace initialization >>>>>>>>>>>>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network >>>>>>>>>>>>> is unreachable >>>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf' >>>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, >>>>>>>>>>>>> please check the issue, fix and redeploy >>>>>>>>>>>>> Log file is located at >>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log >>>>>>>>>>>>> >>>>>>>>>>>>> Interestingly >>>>>>>>>>>>> ============================ >>>>>>>>>>>>> If I try to deploy hosted-engine v3.6, everything goes well in the same >>>>>>>>>>>>> configuration !! : >>>>>>>>>>>>> >>>>>>>>>>>>> .... >>>>>>>>>>>>> [ INFO ] Stage: Transaction setup >>>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>>> [ INFO ] Stage: Package installation >>>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>>> [ INFO ] Configuring libvirt >>>>>>>>>>>>> [ INFO ] Configuring VDSM >>>>>>>>>>>>> [ INFO ] Starting vdsmd >>>>>>>>>>>>> [ INFO ] Waiting for VDSM hardware info >>>>>>>>>>>>> [ INFO ] Configuring the management bridge >>>>>>>>>>>>> [ INFO ] Creating Volume Group >>>>>>>>>>>>> [ INFO ] Creating Storage Domain >>>>>>>>>>>>> [ INFO ] Creating Storage Pool >>>>>>>>>>>>> [ INFO ] Connecting Storage Pool >>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace initialization >>>>>>>>>>>>> [ INFO ] Creating Image for 'hosted-engine.lockspace' ... >>>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.lockspace' created successfully >>>>>>>>>>>>> [ INFO ] Creating Image for 'hosted-engine.metadata' ... >>>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.metadata' created successfully >>>>>>>>>>>>> [ INFO ] Creating VM Image >>>>>>>>>>>>> [ INFO ] Destroying Storage Pool >>>>>>>>>>>>> [ INFO ] Start monitoring domain >>>>>>>>>>>>> [ INFO ] Configuring VM >>>>>>>>>>>>> [ INFO ] Updating hosted-engine configuration >>>>>>>>>>>>> [ INFO ] Stage: Transaction commit >>>>>>>>>>>>> [ INFO ] Stage: Closing up >>>>>>>>>>>>> [ INFO ] Creating VM >>>>>>>>>>>>> You can now connect to the VM with the following command: >>>>>>>>>>>>> /bin/remote-viewer vnc://localhost:5900 >>>>>>>>>>>>> ... >>>>>>>>>>>>> >>>>>>>>>>>>> What could be the problem? >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>> _______________________________________________ >>>> Users mailing list >>>> Users@ovirt.org >>>> http://lists.ovirt.org/mailman/listinfo/users > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users