
On Mon, Jul 25, 2016 at 2:38 PM, <aleksey.maksimov@it-kb.ru> wrote:
telnet kom-ad01-vm31.holding.com 54321 = success connection
telnet ::1 54321 Trying ::1... telnet: connect to address ::1: Network is unreachable
(ipv6 on my server disabled)
Ok, so the issue seams here: now by default vdsm binds on :: and its heuristc can end up using ipv6. See this one: https://bugzilla.redhat.com/show_bug.cgi?id=1350883 Can you please try enabling ipv6 on your host or setting management_ip = 0.0.0.0 under the [address] section in /etc/vdsm/vdsm.conf and then restarting vdsm.
25.07.2016, 15:35, "Simone Tiraboschi" <stirabos@redhat.com>:
On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksimov@it-kb.ru> wrote:
# ss -plutn
Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port
udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6)) udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6)) udp UNCONN 0 0 127.0.0.1:323 *:* users:(("chronyd",pid=795,fd=1)) udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7)) udp UNCONN 0 0 127.0.0.1:25375 *:* users:(("snmpd",pid=1609,fd=8)) udp UNCONN 0 0 127.0.0.1:25376 *:* users:(("cmapeerd",pid=2056,fd=5)) udp UNCONN 0 0 127.0.0.1:25393 *:* users:(("cmanicd",pid=2278,fd=3)) udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9)) udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10)) tcp LISTEN 0 128 *:2381 *:* users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4)) tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8)) tcp LISTEN 0 5 *:54322 *:* users:(("ovirt-imageio-d",pid=753,fd=3)) tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3)) tcp LISTEN 0 100 127.0.0.1:25 *:* users:(("master",pid=1948,fd=13)) tcp LISTEN 0 128 *:2301 *:* users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3)) tcp LISTEN 0 30 *:16514 *:* users:(("libvirtd",pid=10688,fd=13)) tcp LISTEN 0 128 127.0.0.1:199 *:* users:(("snmpd",pid=1609,fd=9)) tcp LISTEN 0 128 :::111 :::* users:(("rpcbind",pid=827,fd=11)) tcp LISTEN 0 5 :::54321 :::* users:(("vdsm",pid=11077,fd=23))
vdsm is properly bind over ipv6.
Can you please check if you can connect to vdsm with: telnet kom-ad01-vm31.holding.com 54321 and with telnet ::1 54321 ?
tcp LISTEN 0 30 :::16514 :::* users:(("libvirtd",pid=10688,fd=14))
25.07.2016, 15:11, "Simone Tiraboschi" <stirabos@redhat.com>:
On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov@it-kb.ru> wrote:
Yes.
# ping $(python -c 'import socket; print(socket.gethostname())')
PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data. 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 time=0.030 ms 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 time=0.015 ms 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 time=0.011 ms ^C --- KOM-AD01-VM31.holding.com ping statistics --- 3 packets transmitted, 3 received, 0% packet loss, time 2001ms rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
but...
and the output of ss -plutn
# vdsClient -s 0 getVdsCaps
Traceback (most recent call last): File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> code, message = commands[command][0](commandArgs) File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap return self.ExecAndExit(self.s.getVdsCapabilities()) File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ return self.__send(self.__name, args) File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request verbose=self.__verbose File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request return self.single_request(host, handler, request_body, verbose) File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request self.send_content(h, request_body) File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content connection.endheaders(request_body) File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders self._send_output(message_body) File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output self.send(msg) File "/usr/lib64/python2.7/httplib.py", line 797, in send self.connect() File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect sock = socket.create_connection((self.host, self.port), self.timeout) File "/usr/lib64/python2.7/socket.py", line 571, in create_connection raise err error: [Errno 101] Network is unreachable
25.07.2016, 14:58, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>:
Ok.
1) I stopped and disabled the service NetworkManager # systemctl stop NetworkManager # systemctl disable NetworkManager
2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in ifcfg-* file.
3) Reboot server
4) Try deploy oVirt HE 4 and I get the same error
[ INFO ] Creating Volume Group [ INFO ] Creating Storage Domain [ INFO ] Creating Storage Pool [ INFO ] Connecting Storage Pool [ INFO ] Verifying sanlock lockspace initialization [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network is unreachable [ INFO ] Stage: Clean up [ INFO ] Generating answer file '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf' [ INFO ] Stage: Pre-termination [ INFO ] Stage: Termination [ ERROR ] Hosted Engine deployment failed: this system is not reliable, please check the issue, fix and redeploy Log file is located at /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
What ideas further?
25.07.2016, 13:06, "Simone Tiraboschi" <stirabos@redhat.com>: > On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksimov@it-kb.ru> wrote: >> What am I supposed to do for successfully deploy ovirt 4 ? >> Any ideas ? > > Can you please try to explicitly configure your DNS with nameserver > under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for > the interface you are going to use? > >> 25.07.2016, 12:47, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>: >>> "Aleksey, was your DNS configured with DNS1 and DNS2 just on the interface you used to create the management bridge on?" >>> >>> Yes. Of course >>> >>> 25.07.2016, 12:27, "Simone Tiraboschi" <stirabos@redhat.com>: >>>> On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski >>>> <piotr.kliczewski@gmail.com> wrote: >>>>> This could be the issue here as well as for BZ #1358530 >>>>> >>>>> On Mon, Jul 25, 2016 at 10:53 AM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>> Could this be due to the fact that the ovirt installer has changed network configuration files (ifcfg-*, resolv.conf) ? >>>>>> After the error in ovirt installation process I see from resolv.conf disappeared on my DNS servers entry and now the server is unable to resolve names. >>>> >>>> So it could be related to https://bugzilla.redhat.com/show_bug.cgi?id=1160423 >>>> >>>> Aleksey, was your DNS configured with DNS1 and DNS2 just on the >>>> interface you used to create the management bridge on? >>>> Can you please try the workaround described here >>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ? >>>> >>>>>> 25.07.2016, 11:26, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>> On Mon, Jul 25, 2016 at 10:22 AM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>> # vdsClient -s 0 getVdsCaps >>>>>>>> >>>>>>>> Traceback (most recent call last): >>>>>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> >>>>>>>> code, message = commands[command][0](commandArgs) >>>>>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap >>>>>>>> return self.ExecAndExit(self.s.getVdsCapabilities()) >>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ >>>>>>>> return self.__send(self.__name, args) >>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request >>>>>>>> verbose=self.__verbose >>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request >>>>>>>> return self.single_request(host, handler, request_body, verbose) >>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request >>>>>>>> self.send_content(h, request_body) >>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content >>>>>>>> connection.endheaders(request_body) >>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders >>>>>>>> self._send_output(message_body) >>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output >>>>>>>> self.send(msg) >>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 797, in send >>>>>>>> self.connect() >>>>>>>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect >>>>>>>> sock = socket.create_connection((self.host, self.port), self.timeout) >>>>>>>> File "/usr/lib64/python2.7/socket.py", line 571, in create_connection >>>>>>>> raise err >>>>>>>> error: [Errno 101] Network is unreachable >>>>>>> >>>>>>> Yaniv, can you please take also a look to this one? >>>>>>> it's exactly the opposite of https://bugzilla.redhat.com/1358530 >>>>>>> Here the jsonrpcclient works but not the xmlrpc one. >>>>>>> >>>>>>>> 25.07.2016, 11:17, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>>>> On Mon, Jul 25, 2016 at 7:51 AM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>>>> Simone, there is something interesting in the vdsm.log? >>>>>>>>> >>>>>>>>> For what I saw the issue is not related to the storage but to the network. >>>>>>>>> ovirt-hosted-engine-setup uses the jsonrpc client, instead the code >>>>>>>>> from ovirt-hosted-engine-ha still uses the xmlrpc client somewhere and >>>>>>>>> this happens also when the setup asks to create the lockspace volume. >>>>>>>>> It seams that in your case the xmlrpc client could not connect vdsm on >>>>>>>>> the localhost. >>>>>>>>> It could be somehow related to: >>>>>>>>> https://bugzilla.redhat.com/1358530 >>>>>>>>> >>>>>>>>> Can you please try executing >>>>>>>>> sudo vdsClient -s 0 getVdsCaps >>>>>>>>> on that host? >>>>>>>>> >>>>>>>>>> 22.07.2016, 19:36, "aleksey.maksimov@it-kb.ru" <aleksey.maksimov@it-kb.ru>: >>>>>>>>>>> Simone, thanks for link. >>>>>>>>>>> vdsm.log attached >>>>>>>>>>> >>>>>>>>>>> 22.07.2016, 19:28, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>>>>>>> On Fri, Jul 22, 2016 at 5:59 PM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>>>>>>> Thank you for your response, Simone. >>>>>>>>>>>>> >>>>>>>>>>>>> Log attached. >>>>>>>>>>>> >>>>>>>>>>>> It seams it comes from VDSM, can you please attach also vdsm.log? >>>>>>>>>>>> >>>>>>>>>>>>> I don't use ovirt-engine-appliance because I have not found "how-to" for ovirt-engine-appliance deployment in hosted engine configuration. >>>>>>>>>>>> >>>>>>>>>>>> yum install ovirt-engine-appliance >>>>>>>>>>>> >>>>>>>>>>>> Then follow the instruction here: >>>>>>>>>>>> http://www.ovirt.org/develop/release-management/features/heapplianceflow/ >>>>>>>>>>>> >>>>>>>>>>>>> 22.07.2016, 17:09, "Simone Tiraboschi" <stirabos@redhat.com>: >>>>>>>>>>>>>> Hi Aleksey, >>>>>>>>>>>>>> Can you please attach hosted-engine-setup logs? >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Fri, Jul 22, 2016 at 3:46 PM, <aleksey.maksimov@it-kb.ru> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hello oVirt guru`s ! >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I have problem with initial deploy of ovirt 4.0 hosted engine. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> My environment : >>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>> * Two servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected (with >>>>>>>>>>>>>>> multipathd) to storage HP 3PAR 7200 >>>>>>>>>>>>>>> * On each server installed CentOS 7.2 Linux (3.10.0-327.22.2.el7.x86_64) >>>>>>>>>>>>>>> * On 3PAR storage I created 2 LUNs for oVirt. >>>>>>>>>>>>>>> - First LUN for oVirt Hosted Engine VM (60GB) >>>>>>>>>>>>>>> - Second LUN for all other VMs (2TB) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> # multipath -ll >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV >>>>>>>>>>>>>>> size=60G features='1 queue_if_no_path' hwhandler='1 alua' wp=rw >>>>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 status=active >>>>>>>>>>>>>>> |- 2:0:1:1 sdd 8:48 active ready running >>>>>>>>>>>>>>> |- 3:0:0:1 sdf 8:80 active ready running >>>>>>>>>>>>>>> |- 2:0:0:1 sdb 8:16 active ready running >>>>>>>>>>>>>>> `- 3:0:1:1 sdh 8:112 active ready running >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 3par-vv2 (360002ac000000000000000160000cec9) dm-1 3PARdata,VV >>>>>>>>>>>>>>> size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw >>>>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 status=active >>>>>>>>>>>>>>> |- 2:0:0:0 sda 8:0 active ready running >>>>>>>>>>>>>>> |- 3:0:0:0 sde 8:64 active ready running >>>>>>>>>>>>>>> |- 2:0:1:0 sdc 8:32 active ready running >>>>>>>>>>>>>>> `- 3:0:1:0 sdg 8:96 active ready running >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> My steps on first server (initial deploy of ovirt 4.0 hosted engine): >>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> # systemctl stop NetworkManager >>>>>>>>>>>>>>> # systemctl disable NetworkManager >>>>>>>>>>>>>>> # yum -y install http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm >>>>>>>>>>>>>>> # yum -y install epel-release >>>>>>>>>>>>>>> # wget >>>>>>>>>>>>>>> http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511... >>>>>>>>>>>>>>> -P /tmp/ >>>>>>>>>>>>>>> # yum install ovirt-hosted-engine-setup >>>>>>>>>>>>>>> # yum install screen >>>>>>>>>>>>>>> # screen -RD >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ...in screen session : >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> # hosted-engine --deploy >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>> in configuration process I chose "fc" as storage type for oVirt hosted >>>>>>>>>>>>>>> engine vm and select 60GB LUN... >>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> --== CONFIGURATION PREVIEW ==-- >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>> Firewall manager : iptables >>>>>>>>>>>>>>> Gateway address : 10.1.0.1 >>>>>>>>>>>>>>> Host name for web application : KOM-AD01-OVIRT1 >>>>>>>>>>>>>>> Storage Domain type : fc >>>>>>>>>>>>>>> Host ID : 1 >>>>>>>>>>>>>>> LUN ID : >>>>>>>>>>>>>>> 360002ac0000000000000001b0000cec9 >>>>>>>>>>>>>>> Image size GB : 40 >>>>>>>>>>>>>>> Console type : vnc >>>>>>>>>>>>>>> Memory size MB : 4096 >>>>>>>>>>>>>>> MAC address : 00:16:3e:77:1d:07 >>>>>>>>>>>>>>> Boot type : cdrom >>>>>>>>>>>>>>> Number of CPUs : 2 >>>>>>>>>>>>>>> ISO image (cdrom boot/cloud-init) : >>>>>>>>>>>>>>> /tmp/CentOS-7-x86_64-NetInstall-1511.iso >>>>>>>>>>>>>> >>>>>>>>>>>>>> Can I ask why you prefer/need to manually create a VM installing from >>>>>>>>>>>>>> a CD instead of using the ready-to-use ovirt-engine-appliance? >>>>>>>>>>>>>> Using the appliance makes the setup process a lot shorted and more comfortable. >>>>>>>>>>>>>> >>>>>>>>>>>>>>> CPU Type : model_Penryn >>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>> and get error after step "Verifying sanlock lockspace initialization" >>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace initialization >>>>>>>>>>>>>>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101] Network >>>>>>>>>>>>>>> is unreachable >>>>>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf' >>>>>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>>>>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, >>>>>>>>>>>>>>> please check the issue, fix and redeploy >>>>>>>>>>>>>>> Log file is located at >>>>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Interestingly >>>>>>>>>>>>>>> ============================ >>>>>>>>>>>>>>> If I try to deploy hosted-engine v3.6, everything goes well in the same >>>>>>>>>>>>>>> configuration !! : >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> .... >>>>>>>>>>>>>>> [ INFO ] Stage: Transaction setup >>>>>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>>>>> [ INFO ] Stage: Package installation >>>>>>>>>>>>>>> [ INFO ] Stage: Misc configuration >>>>>>>>>>>>>>> [ INFO ] Configuring libvirt >>>>>>>>>>>>>>> [ INFO ] Configuring VDSM >>>>>>>>>>>>>>> [ INFO ] Starting vdsmd >>>>>>>>>>>>>>> [ INFO ] Waiting for VDSM hardware info >>>>>>>>>>>>>>> [ INFO ] Configuring the management bridge >>>>>>>>>>>>>>> [ INFO ] Creating Volume Group >>>>>>>>>>>>>>> [ INFO ] Creating Storage Domain >>>>>>>>>>>>>>> [ INFO ] Creating Storage Pool >>>>>>>>>>>>>>> [ INFO ] Connecting Storage Pool >>>>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace initialization >>>>>>>>>>>>>>> [ INFO ] Creating Image for 'hosted-engine.lockspace' ... >>>>>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.lockspace' created successfully >>>>>>>>>>>>>>> [ INFO ] Creating Image for 'hosted-engine.metadata' ... >>>>>>>>>>>>>>> [ INFO ] Image for 'hosted-engine.metadata' created successfully >>>>>>>>>>>>>>> [ INFO ] Creating VM Image >>>>>>>>>>>>>>> [ INFO ] Destroying Storage Pool >>>>>>>>>>>>>>> [ INFO ] Start monitoring domain >>>>>>>>>>>>>>> [ INFO ] Configuring VM >>>>>>>>>>>>>>> [ INFO ] Updating hosted-engine configuration >>>>>>>>>>>>>>> [ INFO ] Stage: Transaction commit >>>>>>>>>>>>>>> [ INFO ] Stage: Closing up >>>>>>>>>>>>>>> [ INFO ] Creating VM >>>>>>>>>>>>>>> You can now connect to the VM with the following command: >>>>>>>>>>>>>>> /bin/remote-viewer vnc://localhost:5900 >>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> What could be the problem? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>> _______________________________________________ >>>>>> Users mailing list >>>>>> Users@ovirt.org >>>>>> http://lists.ovirt.org/mailman/listinfo/users >>> >>> _______________________________________________ >>> Users mailing list >>> Users@ovirt.org >>> http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users