# ss -plutn
Netid State Recv-Q Send-Q
Local Address:Port Peer Address:Port
udp UNCONN 0 0
*:111
*:*
users:(("rpcbind",pid=827,fd=6))
udp UNCONN 0 0
*:161
*:*
users:(("snmpd",pid=1609,fd=6))
udp UNCONN 0 0
127.0.0.1:323
*:*
users:(("chronyd",pid=795,fd=1))
udp UNCONN 0 0
*:959
*:*
users:(("rpcbind",pid=827,fd=7))
udp UNCONN 0 0
127.0.0.1:25375
*:*
users:(("snmpd",pid=1609,fd=8))
udp UNCONN 0 0
127.0.0.1:25376
*:*
users:(("cmapeerd",pid=2056,fd=5))
udp UNCONN 0 0
127.0.0.1:25393
*:*
users:(("cmanicd",pid=2278,fd=3))
udp UNCONN 0 0
:::111
:::*
users:(("rpcbind",pid=827,fd=9))
udp UNCONN 0 0
:::959
:::*
users:(("rpcbind",pid=827,fd=10))
tcp LISTEN 0 128
*:2381
*:*
users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4))
tcp LISTEN 0 128
*:111
*:*
users:(("rpcbind",pid=827,fd=8))
tcp LISTEN 0 5
*:54322
*:*
users:(("ovirt-imageio-d",pid=753,fd=3))
tcp LISTEN 0 128
*:22
*:*
users:(("sshd",pid=1606,fd=3))
tcp LISTEN 0 100
127.0.0.1:25
*:*
users:(("master",pid=1948,fd=13))
tcp LISTEN 0 128
*:2301
*:*
users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
tcp LISTEN 0 30
*:16514
*:*
users:(("libvirtd",pid=10688,fd=13))
tcp LISTEN 0 128
127.0.0.1:199
*:*
users:(("snmpd",pid=1609,fd=9))
tcp LISTEN 0 128
:::111
:::*
users:(("rpcbind",pid=827,fd=11))
tcp LISTEN 0 5
:::54321
:::*
users:(("vdsm",pid=11077,fd=23))
tcp LISTEN 0 30
:::16514
:::*
users:(("libvirtd",pid=10688,fd=14))
25.07.2016, 15:11, "Simone Tiraboschi" <stirabos(a)redhat.com>:
On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov(a)it-kb.ru>
wrote:
> Yes.
>
> # ping $(python -c 'import socket; print(socket.gethostname())')
>
> PING
KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64 time=0.030
ms
> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64 time=0.015
ms
> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64 time=0.011
ms
> ^C
> ---
KOM-AD01-VM31.holding.com ping statistics ---
> 3 packets transmitted, 3 received, 0% packet loss, time 2001ms
> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
>
> but...
and the output of
ss -plutn
> # vdsClient -s 0 getVdsCaps
>
> Traceback (most recent call last):
> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
> code, message = commands[command][0](commandArgs)
> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
> return self.ExecAndExit(self.s.getVdsCapabilities())
> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
> return self.__send(self.__name, args)
> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
> verbose=self.__verbose
> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
> return self.single_request(host, handler, request_body, verbose)
> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
> self.send_content(h, request_body)
> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
> connection.endheaders(request_body)
> File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
> self._send_output(message_body)
> File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
> self.send(msg)
> File "/usr/lib64/python2.7/httplib.py", line 797, in send
> self.connect()
> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in
connect
> sock = socket.create_connection((self.host, self.port), self.timeout)
> File "/usr/lib64/python2.7/socket.py", line 571, in create_connection
> raise err
> error: [Errno 101] Network is unreachable
>
> 25.07.2016, 14:58, "aleksey.maksimov(a)it-kb.ru"
<aleksey.maksimov(a)it-kb.ru>:
>> Ok.
>>
>> 1) I stopped and disabled the service NetworkManager
>> # systemctl stop NetworkManager
>> # systemctl disable NetworkManager
>>
>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in ifcfg-*
file.
>>
>> 3) Reboot server
>>
>> 4) Try deploy oVirt HE 4 and I get the same error
>>
>> [ INFO ] Creating Volume Group
>> [ INFO ] Creating Storage Domain
>> [ INFO ] Creating Storage Pool
>> [ INFO ] Connecting Storage Pool
>> [ INFO ] Verifying sanlock lockspace initialization
>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101]
Network is unreachable
>> [ INFO ] Stage: Clean up
>> [ INFO ] Generating answer file
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
>> [ INFO ] Stage: Pre-termination
>> [ INFO ] Stage: Termination
>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable, please
check the issue, fix and redeploy
>> Log file is located at
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
>>
>> What ideas further?
>>
>> 25.07.2016, 13:06, "Simone Tiraboschi" <stirabos(a)redhat.com>:
>>> On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>> What am I supposed to do for successfully deploy ovirt 4 ?
>>>> Any ideas ?
>>>
>>> Can you please try to explicitly configure your DNS with nameserver
>>> under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
>>> the interface you are going to use?
>>>
>>>> 25.07.2016, 12:47, "aleksey.maksimov(a)it-kb.ru"
<aleksey.maksimov(a)it-kb.ru>:
>>>>> "Aleksey, was your DNS configured with DNS1 and DNS2 just on
the interface you used to create the management bridge on?"
>>>>>
>>>>> Yes. Of course
>>>>>
>>>>> 25.07.2016, 12:27, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
>>>>>> On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>>>> <piotr.kliczewski(a)gmail.com> wrote:
>>>>>>> This could be the issue here as well as for BZ #1358530
>>>>>>>
>>>>>>> On Mon, Jul 25, 2016 at 10:53 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>> Could this be due to the fact that the ovirt
installer has changed network configuration files (ifcfg-*, resolv.conf) ?
>>>>>>>> After the error in ovirt installation process I see
from resolv.conf disappeared on my DNS servers entry and now the server is unable to
resolve names.
>>>>>>
>>>>>> So it could be related to
https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>>>>
>>>>>> Aleksey, was your DNS configured with DNS1 and DNS2 just on
the
>>>>>> interface you used to create the management bridge on?
>>>>>> Can you please try the workaround described here
>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ?
>>>>>>
>>>>>>>> 25.07.2016, 11:26, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
>>>>>>>>> On Mon, Jul 25, 2016 at 10:22 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>> # vdsClient -s 0 getVdsCaps
>>>>>>>>>>
>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>> File
"/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>>>>>>>>> code, message =
commands[command][0](commandArgs)
>>>>>>>>>> File
"/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>>>> return
self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>>>>>>>> return self.__send(self.__name, args)
>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>>>>>>>> verbose=self.__verbose
>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>>>>>>>> return self.single_request(host,
handler, request_body, verbose)
>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>>>>>>>>>> self.send_content(h, request_body)
>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>>>>>>>>>> connection.endheaders(request_body)
>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>>>>>>>> self._send_output(message_body)
>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>>>>>>>> self.send(msg)
>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>>>>>>> self.connect()
>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect
>>>>>>>>>> sock =
socket.create_connection((self.host, self.port), self.timeout)
>>>>>>>>>> File
"/usr/lib64/python2.7/socket.py", line 571, in create_connection
>>>>>>>>>> raise err
>>>>>>>>>> error: [Errno 101] Network is unreachable
>>>>>>>>>
>>>>>>>>> Yaniv, can you please take also a look to this
one?
>>>>>>>>> it's exactly the opposite of
https://bugzilla.redhat.com/1358530
>>>>>>>>> Here the jsonrpcclient works but not the xmlrpc
one.
>>>>>>>>>
>>>>>>>>>> 25.07.2016, 11:17, "Simone
Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>> On Mon, Jul 25, 2016 at 7:51 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>> Simone, there is something
interesting in the vdsm.log?
>>>>>>>>>>>
>>>>>>>>>>> For what I saw the issue is not related
to the storage but to the network.
>>>>>>>>>>> ovirt-hosted-engine-setup uses the
jsonrpc client, instead the code
>>>>>>>>>>> from ovirt-hosted-engine-ha still uses
the xmlrpc client somewhere and
>>>>>>>>>>> this happens also when the setup asks
to create the lockspace volume.
>>>>>>>>>>> It seams that in your case the xmlrpc
client could not connect vdsm on
>>>>>>>>>>> the localhost.
>>>>>>>>>>> It could be somehow related to:
>>>>>>>>>>> https://bugzilla.redhat.com/1358530
>>>>>>>>>>>
>>>>>>>>>>> Can you please try executing
>>>>>>>>>>> sudo vdsClient -s 0 getVdsCaps
>>>>>>>>>>> on that host?
>>>>>>>>>>>
>>>>>>>>>>>> 22.07.2016, 19:36,
"aleksey.maksimov(a)it-kb.ru" <aleksey.maksimov(a)it-kb.ru>:
>>>>>>>>>>>>> Simone, thanks for link.
>>>>>>>>>>>>> vdsm.log attached
>>>>>>>>>>>>>
>>>>>>>>>>>>> 22.07.2016, 19:28,
"Simone Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>>>> On Fri, Jul 22, 2016 at
5:59 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>>>> Thank you for your
response, Simone.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Log attached.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> It seams it comes from
VDSM, can you please attach also vdsm.log?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I don't use
ovirt-engine-appliance because I have not found "how-to" for
ovirt-engine-appliance deployment in hosted engine configuration.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> yum install
ovirt-engine-appliance
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Then follow the
instruction here:
>>>>>>>>>>>>>>
http://www.ovirt.org/develop/release-management/features/heapplianceflow/
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> 22.07.2016, 17:09,
"Simone Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>>>>>> Hi Aleksey,
>>>>>>>>>>>>>>>> Can you please
attach hosted-engine-setup logs?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Fri, Jul 22,
2016 at 3:46 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hello oVirt
guru`s !
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> I have
problem with initial deploy of ovirt 4.0 hosted engine.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> My
environment :
>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>> * Two
servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected (with
>>>>>>>>>>>>>>>>> multipathd)
to storage HP 3PAR 7200
>>>>>>>>>>>>>>>>> * On each
server installed CentOS 7.2 Linux (3.10.0-327.22.2.el7.x86_64)
>>>>>>>>>>>>>>>>> * On 3PAR
storage I created 2 LUNs for oVirt.
>>>>>>>>>>>>>>>>> - First LUN
for oVirt Hosted Engine VM (60GB)
>>>>>>>>>>>>>>>>> - Second
LUN for all other VMs (2TB)
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> # multipath
-ll
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> 3par-vv1
(360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV
>>>>>>>>>>>>>>>>> size=60G
features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>> `-+-
policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>> |-
2:0:1:1 sdd 8:48 active ready running
>>>>>>>>>>>>>>>>> |-
3:0:0:1 sdf 8:80 active ready running
>>>>>>>>>>>>>>>>> |-
2:0:0:1 sdb 8:16 active ready running
>>>>>>>>>>>>>>>>> `-
3:0:1:1 sdh 8:112 active ready running
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> 3par-vv2
(360002ac000000000000000160000cec9) dm-1 3PARdata,VV
>>>>>>>>>>>>>>>>> size=2.0T
features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>> `-+-
policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>> |-
2:0:0:0 sda 8:0 active ready running
>>>>>>>>>>>>>>>>> |-
3:0:0:0 sde 8:64 active ready running
>>>>>>>>>>>>>>>>> |-
2:0:1:0 sdc 8:32 active ready running
>>>>>>>>>>>>>>>>> `-
3:0:1:0 sdg 8:96 active ready running
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> My steps on
first server (initial deploy of ovirt 4.0 hosted engine):
>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> # systemctl
stop NetworkManager
>>>>>>>>>>>>>>>>> # systemctl
disable NetworkManager
>>>>>>>>>>>>>>>>> # yum -y
install
http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
>>>>>>>>>>>>>>>>> # yum -y
install epel-release
>>>>>>>>>>>>>>>>> # wget
>>>>>>>>>>>>>>>>>
http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>> -P /tmp/
>>>>>>>>>>>>>>>>> # yum
install ovirt-hosted-engine-setup
>>>>>>>>>>>>>>>>> # yum
install screen
>>>>>>>>>>>>>>>>> # screen
-RD
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> ...in
screen session :
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> #
hosted-engine --deploy
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>> in
configuration process I chose "fc" as storage type for oVirt hosted
>>>>>>>>>>>>>>>>> engine vm
and select 60GB LUN...
>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> --==
CONFIGURATION PREVIEW ==--
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>
Firewall manager : iptables
>>>>>>>>>>>>>>>>>
Gateway address : 10.1.0.1
>>>>>>>>>>>>>>>>>
Host name for web application : KOM-AD01-OVIRT1
>>>>>>>>>>>>>>>>>
Storage Domain type : fc
>>>>>>>>>>>>>>>>>
Host ID : 1
>>>>>>>>>>>>>>>>>
LUN ID :
>>>>>>>>>>>>>>>>>
360002ac0000000000000001b0000cec9
>>>>>>>>>>>>>>>>>
Image size GB : 40
>>>>>>>>>>>>>>>>>
Console type : vnc
>>>>>>>>>>>>>>>>>
Memory size MB : 4096
>>>>>>>>>>>>>>>>>
MAC address : 00:16:3e:77:1d:07
>>>>>>>>>>>>>>>>>
Boot type : cdrom
>>>>>>>>>>>>>>>>>
Number of CPUs : 2
>>>>>>>>>>>>>>>>>
ISO image (cdrom boot/cloud-init) :
>>>>>>>>>>>>>>>>>
/tmp/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Can I ask why
you prefer/need to manually create a VM installing from
>>>>>>>>>>>>>>>> a CD instead of
using the ready-to-use ovirt-engine-appliance?
>>>>>>>>>>>>>>>> Using the
appliance makes the setup process a lot shorted and more comfortable.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
CPU Type : model_Penryn
>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>> and get
error after step "Verifying sanlock lockspace initialization"
>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> [ INFO ]
Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>> [ ERROR ]
Failed to execute stage 'Misc configuration': [Errno 101] Network
>>>>>>>>>>>>>>>>> is
unreachable
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Clean up
>>>>>>>>>>>>>>>>> [ INFO ]
Generating answer file
>>>>>>>>>>>>>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Pre-termination
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Termination
>>>>>>>>>>>>>>>>> [ ERROR ]
Hosted Engine deployment failed: this system is not reliable,
>>>>>>>>>>>>>>>>> please
check the issue, fix and redeploy
>>>>>>>>>>>>>>>>>
Log file is located at
>>>>>>>>>>>>>>>>>
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
Interestingly
>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>> If I try to
deploy hosted-engine v3.6, everything goes well in the same
>>>>>>>>>>>>>>>>>
configuration !! :
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> ....
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Transaction setup
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Misc configuration
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Package installation
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Misc configuration
>>>>>>>>>>>>>>>>> [ INFO ]
Configuring libvirt
>>>>>>>>>>>>>>>>> [ INFO ]
Configuring VDSM
>>>>>>>>>>>>>>>>> [ INFO ]
Starting vdsmd
>>>>>>>>>>>>>>>>> [ INFO ]
Waiting for VDSM hardware info
>>>>>>>>>>>>>>>>> [ INFO ]
Configuring the management bridge
>>>>>>>>>>>>>>>>> [ INFO ]
Creating Volume Group
>>>>>>>>>>>>>>>>> [ INFO ]
Creating Storage Domain
>>>>>>>>>>>>>>>>> [ INFO ]
Creating Storage Pool
>>>>>>>>>>>>>>>>> [ INFO ]
Connecting Storage Pool
>>>>>>>>>>>>>>>>> [ INFO ]
Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>> [ INFO ]
Creating Image for 'hosted-engine.lockspace' ...
>>>>>>>>>>>>>>>>> [ INFO ]
Image for 'hosted-engine.lockspace' created successfully
>>>>>>>>>>>>>>>>> [ INFO ]
Creating Image for 'hosted-engine.metadata' ...
>>>>>>>>>>>>>>>>> [ INFO ]
Image for 'hosted-engine.metadata' created successfully
>>>>>>>>>>>>>>>>> [ INFO ]
Creating VM Image
>>>>>>>>>>>>>>>>> [ INFO ]
Destroying Storage Pool
>>>>>>>>>>>>>>>>> [ INFO ]
Start monitoring domain
>>>>>>>>>>>>>>>>> [ INFO ]
Configuring VM
>>>>>>>>>>>>>>>>> [ INFO ]
Updating hosted-engine configuration
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Transaction commit
>>>>>>>>>>>>>>>>> [ INFO ]
Stage: Closing up
>>>>>>>>>>>>>>>>> [ INFO ]
Creating VM
>>>>>>>>>>>>>>>>>
You can now connect to the VM with the following command:
>>>>>>>>>>>>>>>>>
/bin/remote-viewer vnc://localhost:5900
>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> What could
be the problem?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>> Users
mailing list
>>>>>>>>>>>>>>>>>
Users(a)ovirt.org
>>>>>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list
>>>>>>>> Users(a)ovirt.org
>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> Users(a)ovirt.org
>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>
>> _______________________________________________
>> Users mailing list
>> Users(a)ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users