# ss -plutn
Netid State Recv-Q Send-Q
Local Address:Port Peer Address:Port
udp UNCONN 0 0
*:111
*:*
users:(("rpcbind",pid=827,fd=6))
udp UNCONN 0 0
*:161
*:*
users:(("snmpd",pid=1609,fd=6))
udp UNCONN 0 0
127.0.0.1:323
*:*
users:(("chronyd",pid=795,fd=1))
udp UNCONN 0 0
*:959
*:*
users:(("rpcbind",pid=827,fd=7))
udp UNCONN 0 0
127.0.0.1:25375
*:*
users:(("snmpd",pid=1609,fd=8))
udp UNCONN 0 0
127.0.0.1:25376
*:*
users:(("cmapeerd",pid=2056,fd=5))
udp UNCONN 0 0
127.0.0.1:25393
*:*
users:(("cmanicd",pid=2278,fd=3))
udp UNCONN 0 0
:::111
:::*
users:(("rpcbind",pid=827,fd=9))
udp UNCONN 0 0
:::959
:::*
users:(("rpcbind",pid=827,fd=10))
tcp LISTEN 0 128
*:2381
*:*
users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4))
tcp LISTEN 0 128
*:111
*:*
users:(("rpcbind",pid=827,fd=8))
tcp LISTEN 0 5
*:54322
*:*
users:(("ovirt-imageio-d",pid=753,fd=3))
tcp LISTEN 0 128
*:22
*:*
users:(("sshd",pid=1606,fd=3))
tcp LISTEN 0 100
127.0.0.1:25
*:*
users:(("master",pid=1948,fd=13))
tcp LISTEN 0 128
*:2301
*:*
users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3))
tcp LISTEN 0 30
*:16514
*:*
users:(("libvirtd",pid=10688,fd=13))
tcp LISTEN 0 128
127.0.0.1:199
*:*
users:(("snmpd",pid=1609,fd=9))
tcp LISTEN 0 128
:::111
:::*
users:(("rpcbind",pid=827,fd=11))
tcp LISTEN 0 5
:::54321
:::*
users:(("vdsm",pid=11077,fd=23))
vdsm is properly bind over ipv6.
Can you please check if you can connect to vdsm with:
telnet
tcp LISTEN 0 30
:::16514
:::*
users:(("libvirtd",pid=10688,fd=14))
25.07.2016, 15:11, "Simone Tiraboschi" <stirabos(a)redhat.com>:
> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>> Yes.
>>
>> # ping $(python -c 'import socket; print(socket.gethostname())')
>>
>> PING
KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data.
>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 ttl=64
time=0.030 ms
>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 ttl=64
time=0.015 ms
>> 64 bytes from
kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 ttl=64
time=0.011 ms
>> ^C
>> ---
KOM-AD01-VM31.holding.com ping statistics ---
>> 3 packets transmitted, 3 received, 0% packet loss, time 2001ms
>> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms
>>
>> but...
>
> and the output of
> ss -plutn
>
>> # vdsClient -s 0 getVdsCaps
>>
>> Traceback (most recent call last):
>> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>> code, message = commands[command][0](commandArgs)
>> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>> return self.ExecAndExit(self.s.getVdsCapabilities())
>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>> return self.__send(self.__name, args)
>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>> verbose=self.__verbose
>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>> return self.single_request(host, handler, request_body, verbose)
>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in
single_request
>> self.send_content(h, request_body)
>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in
send_content
>> connection.endheaders(request_body)
>> File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>> self._send_output(message_body)
>> File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>> self.send(msg)
>> File "/usr/lib64/python2.7/httplib.py", line 797, in send
>> self.connect()
>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203,
in connect
>> sock = socket.create_connection((self.host, self.port), self.timeout)
>> File "/usr/lib64/python2.7/socket.py", line 571, in
create_connection
>> raise err
>> error: [Errno 101] Network is unreachable
>>
>> 25.07.2016, 14:58, "aleksey.maksimov(a)it-kb.ru"
<aleksey.maksimov(a)it-kb.ru>:
>>> Ok.
>>>
>>> 1) I stopped and disabled the service NetworkManager
>>> # systemctl stop NetworkManager
>>> # systemctl disable NetworkManager
>>>
>>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add PEERDNS=no in
ifcfg-* file.
>>>
>>> 3) Reboot server
>>>
>>> 4) Try deploy oVirt HE 4 and I get the same error
>>>
>>> [ INFO ] Creating Volume Group
>>> [ INFO ] Creating Storage Domain
>>> [ INFO ] Creating Storage Pool
>>> [ INFO ] Connecting Storage Pool
>>> [ INFO ] Verifying sanlock lockspace initialization
>>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno 101]
Network is unreachable
>>> [ INFO ] Stage: Clean up
>>> [ INFO ] Generating answer file
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf'
>>> [ INFO ] Stage: Pre-termination
>>> [ INFO ] Stage: Termination
>>> [ ERROR ] Hosted Engine deployment failed: this system is not reliable,
please check the issue, fix and redeploy
>>> Log file is located at
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160725142534-t81kwf.log
>>>
>>> What ideas further?
>>>
>>> 25.07.2016, 13:06, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
>>>> On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksimov(a)it-kb.ru>
wrote:
>>>>> What am I supposed to do for successfully deploy ovirt 4 ?
>>>>> Any ideas ?
>>>>
>>>> Can you please try to explicitly configure your DNS with nameserver
>>>> under /etc/resolv.conf and remove DNS1 and DNS2 and set PEERDNS=no for
>>>> the interface you are going to use?
>>>>
>>>>> 25.07.2016, 12:47, "aleksey.maksimov(a)it-kb.ru"
<aleksey.maksimov(a)it-kb.ru>:
>>>>>> "Aleksey, was your DNS configured with DNS1 and DNS2 just
on the interface you used to create the management bridge on?"
>>>>>>
>>>>>> Yes. Of course
>>>>>>
>>>>>> 25.07.2016, 12:27, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
>>>>>>> On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski
>>>>>>> <piotr.kliczewski(a)gmail.com> wrote:
>>>>>>>> This could be the issue here as well as for BZ
#1358530
>>>>>>>>
>>>>>>>> On Mon, Jul 25, 2016 at 10:53 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>> Could this be due to the fact that the ovirt
installer has changed network configuration files (ifcfg-*, resolv.conf) ?
>>>>>>>>> After the error in ovirt installation process I
see from resolv.conf disappeared on my DNS servers entry and now the server is unable to
resolve names.
>>>>>>>
>>>>>>> So it could be related to
https://bugzilla.redhat.com/show_bug.cgi?id=1160423
>>>>>>>
>>>>>>> Aleksey, was your DNS configured with DNS1 and DNS2 just
on the
>>>>>>> interface you used to create the management bridge on?
>>>>>>> Can you please try the workaround described here
>>>>>>>
https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25
?
>>>>>>>
>>>>>>>>> 25.07.2016, 11:26, "Simone Tiraboschi"
<stirabos(a)redhat.com>:
>>>>>>>>>> On Mon, Jul 25, 2016 at 10:22 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>> # vdsClient -s 0 getVdsCaps
>>>>>>>>>>>
>>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>>> File
"/usr/share/vdsm/vdsClient.py", line 2980, in <module>
>>>>>>>>>>> code, message =
commands[command][0](commandArgs)
>>>>>>>>>>> File
"/usr/share/vdsm/vdsClient.py", line 543, in do_getCap
>>>>>>>>>>> return
self.ExecAndExit(self.s.getVdsCapabilities())
>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__
>>>>>>>>>>> return self.__send(self.__name,
args)
>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request
>>>>>>>>>>> verbose=self.__verbose
>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request
>>>>>>>>>>> return self.single_request(host,
handler, request_body, verbose)
>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1301, in single_request
>>>>>>>>>>> self.send_content(h, request_body)
>>>>>>>>>>> File
"/usr/lib64/python2.7/xmlrpclib.py", line 1448, in send_content
>>>>>>>>>>>
connection.endheaders(request_body)
>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 975, in endheaders
>>>>>>>>>>> self._send_output(message_body)
>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 835, in _send_output
>>>>>>>>>>> self.send(msg)
>>>>>>>>>>> File
"/usr/lib64/python2.7/httplib.py", line 797, in send
>>>>>>>>>>> self.connect()
>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect
>>>>>>>>>>> sock =
socket.create_connection((self.host, self.port), self.timeout)
>>>>>>>>>>> File
"/usr/lib64/python2.7/socket.py", line 571, in create_connection
>>>>>>>>>>> raise err
>>>>>>>>>>> error: [Errno 101] Network is
unreachable
>>>>>>>>>>
>>>>>>>>>> Yaniv, can you please take also a look to
this one?
>>>>>>>>>> it's exactly the opposite of
https://bugzilla.redhat.com/1358530
>>>>>>>>>> Here the jsonrpcclient works but not the
xmlrpc one.
>>>>>>>>>>
>>>>>>>>>>> 25.07.2016, 11:17, "Simone
Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>> On Mon, Jul 25, 2016 at 7:51 AM,
<aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>> Simone, there is something
interesting in the vdsm.log?
>>>>>>>>>>>>
>>>>>>>>>>>> For what I saw the issue is not
related to the storage but to the network.
>>>>>>>>>>>> ovirt-hosted-engine-setup uses the
jsonrpc client, instead the code
>>>>>>>>>>>> from ovirt-hosted-engine-ha still
uses the xmlrpc client somewhere and
>>>>>>>>>>>> this happens also when the setup
asks to create the lockspace volume.
>>>>>>>>>>>> It seams that in your case the
xmlrpc client could not connect vdsm on
>>>>>>>>>>>> the localhost.
>>>>>>>>>>>> It could be somehow related to:
>>>>>>>>>>>>
https://bugzilla.redhat.com/1358530
>>>>>>>>>>>>
>>>>>>>>>>>> Can you please try executing
>>>>>>>>>>>> sudo vdsClient -s 0 getVdsCaps
>>>>>>>>>>>> on that host?
>>>>>>>>>>>>
>>>>>>>>>>>>> 22.07.2016, 19:36,
"aleksey.maksimov(a)it-kb.ru" <aleksey.maksimov(a)it-kb.ru>:
>>>>>>>>>>>>>> Simone, thanks for link.
>>>>>>>>>>>>>> vdsm.log attached
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> 22.07.2016, 19:28,
"Simone Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>>>>> On Fri, Jul 22, 2016
at 5:59 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>>>>> Thank you for
your response, Simone.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Log attached.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> It seams it comes
from VDSM, can you please attach also vdsm.log?
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I don't use
ovirt-engine-appliance because I have not found "how-to" for
ovirt-engine-appliance deployment in hosted engine configuration.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> yum install
ovirt-engine-appliance
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Then follow the
instruction here:
>>>>>>>>>>>>>>>
http://www.ovirt.org/develop/release-management/features/heapplianceflow/
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 22.07.2016,
17:09, "Simone Tiraboschi" <stirabos(a)redhat.com>:
>>>>>>>>>>>>>>>>> Hi Aleksey,
>>>>>>>>>>>>>>>>> Can you
please attach hosted-engine-setup logs?
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Fri, Jul
22, 2016 at 3:46 PM, <aleksey.maksimov(a)it-kb.ru> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hello
oVirt guru`s !
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> I have
problem with initial deploy of ovirt 4.0 hosted engine.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> My
environment :
>>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>>> * Two
servers HP ProLiant DL 360 G5 with Qlogic FC HBA connected (with
>>>>>>>>>>>>>>>>>>
multipathd) to storage HP 3PAR 7200
>>>>>>>>>>>>>>>>>> * On
each server installed CentOS 7.2 Linux (3.10.0-327.22.2.el7.x86_64)
>>>>>>>>>>>>>>>>>> * On
3PAR storage I created 2 LUNs for oVirt.
>>>>>>>>>>>>>>>>>> - First
LUN for oVirt Hosted Engine VM (60GB)
>>>>>>>>>>>>>>>>>> -
Second LUN for all other VMs (2TB)
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> #
multipath -ll
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
3par-vv1 (360002ac0000000000000001b0000cec9) dm-0 3PARdata,VV
>>>>>>>>>>>>>>>>>>
size=60G features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>> `-+-
policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>> |-
2:0:1:1 sdd 8:48 active ready running
>>>>>>>>>>>>>>>>>> |-
3:0:0:1 sdf 8:80 active ready running
>>>>>>>>>>>>>>>>>> |-
2:0:0:1 sdb 8:16 active ready running
>>>>>>>>>>>>>>>>>> `-
3:0:1:1 sdh 8:112 active ready running
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
3par-vv2 (360002ac000000000000000160000cec9) dm-1 3PARdata,VV
>>>>>>>>>>>>>>>>>>
size=2.0T features='1 queue_if_no_path' hwhandler='1 alua' wp=rw
>>>>>>>>>>>>>>>>>> `-+-
policy='round-robin 0' prio=50 status=active
>>>>>>>>>>>>>>>>>> |-
2:0:0:0 sda 8:0 active ready running
>>>>>>>>>>>>>>>>>> |-
3:0:0:0 sde 8:64 active ready running
>>>>>>>>>>>>>>>>>> |-
2:0:1:0 sdc 8:32 active ready running
>>>>>>>>>>>>>>>>>> `-
3:0:1:0 sdg 8:96 active ready running
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> My
steps on first server (initial deploy of ovirt 4.0 hosted engine):
>>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> #
systemctl stop NetworkManager
>>>>>>>>>>>>>>>>>> #
systemctl disable NetworkManager
>>>>>>>>>>>>>>>>>> # yum
-y install
http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm
>>>>>>>>>>>>>>>>>> # yum
-y install epel-release
>>>>>>>>>>>>>>>>>> # wget
>>>>>>>>>>>>>>>>>>
http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1...
>>>>>>>>>>>>>>>>>> -P
/tmp/
>>>>>>>>>>>>>>>>>> # yum
install ovirt-hosted-engine-setup
>>>>>>>>>>>>>>>>>> # yum
install screen
>>>>>>>>>>>>>>>>>> #
screen -RD
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> ...in
screen session :
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> #
hosted-engine --deploy
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>> in
configuration process I chose "fc" as storage type for oVirt hosted
>>>>>>>>>>>>>>>>>> engine
vm and select 60GB LUN...
>>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> --==
CONFIGURATION PREVIEW ==--
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>>
Firewall manager : iptables
>>>>>>>>>>>>>>>>>>
Gateway address : 10.1.0.1
>>>>>>>>>>>>>>>>>>
Host name for web application : KOM-AD01-OVIRT1
>>>>>>>>>>>>>>>>>>
Storage Domain type : fc
>>>>>>>>>>>>>>>>>>
Host ID : 1
>>>>>>>>>>>>>>>>>>
LUN ID :
>>>>>>>>>>>>>>>>>>
360002ac0000000000000001b0000cec9
>>>>>>>>>>>>>>>>>>
Image size GB : 40
>>>>>>>>>>>>>>>>>>
Console type : vnc
>>>>>>>>>>>>>>>>>>
Memory size MB : 4096
>>>>>>>>>>>>>>>>>>
MAC address : 00:16:3e:77:1d:07
>>>>>>>>>>>>>>>>>>
Boot type : cdrom
>>>>>>>>>>>>>>>>>>
Number of CPUs : 2
>>>>>>>>>>>>>>>>>>
ISO image (cdrom boot/cloud-init) :
>>>>>>>>>>>>>>>>>>
/tmp/CentOS-7-x86_64-NetInstall-1511.iso
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Can I ask
why you prefer/need to manually create a VM installing from
>>>>>>>>>>>>>>>>> a CD instead
of using the ready-to-use ovirt-engine-appliance?
>>>>>>>>>>>>>>>>> Using the
appliance makes the setup process a lot shorted and more comfortable.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
CPU Type : model_Penryn
>>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>> and get
error after step "Verifying sanlock lockspace initialization"
>>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> [ INFO
] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>>> [ ERROR
] Failed to execute stage 'Misc configuration': [Errno 101] Network
>>>>>>>>>>>>>>>>>> is
unreachable
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Clean up
>>>>>>>>>>>>>>>>>> [ INFO
] Generating answer file
>>>>>>>>>>>>>>>>>>
'/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf'
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Pre-termination
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Termination
>>>>>>>>>>>>>>>>>> [ ERROR
] Hosted Engine deployment failed: this system is not reliable,
>>>>>>>>>>>>>>>>>> please
check the issue, fix and redeploy
>>>>>>>>>>>>>>>>>>
Log file is located at
>>>>>>>>>>>>>>>>>>
/var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20160722123404-t26vw0.log
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
Interestingly
>>>>>>>>>>>>>>>>>>
============================
>>>>>>>>>>>>>>>>>> If I
try to deploy hosted-engine v3.6, everything goes well in the same
>>>>>>>>>>>>>>>>>>
configuration !! :
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> ....
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Transaction setup
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Misc configuration
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Package installation
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Misc configuration
>>>>>>>>>>>>>>>>>> [ INFO
] Configuring libvirt
>>>>>>>>>>>>>>>>>> [ INFO
] Configuring VDSM
>>>>>>>>>>>>>>>>>> [ INFO
] Starting vdsmd
>>>>>>>>>>>>>>>>>> [ INFO
] Waiting for VDSM hardware info
>>>>>>>>>>>>>>>>>> [ INFO
] Configuring the management bridge
>>>>>>>>>>>>>>>>>> [ INFO
] Creating Volume Group
>>>>>>>>>>>>>>>>>> [ INFO
] Creating Storage Domain
>>>>>>>>>>>>>>>>>> [ INFO
] Creating Storage Pool
>>>>>>>>>>>>>>>>>> [ INFO
] Connecting Storage Pool
>>>>>>>>>>>>>>>>>> [ INFO
] Verifying sanlock lockspace initialization
>>>>>>>>>>>>>>>>>> [ INFO
] Creating Image for 'hosted-engine.lockspace' ...
>>>>>>>>>>>>>>>>>> [ INFO
] Image for 'hosted-engine.lockspace' created successfully
>>>>>>>>>>>>>>>>>> [ INFO
] Creating Image for 'hosted-engine.metadata' ...
>>>>>>>>>>>>>>>>>> [ INFO
] Image for 'hosted-engine.metadata' created successfully
>>>>>>>>>>>>>>>>>> [ INFO
] Creating VM Image
>>>>>>>>>>>>>>>>>> [ INFO
] Destroying Storage Pool
>>>>>>>>>>>>>>>>>> [ INFO
] Start monitoring domain
>>>>>>>>>>>>>>>>>> [ INFO
] Configuring VM
>>>>>>>>>>>>>>>>>> [ INFO
] Updating hosted-engine configuration
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Transaction commit
>>>>>>>>>>>>>>>>>> [ INFO
] Stage: Closing up
>>>>>>>>>>>>>>>>>> [ INFO
] Creating VM
>>>>>>>>>>>>>>>>>>
You can now connect to the VM with the following command:
>>>>>>>>>>>>>>>>>>
/bin/remote-viewer vnc://localhost:5900
>>>>>>>>>>>>>>>>>> ...
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> What
could be the problem?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
_______________________________________________
>>>>>>>>>>>>>>>>>> Users
mailing list
>>>>>>>>>>>>>>>>>>
Users(a)ovirt.org
>>>>>>>>>>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list
>>>>>>>>> Users(a)ovirt.org
>>>>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list
>>>>>> Users(a)ovirt.org
>>>>>>
http://lists.ovirt.org/mailman/listinfo/users
>>>
>>> _______________________________________________
>>> Users mailing list
>>> Users(a)ovirt.org
>>>
http://lists.ovirt.org/mailman/listinfo/users