<html>
<head>
<meta content="text/html; charset=ISO-8859-1"
http-equiv="Content-Type">
</head>
<body style="font-family: sans-serif; font-size: 16px;"
bgcolor="#FFFFFF" text="#000000">
<div id="QCMcontainer"
style="font-family:sans-serif;font-size:16px;"><br>
<br>
<span style="color:#000000; font-family:sans-serif;
font-size:12px; font-weight:normal" class="headerSpan">
<div class="moz-cite-prefix">Il 31/10/2013 08:02, Sahina Bose ha
scritto:<br>
</div>
</span>
<blockquote style="border-left: 2px solid rgb(0, 153, 0) !
important; border-right: 2px solid rgb(0, 153, 0) ! important;
padding: 0px 15px; margin: 8px 2px; font-size: medium;"
cite="mid:52720069.8010708@redhat.com" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->
<br>
On 10/31/2013 11:51 AM, Sahina Bose wrote:
<br>
<blockquote style="border-left: 2px solid #009900 !important;
border-right: 2px solid #009900 !important; padding: 0px 15px
0px 15px; margin: 8px 2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->
<br>
On 10/31/2013 09:22 AM, Zhou Zheng Sheng wrote:
<br>
<blockquote style="border-left: 2px solid #009900 !important;
border-right: 2px solid #009900 !important; padding: 0px
15px 0px 15px; margin: 8px 2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->
<br>
on 2013/10/31 06:24, Dan Kenigsberg wrote:
<br>
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900 !important;
padding: 0px 15px 0px 15px; margin: 8px 2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->On
Wed, Oct 30, 2013 at 08:41:43PM +0100, Alessandro Bianchi
wrote:
<br>
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900 !important;
padding: 0px 15px 0px 15px; margin: 8px 2px;"
type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->Il
30/10/2013 18:04, Dan Kenigsberg ha scritto:
<br>
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900
!important; padding: 0px 15px 0px 15px; margin: 8px
2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->On
Wed, Oct 30, 2013 at 02:40:02PM +0100, Alessandro
Bianchi wrote:
<br>
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900
!important; padding: 0px 15px 0px 15px; margin: 8px
2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->
Il 30/10/2013 13:58, Dan Kenigsberg ha scritto:
<br>
<br>
On Wed, Oct 30, 2013 at 11:34:21AM +0100,
Alessandro Bianchi wrote:
<br>
<br>
Hi everyone
<br>
<br>
I've set up a gluster storage with two
replicated bricks
<br>
<br>
DC is up and I created a VM to test gluster
storage
<br>
<br>
If I start the VM WITHOUT any disk attached
(only one virtual DVD) it
<br>
starts fine.
<br>
<br>
If I attach a gluster domain disk thin
provisioning 30 Gb the Vm stucks in
<br>
"waiting for launch" state
<br>
<br>
I see no special activity on the gluster
servers (they serve several other
<br>
shares with no troubles at all and even the ISO
domain is a NFS on
<br>
locally mounted gluster and works fine)
<br>
<br>
I've double checked all the pre requisites and
they look fine (F 19 -
<br>
gluster setup insecure in both glusterd.vol
and volume options -
<br>
uid/gid/insecure )
<br>
<br>
Am I doing something wrong?
<br>
<br>
I'm even unable to stop the VM from the engine
GUI
<br>
<br>
Any advise?
<br>
<br>
Which version of ovirt are you using? Hopefully
ovirt-3.3.0.1.
<br>
For how long is the VM stuck in its "wait for
launch" state?
<br>
What does `virsh -r list` has to say while startup
stalls?
<br>
Would you provide more content of your vdsm.log
and possibly
<br>
libvirtd.log so we can understand what blocks the
VM start-up? Please
<br>
use attachement of pastebin, as your mail agents
wreaks havoc to the log
<br>
lines.
<br>
<br>
<br>
Thank you for your answer.
<br>
<br>
Here are the "facts"
<br>
<br>
In the GUI I see
<br>
<br>
"waiting for launch 3 h"
<br>
<br>
virsh -r list
<br>
Id Nome Stato
<br>
----------------------------------------------------
<br>
3 CentOS_30 terminato
<br>
<br>
vdsClient -s 0 list table
<br>
200dfb05-461e-49d9-95a2-c0a7c7ced669 0
CentOS_30
<br>
WaitForLaunch
<br>
<br>
Packages:
<br>
<br>
ovirt-engine-userportal-3.3.0.1-1.fc19.noarch
<br>
ovirt-log-collector-3.3.1-1.fc19.noarch
<br>
ovirt-engine-restapi-3.3.0.1-1.fc19.noarch
<br>
ovirt-engine-setup-3.3.0.1-1.fc19.noarch
<br>
ovirt-engine-backend-3.3.0.1-1.fc19.noarch
<br>
ovirt-host-deploy-java-1.1.1-1.fc19.noarch
<br>
ovirt-release-fedora-8-1.noarch
<br>
ovirt-engine-setup-plugin-allinone-3.3.0.1-1.fc19.noarch
<br>
ovirt-engine-webadmin-portal-3.3.0.1-1.fc19.noarch
<br>
ovirt-engine-sdk-python-3.3.0.7-1.fc19.noarch
<br>
ovirt-iso-uploader-3.3.1-1.fc19.noarch
<br>
ovirt-engine-websocket-proxy-3.3.0.1-1.fc19.noarch
<br>
ovirt-engine-dbscripts-3.3.0.1-1.fc19.noarch
<br>
ovirt-host-deploy-offline-1.1.1-1.fc19.noarch
<br>
ovirt-engine-cli-3.3.0.5-1.fc19.noarch
<br>
ovirt-engine-tools-3.3.0.1-1.fc19.noarch
<br>
ovirt-engine-lib-3.3.0.1-1.fc19.noarch
<br>
ovirt-image-uploader-3.3.1-1.fc19.noarch
<br>
ovirt-engine-3.3.0.1-1.fc19.noarch
<br>
ovirt-host-deploy-1.1.1-1.fc19.noarch
<br>
<br>
I attach the full vdsm log
<br>
<br>
Look around 30-10 10:30 to see all what happens
<br>
<br>
Despite the "terminated" label in output from
virsh I still see the VM
<br>
"waiting for launch" in the GUI, so I suspect
the answer to "how long" may
<br>
be "forever"
<br>
<br>
Since this is a test VM I can do whatever test
you may need to track the
<br>
problem included destroy and rebuild
<br>
<br>
It would be great to have gluster support stable
in ovirt!
<br>
<br>
Thank you for your efforts
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
The log has an ominous failed attempt to start the VM,
followed by an
<br>
immediate vdsm crash. Is it reproducible?
<br>
<br>
We have plenty of issues lurking here:
<br>
1. Why has libvirt failed to create the VM? For this,
please find clues
<br>
in the complete non-line-broken CentOS_30.log and
libvirtd.log.
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
attached to this messages
<br>
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900
!important; padding: 0px 15px 0px 15px; margin: 8px
2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->2.
Why was vdsm killed? Does /var/log/message has a clue
from systemd?
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
result of cat /var/log/messages | grep vdsm attached
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
I do not see an explicit attempt to take vdsmd down. Do
you see any other
<br>
incriminating message correlated with
<br>
<br>
Oct 30 08:51:15 hypervisor respawn: slave
'/usr/share/vdsm/vdsm' died, respawning slave
<br>
<br>
<br>
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900 !important;
padding: 0px 15px 0px 15px; margin: 8px 2px;"
type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900
!important; padding: 0px 15px 0px 15px; margin: 8px
2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->3.
We may have a nasty race: if Vdsm crashes just before
it has
<br>
registered that the VM is down.
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
<!--[if !IE]></DIV><![endif]--></blockquote>
Actually, this is not the issue: vdsm tries (and fails,
due to qemu/libvirt
<br>
bug) to destroy the VM.
<br>
<br>
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900 !important;
padding: 0px 15px 0px 15px; margin: 8px 2px;"
type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->
<blockquote style="border-left: 2px solid #009900
!important; border-right: 2px solid #009900
!important; padding: 0px 15px 0px 15px; margin: 8px
2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->4.
We used to force Vdsm to run with LC_ALL=C. It seems
that the grand
<br>
service rewrite by Zhou
(<a class="moz-txt-link-freetext" href="http://gerrit.ovirt.org/15578">http://gerrit.ovirt.org/15578</a>) has changed
<br>
that. This may have adverse effects, since AFAIR
we sometimes parse
<br>
application output, and assume that it's in C.
Having a non-English
<br>
log file is problematic on it's own for support
personal, used to
<br>
grep for keywords. ybronhei, was it intensional?
Can it be reverted
<br>
or at least scrutinized?
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
currentely it still says "waiting for launch 9h"
<br>
<br>
I don't abort it so if you need any other info I can
have them
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
libvirtd fails to connect to qemu's monitor. This smells
like a qemu bug that
<br>
is beyond my over-the-mailing-list debugging abilities :-(
<br>
You may want to strace or gdb the running qemu process, or
to try to re-attach
<br>
to it by restarting libvirtd.
<br>
<br>
2013-10-30 07:51:10.045+0000: 8304: debug :
qemuProcessStart:3804 : Waiting for monitor to show up
<br>
2013-10-30 07:51:10.045+0000: 8304: debug :
qemuProcessWaitForMonitor:1707 : Connect monitor to
0x7fc1640eab80 'CentOS_30'
<br>
2013-10-30 07:51:10.246+0000: 8304: debug :
qemuMonitorOpenInternal:751 : QEMU_MONITOR_NEW:
mon=0x7fc1640ef6b0 refs=2 fd=27
<br>
2013-10-30 07:51:10.246+0000: 8304: debug :
qemuMonitorSetCapabilities:1145 : mon=0x7fc1640ef6b0
<br>
2013-10-30 07:51:10.246+0000: 8304: debug :
qemuMonitorSend:887 : QEMU_MONITOR_SEND_MSG:
mon=0x7fc1640ef6b0
msg={"execute":"qmp_capabilities","id":"libvirt-1"}
<br>
<br>
fd=-1
<br>
2013-10-30 07:51:13.097+0000: 8296: error :
qemuMonitorIORead:505 : Unable to read from monitor:
Connessione interrotta dal corrispondente
<br>
2013-10-30 07:51:13.097+0000: 8296: debug :
qemuMonitorIO:638 : Error on monitor Unable to read from
monitor: Connessione interrotta dal corrispondente
<br>
2013-10-30 07:51:13.097+0000: 8296: debug :
qemuMonitorIO:672 : Triggering error callback
<br>
2013-10-30 07:51:13.097+0000: 8296: debug :
qemuProcessHandleMonitorError:351 : Received error on
0x7fc1640eab80 'CentOS_30'
<br>
2013-10-30 07:51:13.097+0000: 8304: debug :
qemuMonitorSend:899 : Send command resulted in error
Unable to read from monitor: Connessione interrotta dal
corrispondente
<br>
2013-10-30 07:51:13.097+0000: 8304: debug :
qemuProcessStop:3992 : Shutting down VM 'CentOS_30'
pid=7655 flags=0
<br>
2013-10-30 07:51:13.097+0000: 8296: debug :
qemuMonitorIO:638 : Error on monitor Unable to read from
monitor: Connessione interrotta dal corrispondente
<br>
2013-10-30 07:51:13.097+0000: 8296: debug :
qemuMonitorIO:661 : Triggering EOF callback
<br>
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
I noticed in the previous attached CentOS_30.log, qemu said
"Gluster
<br>
connection failed for server=172.16.0.100" for each of the
first few
<br>
attempts to start qemu. That's why libvirt said it could not
connect to
<br>
qemu monitor, the real reason was that qemu process exited
after it
<br>
checked the gluster storage. libvirt error message was not
very friendly
<br>
in this case.
<br>
<br>
Then I see on "2013-10-30 08:28:51.047+0000" in
CentOS_30.log, qemu
<br>
started successfully without reporting gluster error, this
was fine. In
<br>
libvirtd.log on the corresponding time spot, it libvirt
connected to the
<br>
monitor and reported it was about to issue a handshake.
<br>
<br>
Investigate the CentOS_30.log further, I find the attempts
to start qemu
<br>
after "2013-10-30 08:28:51" were failed with the same error
"Gluster
<br>
connection failed for server=172.16.0.100". Maybe there is a
problem in
<br>
the gluster server or firewall? I'm not sure.
<br>
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
<br>
<a class="moz-txt-link-freetext" href="http://www.ovirt.org/Features/GlusterFS_Storage_Domain#Important_Pre-requisites">http://www.ovirt.org/Features/GlusterFS_Storage_Domain#Important_Pre-requisites</a>
- I assume these have been followed w.r.t setting permissions.
<br>
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
<br>
Sorry, I missed that you had set all the insecure options on the
volume.
<br>
<br>
"Gluster connection failed for server" has a "Transport endpoint
is not connected" error message.
<br>
<br>
Are you able to manually mount the gluster volume?
<br>
<br>
<blockquote style="border-left: 2px solid #009900 !important;
border-right: 2px solid #009900 !important; padding: 0px 15px
0px 15px; margin: 8px 2px;" type="cite"><!--[if !IE]><DIV style="border-left: 2px solid #009900; border-right: 2px solid #009900; padding: 0px 15px; margin: 2px 0px;"><![endif]-->
<br>
_______________________________________________
<br>
Users mailing list
<br>
<a class="moz-txt-link-abbreviated" href="mailto:Users@ovirt.org">Users@ovirt.org</a>
<br>
<a class="moz-txt-link-freetext" href="http://lists.ovirt.org/mailman/listinfo/users">http://lists.ovirt.org/mailman/listinfo/users</a>
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
<br>
<br>
<!--[if !IE]></DIV><![endif]--></blockquote>
Yes<br>
<br>
The errors you see before 10:30 were due to this bug<br>
<br>
<a href="https://bugzilla.redhat.com/show_bug.cgi?id=988299">https://bugzilla.redhat.com/show_bug.cgi?id=988299</a><br>
<br>
I specified the storage path as 172.16.0.100:/vms so the result
was in connection failure even if the gluster share were properly
mounted in the DS on the hypervisor<br>
<br>
I had to remove the storage and redifine it as follows<br>
<br>
172.16.0.100:vms<br>
<br>
Without the trailing / and connection errors went away<br>
<br>
So please ignore the earlier glusterfs connection errors, I
suggest to look at logsta starting around 10:00 when the
connection was properly recognized<br>
<br>
Thnak you all for your efforts<br>
<br>
P.S. GUI now says "waiting for launch 21 h": does it have any
timeout?<br>
<br>
<div class="moz-signature">-- <br>
<meta http-equiv="CONTENT-TYPE" content="text/html;
charset=ISO-8859-1">
<title></title>
<meta name="generator" content="Bluefish 2.0.3">
<meta name="author" content="Alessandro Bianchi">
<meta name="CREATED" content="20100306;9474300">
<meta name="CHANGEDBY" content="Alessandro ">
<meta name="CHANGED" content="20100306;10212100">
<style type="text/css">
        <!--
                P { font-family: "Arial", "Helvetica", sans-serif; font-size: 10pt }
                P.nome { color: #ff8000; font-family: "Arial", "Helvetica", sans-serif; font-size: 12pt; font-weight: bold; text-align: center }
                P.indirizzo { color: #0084d1; font-family: "Arial", "Helvetica", sans-serif; font-size: 10pt; font-weight: bold; line-height: 0.48cm; text-align: center }
                P.info { color: #b3b3b3; font-family: "Arial", "Helvetica", sans-serif; font-size: 9pt }
                A:link { color: #005dff; text-decoration: none }
                A:visited { color: #005dff; text-decoration: none }
        -->
        </style>
<p class="nome">SkyNet SRL</p>
<p class="indirizzo">Via Maggiate 67/a - 28021 Borgomanero (NO)
- tel.
+39 0322-836487/834765 - fax +39 0322-836608</p>
<p align="CENTER"><a href="http://www.skynet.it/">http://www.skynet.it</a></p>
<p class="indirizzo">Autorizzazione Ministeriale n.197</p>
<p class="info">Le informazioni contenute in questo messaggio
sono
riservate e confidenziali ed è vietata la diffusione in
qualunque
modo eseguita.<br>
Qualora Lei non fosse la persona a cui il presente
messaggio è destinato, La invitiamo ad eliminarlo ed a
distruggerlo
non divulgandolo, dandocene gentilmente comunicazione. <br>
Per
qualsiasi informazione si prega di contattare <a class="moz-txt-link-abbreviated" href="mailto:info@skynet.it">info@skynet.it</a>
(e-mail dell'azienda). Rif. D.L. 196/2003</p>
</div>
<br>
</div>
</body>
</html>