<div dir="ltr">sanlock is at the latest version (this solved another problem we had a few days ago):<div><br></div><div><div>$ rpm -q sanlock</div><div>sanlock-2.6-7.fc18.x86_64</div><div><br></div><div>the storage is on the same machine as the engine and vdsm.</div>
<div>iptables is up but there is a rule to allow all localhost traffic.</div><div><br></div><br><div class="gmail_quote">On Sun, Mar 24, 2013 at 11:34 AM, Maor Lipchuk <span dir="ltr"><<a href="mailto:mlipchuk@redhat.com" target="_blank">mlipchuk@redhat.com</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">From the VDSM log, it seems that the master storage domain was not<br>
responding.<br>
<br>
Thread-23::DEBUG::2013-03-22<br>
18:50:20,263::domainMonitor::216::Storage.DomainMonitorThread::(_monitorDomain)<br>
Domain 1083422e-a5db-41b6-b667-b9ef1ef244f0 changed its status to Invalid<br>
....<br>
Traceback (most recent call last):<br>
File "/usr/share/vdsm/storage/domainMonitor.py", line 186, in<br>
_monitorDomain<br>
self.domain.selftest()<br>
File "/usr/share/vdsm/storage/nfsSD.py", line 108, in selftest<br>
fileSD.FileStorageDomain.selftest(self)<br>
File "/usr/share/vdsm/storage/fileSD.py", line 480, in selftest<br>
self.oop.os.statvfs(self.domaindir)<br>
File "/usr/share/vdsm/storage/remoteFileHandler.py", line 280, in<br>
callCrabRPCFunction<br>
*args, **kwargs)<br>
File "/usr/share/vdsm/storage/remoteFileHandler.py", line 180, in<br>
callCrabRPCFunction<br>
rawLength = self._recvAll(LENGTH_STRUCT_LENGTH, timeout)<br>
File "/usr/share/vdsm/storage/remoteFileHandler.py", line 146, in _recvAll<br>
raise Timeout()<br>
Timeout<br>
.....<br>
<br>
I'm also see a san lock issue, but I think that is because the storage<br>
could not be reached:<br>
ReleaseHostIdFailure: Cannot release host id:<br>
('1083422e-a5db-41b6-b667-b9ef1ef244f0', SanlockException(16, 'Sanlock<br>
lockspace remove failure', 'Device or resource busy'))<br>
<br>
Can you try to see if the ip tables are running on your host, and if so,<br>
please check if it is blocking the storage server by any chance?<br>
Can you try to manually mount this NFS and see if it works?<br>
Is it possible the storage server got connectivity issues?<br>
<br>
<br>
Regards,<br>
Maor<br>
<div><div class="h5"><br>
On 03/22/2013 08:24 PM, Limor Gavish wrote:<br>
> Hello,<br>
><br>
> I am using Ovirt 3.2 on Fedora 18:<br>
> [wil@bufferoverflow ~]$ rpm -q vdsm<br>
> vdsm-4.10.3-7.fc18.x86_64<br>
><br>
> (the engine is built from sources).<br>
><br>
> I seem to have hit this bug:<br>
> <a href="https://bugzilla.redhat.com/show_bug.cgi?id=922515" target="_blank">https://bugzilla.redhat.com/show_bug.cgi?id=922515</a><br>
><br>
> in the following configuration:<br>
> Single host (no migrations)<br>
> Created a VM, installed an OS inside (Fedora18)<br>
> stopped the VM.<br>
> created template from it.<br>
> Created an additional VM from the template using thin provision.<br>
> Started the second VM.<br>
><br>
> in addition to the errors in the logs the storage domains (both data and<br>
> ISO) crashed, i.e went to "unknown" and "inactive" states respectively.<br>
> (see the attached engine.log)<br>
><br>
> I attached the VDSM and engine logs.<br>
><br>
> is there a way to work around this problem?<br>
> It happens repeatedly.<br>
><br>
> Yuval Meir<br>
><br>
><br>
><br>
</div></div>> _______________________________________________<br>
> Users mailing list<br>
> <a href="mailto:Users@ovirt.org">Users@ovirt.org</a><br>
> <a href="http://lists.ovirt.org/mailman/listinfo/users" target="_blank">http://lists.ovirt.org/mailman/listinfo/users</a><br>
><br>
<br>
<br>
</blockquote></div><br></div></div>