[Users] node can not access disks
Gadi Ickowicz
gickowic at redhat.com
Tue Feb 11 08:50:30 UTC 2014
Hi,
well at this stage it looks very strange:
* it looks like the disks are really not on the storage domain - the lv for the vm you are attemtping to start (From the previous mail's logs) is not visible at all.
* did something happen to the storage server itself/is it possible that things were deleted from the storage itself? (maybe you deleted the vms' disks?)
* On the problematic node I see that the lvm command gave some errors regarding missing devices:
/dev/mapper/360a9800042415569305d434565795a54: read failed after 0 of 4096 at 10737352704: Chyba vstupu/výstupu
/dev/mapper/360a9800042415569305d434565795a54: read failed after 0 of 4096 at 10737410048: Chyba vstupu/výstupu
/dev/mapper/360a9800042415569305d434565795a54: read failed after 0 of 4096 at 0: Chyba vstupu/výstupu
/dev/mapper/360a9800042415569305d434565795a54: read failed after 0 of 4096 at 4096: Chyba vstupu/výstupu
* Is the domain a single LUN or multiple LUNs? what is the output of
# multipath -ll
# vgs
could you attach the full vdsm log from the problematic host maybe?
Gadi Ickowicz
----- Original Message -----
From: "Jakub Bittner" <j.bittner at nbu.cz>
To: "Gadi Ickowicz" <gickowic at redhat.com>
Cc: users at ovirt.org
Sent: Tuesday, February 11, 2014 10:25:33 AM
Subject: Re: [Users] node can not access disks
Dne 11.2.2014 07:47, Gadi Ickowicz napsal(a):
> lvs ce77262f-8346-42e5-823a-bd321f0814e7
Hello,
it is ISCSI domain. All nodes are problematic.
lvs output:
http://fpaste.org/76058/
I restarted problematic node but it did not help. So I removed it from
cluster, but many other VMs on ther nodes are dead (can not run, or can
not boot from disk because they do not see it). About 5% of VM we were
able to export or restart to stable status
More information about the Users
mailing list