
On Mon, Jun 27, 2016 at 6:32 PM, David Gerner <davegerner1975@gmail.com> wrote:
Hi,
I'm running into a strange issue that I've been trying to solve unsuccessfully for a few days now, and I was hoping someone could offer some insight.
A few days ago I needed to reboot the the server that hosts the management engine and is also a node in the system. Following proper procedure I selected a new host to be the SPM, migrated VMs off the Host, and put it into maintenance mode. After the host came back up (and the management engine was back online) I noticed that one of my VMs had halted on a storage error, to rule out the SPM being the issue I asked oVirt to select a new SPM and it was stuck in a contending loop where each host tries to contend for SPM status but ultimately fails (every other VM also halted by this point). The error was "BlockSD master file system FSCK error", after researching the error I found a post on this list that was the same error and the author said that a simple FSCK on the offending file system fixed his issue. I had to force shutdown every VM from the halted state and put all but one host into maintenance mode. On that host I ran a FSCK on the offending volume which found a lot of short read errors which it fixed and afterwards the contending loop was broke and hosts could now successfully be an SPM.
This means that the master file system is probably ok on the current master domain, which may be different domain - do you have more than one storage domain?
Now every VM halts on start or resume, even ones that were offline at the time of the earlier incident, with a Storage Error "abnormal vm stop device virtio-disk0 error eio". I can't even create new disks because it fails with an error. I've attached what I think is the relevant VDSM log portion of a VM trying to resume, if more is needed please just let me know.
We need full vdsm logs and engine logs to understand such errors. I think the best thing would be to open a bug and attach these logs.
I'm worried FSCK and I mangled the file system, and have no idea how to repair it.
It can be useful to investigate the master that failed and you fixed using fsck. You can copy this lv contents like this: 1. Make sure the domain is not the master domain, and that the master lv is not mounted. If the lv is mounted, you can force another domain to be the master domain by deactivating this domain. Since your vms are not running, this should not be a problem. If you deactivated this domain, you may need to connect to it again, you can do this using iscsiadm. If you connect to the target manually using iscsiadm, do delete the node when you finish, so it will not be connected automatically on reboot. 2. Activate the master lv if needed: lvchange -ay vgname/master Warning: do not mount this lv 3. Copy the lv using dd to other storage: dd if=/dev/vgname/master of=master.backup bs=1M iflag=direct oflag=direct 4. Deactivate the master lv lvchange -an vgname/master 5. Compress the backup And share the image. Nir