On 11/06/15 11:00 +0000, Soeren Malchow wrote:
We are still having this problem and we can not figure out what to do,
i
sent the logs already as download, can i do anything else to help ?
Hi. I'm sorry but I don't have any new information for you yet. One
thing you could do is create a new bug for this issue so we can track
it better. Please try to include as much information as possible from
this discussion (including relevant log files) in your report. So far
you are the only one reporting these issues so we'll want to work to
narrow down the specific scenario that is causing this problem and get
the right people working on the solution.
On 04/06/15 17:08, "Soeren Malchow"
<soeren.malchow(a)mcon.net> wrote:
>Hi,
>
>I would send those, but unfortunately we did not think about the journals
>getting deleted after a reboot.
>
>I just made the journals persistent on the servers, we are trying to
>trigger the error again last time we only got half way through the VM’s
>when removing the snapshots so we have a good chance that it comes up
>again.
>
>Also the libvirt logs to the journal not to libvirtd.log, i would send the
>journal directly to you and Eric via our data exchange servers
>
>
>Soeren
>
>On 04/06/15 16:17, "Adam Litke" <alitke(a)redhat.com> wrote:
>
>>On 04/06/15 13:08 +0000, Soeren Malchow wrote:
>>>Hi Adam, Hi Eric,
>>>
>>>We had this issue again a few minutes ago.
>>>
>>>One machine went down exactly the same way as described, the machine had
>>>only one snapshot and it was the only snapshot that was removed, before
>>>that in the same scriptrun we deleted the snapshots of 15 other Vms,
>>>some
>>>without, some with 1 and some with several snapshots.
>>>
>>>Can i provide anything from the logs that helps ?
>>
>>Let's start with the libvirtd.log on that host. It might be rather
>>large so we may need to find a creative place to host it.
>>
>>>
>>>Regards
>>>Soeren
>>>
>>>
>>>
>>>On 03/06/15 18:07, "Soeren Malchow" <soeren.malchow(a)mcon.net>
wrote:
>>>
>>>>Hi,
>>>>
>>>>This is not happening every time, the last time i had this, it was a
>>>>script runnning, and something like th 9. Vm and the 23. Vm had a
>>>>problem,
>>>>and it is not always the same VMS, it is not about the OS (happen for
>>>>Windows and Linux alike)
>>>>
>>>>And as i said it also happened when i tried to remove the snapshots
>>>>sequentially, here is the code (i know it is probably not the elegant
>>>>way,
>>>>but i am not a developer) and the code actually has correct indentions.
>>>>
>>>><― snip ―>
>>>>
>>>>print "Snapshot deletion"
>>>>try:
>>>> time.sleep(300)
>>>> Connect()
>>>> vms = api.vms.list()
>>>> for vm in vms:
>>>> print ("Deleting snapshots for %s ") % vm.name
>>>> snapshotlist = vm.snapshots.list()
>>>> for snapshot in snapshotlist:
>>>> if snapshot.description != "Active VM":
>>>> time.sleep(30)
>>>> snapshot.delete()
>>>> try:
>>>> while
>>>>api.vms.get(name=vm.name).snapshots.get(id=snapshot.id).snapshot_status
>>>>==
>>>>"locked":
>>>> print("Waiting for snapshot %s on %s
deletion
>>>>to
>>>>finish") % (snapshot.description, vm.name)
>>>> time.sleep(60)
>>>> except Exception as e:
>>>> print ("Snapshot %s does not exist anymore")
%
>>>>snapshot.description
>>>> print ("Snapshot deletion for %s done") % vm.name
>>>> print ("Deletion of snapshots done")
>>>> api.disconnect()
>>>>except Exception as e:
>>>> print ("Something went wrong when deleting the
snapshots\n%s") %
>>>>str(e)
>>>>
>>>>
>>>>
>>>><― snip ―>
>>>>
>>>>
>>>>Cheers
>>>>Soeren
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>On 03/06/15 15:20, "Adam Litke" <alitke(a)redhat.com>
wrote:
>>>>
>>>>>On 03/06/15 07:36 +0000, Soeren Malchow wrote:
>>>>>>Dear Adam
>>>>>>
>>>>>>First we were using a python script that was working on 4 threads
and
>>>>>>therefore removing 4 snapshots at the time throughout the
cluster,
>>>>>>that
>>>>>>still caused problems.
>>>>>>
>>>>>>Now i took the snapshot removing out of the threaded part an i am
>>>>>>just
>>>>>>looping through each snapshot on each VM one after another, even
with
>>>>>>³sleeps² inbetween, but the problem remains.
>>>>>>But i am getting the impression that it is a problem with the
amount
>>>>>>of
>>>>>>snapshots that are deleted in a certain time, if i delete
manually
>>>>>>and
>>>>>>one
>>>>>>after another (meaning every 10 min or so) i do not have problems,
if
>>>>>>i
>>>>>>delete manually and do several at once and on one VM the next one
>>>>>>just
>>>>>>after one finished, the risk seems to increase.
>>>>>
>>>>>Hmm. In our lab we extensively tested removing a snapshot for a VM
>>>>>with 4 disks. This means 4 block jobs running simultaneously. Less
>>>>>than 10 minutes later (closer to 1 minute) we would remove a second
>>>>>snapshot for the same VM (again involving 4 block jobs). I guess we
>>>>>should rerun this flow on a fully updated CentOS 7.1 host to see
about
>>>>>local reproduction. Seems your case is much simpler than this
though.
>>>>>Is this happening every time or intermittently?
>>>>>
>>>>>>I do not think it is the number of VMS because we had this on
hosts
>>>>>>with
>>>>>>only 3 or 4 Vms running
>>>>>>
>>>>>>I will try restarting the libvirt and see what happens.
>>>>>>
>>>>>>We are not using RHEL 7.1 only CentOS 7.1
>>>>>>
>>>>>>Is there anything else we can look at when this happens again ?
>>>>>
>>>>>I'll defer to Eric Blake for the libvirt side of this. Eric,
would
>>>>>enabling debug logging in libvirtd help to shine some light on the
>>>>>problem?
>>>>>
>>>>>--
>>>>>Adam Litke
>>>>
>>>>_______________________________________________
>>>>Users mailing list
>>>>Users(a)ovirt.org
>>>>http://lists.ovirt.org/mailman/listinfo/users
>>>
>>
>>--
>>Adam Litke
>
>_______________________________________________
>Users mailing list
>Users(a)ovirt.org
>http://lists.ovirt.org/mailman/listinfo/users
--
Adam Litke