Hi,
On Fri, 25 May 2018 08:11:13 +0000
"Vrgotic, Marko" <M.Vrgotic(a)activevideo.com> wrote:
Dear Nir, Arik and Richard,
I hope discussion will continue somewhere where i am able to join as watcher at least.
please open a bug on VDSM. This is something we need to deal with during
import -- or at least prevent users from importing.
I have not seen any communication since Nir’s proposal. Please, if
possible allow me to somehow track in which direction you are leaning.
In the meantime, as experienced engineers, do you have any suggestion how I could
“workaround” current problem?
Rebasing image using qemu-img to remove backing file did not help (VM was able to start,
but No Boot Device) and I think now that is due to image having functional dependency of
base image.
What do you mean by rebasing? On which backing image did you rebase it?
I'm not too familiar with openstack, but I'd suggest doing a 'qemu-img
convert' on the disk in openstack to squash the backing change into new (and complete)
image, assign this new disk to your VM and import it to oVirt.
Tomas
Like with VMs in oVrt, where template can not be deleted if there are still VMs existing,
which were created from that template.
Please advise
— — —
Met vriendelijke groet / Best regards,
Marko Vrgotic
Sr. System Engineer
ActiveVideo
Tel. +31 (0)35 677 4131
email: m.vrgotic(a)activevideo.com
skype: av.mvrgotic.se
www.activevideo.com
________________________________
From: Vrgotic, Marko
Sent: Thursday, May 24, 2018 5:26:30 PM
To: Nir Soffer
Cc: users(a)ovirt.org; Richard W.M. Jones; Arik Hadas
Subject: Re: [ovirt-users] Libvirt ERROR cannot access backing file after importing VM
from OpenStack
Dear Nir,
I believe i understand now. The image imported is not base image, but required backing
file to be able to work properly.
Maybe silly move, but i have tried to “solve/workaround” around the problem by rebasing
image to remove backing file dependency, but it’s clear now why I than saw that “no
bootable device found” during imported VM boot.
I support you suggestion to solve the import by either importing complete chain or
recreating image so in a way it’s independent from former chain.
If you decide to go this way, please let me know which issue to track and if you need any
more data provided from me.
I still need to solve problem with 200+ VM wanting to move to oVirt.
Kindly awaiting further updates.
— — —
Met vriendelijke groet / Best regards,
Marko Vrgotic
Sr. System Engineer
ActiveVideo
Tel. +31 (0)35 677 4131
email: m.vrgotic(a)activevideo.com
skype: av.mvrgotic.se
www.activevideo.com
________________________________
From: Nir Soffer <nsoffer(a)redhat.com>
Sent: Thursday, May 24, 2018 5:13:47 PM
To: Vrgotic, Marko
Cc: users(a)ovirt.org; Richard W.M. Jones; Arik Hadas
Subject: Re: [ovirt-users] Libvirt ERROR cannot access backing file after importing VM
from OpenStack
On Thu, May 24, 2018 at 6:06 PM Vrgotic, Marko
<M.Vrgotic@activevideo.com<mailto:M.Vrgotic@activevideo.com>> wrote:
Dear Nir,
Thank you for quick reply.
Ok, why it will not work?
Because the image has a backing file which is not accessible to oVirt.
I used qemu+tcp connection, via import method through engine admin UI.
Images was imported and converted according logs, still “backing file” invalid entry
remained.
Also, I did use same method before, connecting to plain “libvirt kvm” host, import and
conversion went smooth, no backend file.
Image format is qcow(2) which is supported by oVirt.
What am I missing? Should I use different method?
I guess this is not a problem on your side, but a bug in our side.
Either we should block the operation that cannot work, or fix the process
so we don't refer to non-existing image.
When importing we have 2 options:
- import the entire chain, importing all images in the chain, converting
each image to oVirt volume, and updating the backing file of each layer
to point to the oVirt image.
- import the current state of the image into a new image, using either raw
or qcow2, but without any backing file.
Arik, do you know why we create qcow2 file with invalid backing file?
Nir
Kindly awaiting your reply.
— — —
Met vriendelijke groet / Best regards,
Marko Vrgotic
Sr. System Engineer
ActiveVideo
Tel. +31 (0)35 677 4131<tel:+31%2035%20677%204131>
email: m.vrgotic@activevideo.com<mailto:m.vrgotic@activevideo.com>
skype: av.mvrgotic.se<http://av.mvrgotic.se>
www.activevideo.com<http://www.activevideo.com>
________________________________
From: Nir Soffer <nsoffer@redhat.com<mailto:nsoffer@redhat.com>>
Sent: Thursday, May 24, 2018 4:09:40 PM
To: Vrgotic, Marko
Cc: users@ovirt.org<mailto:users@ovirt.org>; Richard W.M. Jones; Arik Hadas
Subject: Re: [ovirt-users] Libvirt ERROR cannot access backing file after importing VM
from OpenStack
On Thu, May 24, 2018 at 5:05 PM Vrgotic, Marko
<M.Vrgotic@activevideo.com<mailto:M.Vrgotic@activevideo.com>> wrote:
Dear oVirt team,
When trying to start imported VM, it fails with following message:
ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ForkJoinPool-1-worker-2) [] EVENT_ID: VM_DOWN_ERROR(119), VM instance-00000673 is down
with error. Exit message: Cannot access backing file
'/var/lib/nova/instances/_base/2f4f8c5fc11bb83bcab03f4c829ddda4da8c0bce' of
storage file
'/rhev/data-center/mnt/glusterSD/aws-gfs-01.awesome.lan:_gv0__he/2607c265-248c-40ad-b020-f3756454839e/images/816ac00f-ba98-4827-b5c8-42a8ba496089/8ecfcd5b-db67-4c23-9869-0e20d7553aba'
(as uid:107, gid:107): No such file or directory.
Platform details:
Ovirt SHE
Version 4.2.2.6-1.el7.centos
GlusterFS, unmanaged by oVirt.
VM is imported & converted from OpenStack, according to log files, successfully (one
WARN, related to different MAC address):
2018-05-24 12:03:31,028+02 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetVmsNamesFromExternalProviderVDSCommand]
(default task-29) [cc5931a2-1af5-4d65-b0b3-362588db9d3f] FINISH,
GetVmsNamesFromExternalProviderVDSCommand, return: [VM [instance-0001f94c], VM
[instance-00078f6a], VM [instance-00000814], VM [instance-0001f9ac], VM
[instance-000001ff], VM [instance-0001f718], VM [instance-00000673], VM
[instance-0001ecf2], VM [instance-00078d38]], log id: 7f178a5e
2018-05-24 12:48:33,722+02 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetVmsNamesFromExternalProviderVDSCommand]
(default task-8) [103d56e1-7449-4853-ae50-48ee94d43d77] FINISH,
GetVmsNamesFromExternalProviderVDSCommand, return: [VM [instance-0001f94c], VM
[instance-00078f6a], VM [instance-00000814], VM [instance-0001f9ac], VM
[instance-000001ff], VM [instance-0001f718], VM [instance-00000673], VM
[instance-0001ecf2], VM [instance-00078d38]], log id: 3aa178c5
2018-05-24 12:48:47,291+02 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetVmsFullInfoFromExternalProviderVDSCommand]
(default task-17) [4bf555c7-9d64-4ecc-b059-8a60a4b27bdd] START,
GetVmsFullInfoFromExternalProviderVDSCommand(HostName = aws-ovhv-01,
GetVmsFromExternalProviderParameters:{hostId='cbabe1e8-9e7f-4c4b-be9c-49154953564d',
url='qemu+tcp://root@172.19.0.12/system<http://root@172.19.0.12/system>',
username='null', originType='KVM',
namesOfVms='[instance-00000673]'}), log id: 4c445109
2018-05-24 12:48:47,318+02 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetVmsFullInfoFromExternalProviderVDSCommand]
(default task-17) [4bf555c7-9d64-4ecc-b059-8a60a4b27bdd] FINISH,
GetVmsFullInfoFromExternalProviderVDSCommand, return: [VM [instance-00000673]], log id:
4c445109
2018-05-24 12:49:20,466+02 INFO
[org.ovirt.engine.core.bll.exportimport.ImportVmFromExternalProviderCommand] (default
task-41) [14edb003-b4a0-4355-b3de-da2b68774fe3] Lock Acquired to object
'EngineLock:{exclusiveLocks='[instance-00000673=VM_NAME,
1f0b608f-7cfc-4b27-a876-b5d8073011a1=VM]', sharedLocks=''}'
2018-05-24 12:49:20,586+02 WARN
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engine-Thread-653408) [14edb003-b4a0-4355-b3de-da2b68774fe3]
EVENT_ID: MAC_ADDRESS_IS_EXTERNAL(925), VM instance-00000673 has MAC address(es)
fa:16:3e:74:18:50, which is/are out of its MAC pool definitions.
2018-05-24 12:49:21,021+02 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engine-Thread-653408) [14edb003-b4a0-4355-b3de-da2b68774fe3]
EVENT_ID: IMPORTEXPORT_STARTING_IMPORT_VM(1,165), Starting to import Vm instance-00000673
to Data Center AVEUNL, Cluster AWSEUOPS
2018-05-24 12:49:28,816+02 INFO
[org.ovirt.engine.core.bll.exportimport.ImportVmFromExternalProviderCommand]
(EE-ManagedThreadFactory-engine-Thread-653407) [] Lock freed to object
'EngineLock:{exclusiveLocks='[instance-00000673=VM_NAME,
1f0b608f-7cfc-4b27-a876-b5d8073011a1=VM]', sharedLocks=''}'
2018-05-24 12:49:28,911+02 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.ConvertVmVDSCommand]
(EE-ManagedThreadFactory-commandCoordinator-Thread-2) [2047673e] START,
ConvertVmVDSCommand(HostName = aws-ovhv-01,
ConvertVmVDSParameters:{hostId='cbabe1e8-9e7f-4c4b-be9c-49154953564d',
url='qemu+tcp://root@172.19.0.12/system<http://root@172.19.0.12/system>',
username='null', vmId='1f0b608f-7cfc-4b27-a876-b5d8073011a1',
vmName='instance-00000673',
storageDomainId='2607c265-248c-40ad-b020-f3756454839e',
storagePoolId='5a5de92c-0120-0167-03cb-00000000038a',
virtioIsoPath='null', compatVersion='null',
Disk0='816ac00f-ba98-4827-b5c8-42a8ba496089'}), log id: 53408517
2018-05-24 12:49:29,010+02 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-commandCoordinator-Thread-2) [2047673e] EVENT_ID:
IMPORTEXPORT_STARTING_CONVERT_VM(1,193), Starting to convert Vm instance-00000673
2018-05-24 12:52:57,982+02 INFO [org.ovirt.engine.core.bll.UpdateVmCommand] (default
task-16) [df1d5f72-eb17-46e4-9946-20ca9809b54c] Failed to Acquire Lock to object
'EngineLock:{exclusiveLocks='[instance-00000673=VM_NAME]',
sharedLocks='[1f0b608f-7cfc-4b27-a876-b5d8073011a1=VM]'}'
2018-05-24 12:59:24,575+02 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engineScheduled-Thread-20) [2047673e] EVENT_ID:
IMPORTEXPORT_IMPORT_VM(1,152), Vm instance-00000673 was imported successfully to Data
Center AVEUNL, Cluster AWSEUOPS
Than trying to start VM fails with following messages:
2018-05-24 13:00:32,085+02 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engine-Thread-653729) [] EVENT_ID: USER_STARTED_VM(153), VM
instance-00000673 was started by admin@internal-authz (Host: aws-ovhv-06).
2018-05-24 13:00:33,417+02 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-9) [] VM
'1f0b608f-7cfc-4b27-a876-b5d8073011a1'(instance-00000673) moved from
'WaitForLaunch' --> 'Down'
2018-05-24 13:00:33,436+02 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(ForkJoinPool-1-worker-9) [] EVENT_ID: VM_DOWN_ERROR(119), VM instance-00000673 is down
with error. Exit message: Cannot access backing file
'/var/lib/nova/instances/_base/2f4f8c5fc11bb83bcab03f4c829ddda4da8c0bce' of
storage file
'/rhev/data-center/mnt/glusterSD/aws-gfs-01.awesome.lan:_gv0__he/2607c265-248c-40ad-b020-f3756454839e/images/816ac00f-ba98-4827-b5c8-42a8ba496089/8ecfcd5b-db67-4c23-9869-0e20d7553aba'
(as uid:107, gid:107): No such file or directory.
2018-05-24 13:00:33,437+02 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
(ForkJoinPool-1-worker-9) [] add VM
'1f0b608f-7cfc-4b27-a876-b5d8073011a1'(instance-00000673) to rerun treatment
2018-05-24 13:00:33,455+02 WARN
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engine-Thread-653732) [] EVENT_ID:
USER_INITIATED_RUN_VM_FAILED(151), Failed to run VM instance-00000673 on Host
aws-ovhv-06.
2018-05-24 13:00:33,460+02 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engine-Thread-653732) [] EVENT_ID: USER_FAILED_RUN_VM(54), Failed
to run VM instance-00000673 (User: admin@internal-authz).
Checking on the Gluster volume, directory and files exist, permissions are in order:
[root@aws-ovhv-01 816ac00f-ba98-4827-b5c8-42a8ba496089]
-rw-rw----. 1 vdsm kvm 14G May 24 12:59 8ecfcd5b-db67-4c23-9869-0e20d7553aba
-rw-rw----. 1 vdsm kvm 1.0M May 24 12:49 8ecfcd5b-db67-4c23-9869-0e20d7553aba.lease
-rw-r--r--. 1 vdsm kvm 310 May 24 12:49 8ecfcd5b-db67-4c23-9869-0e20d7553aba.meta
Than I have checked image info, and noticed that backing file entry is pointing to
non-existing location, which does and should not exist on oVirt hosts:
[root@aws-ovhv-01 816ac00f-ba98-4827-b5c8-42a8ba496089]# qemu-img info
8ecfcd5b-db67-4c23-9869-0e20d7553aba
image: 8ecfcd5b-db67-4c23-9869-0e20d7553aba
file format: qcow2
virtual size: 160G (171798691840 bytes)
disk size: 14G
cluster_size: 65536
backing file: /var/lib/nova/instances/_base/2f4f8c5fc11bb83bcab03f4c829ddda4da8c0bce
Format specific information:
compat: 1.1
lazy refcounts: false
refcount bits: 16
corrupt: false
Can somebody advise me how to fix, address this, as I am in need of importing 200+ VMs
from OpenStack to oVirt?
Sure this qcow2 file will not work in oVirt.
I wonder how you did the import?
Nir