Looks like LSM with libgfapi enabled can't work at the moment[1]


[1] - https://bugzilla.redhat.com/show_bug.cgi?id=1481688#c38

On Thu, Jul 18, 2019 at 12:05 PM Christoph Köhler <koehler@luis.uni-hannover.de> wrote:
The engine runs on version 4.3.4

The hypervisors run on 4.2.8, (version vdsm: 4.20.46-1.el7)
with libvirt 4.5.0-10.el7_6.7

By vdsm.log :

2019-07-18 09:29:09,744+0200 INFO  (jsonrpc/2) [storage.StorageDomain]
Creating domain run directory
u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640' (fileSD:577)
2019-07-18 09:29:09,744+0200 INFO  (jsonrpc/2) [storage.fileUtils]
Creating directory:
/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640 mode: None
(fileUtils:197)

2019-07-18 09:29:09,744+0200 INFO  (jsonrpc/2) [storage.StorageDomain]
Creating symlink from
/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50
to /var/run/vdsm/storage/4dabb6d6-4be5-458
c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 (fileSD:580)

2019-07-18 09:29:09,789+0200 INFO  (jsonrpc/2) [vdsm.api] FINISH
prepareImage return={'info': {'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'type': 'network', 'hosts': [{'port':
'0', 'transport': 'tcp', 'name': '192.168.11.20'}, {'port': '0',
'transport': 'tcp', 'name': 'glusrv01'}, {'port': '0', 'transport':
'tcp', 'name': 'glusrv02'}, {'port': '0', 'transport': 'tcp', 'name':
'glusrv03'}], 'protocol': 'gluster'}, 'path': u'/rhev/data-center/mn
t/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'imgVolumesInfo': [{'domainID': u'4dabb6d6-4be5-458c-811d-6d5e87699640',
'leaseOffset': 0, 'path': u'gluvol4/4dabb6d
6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'volumeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath':
u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e8769964
0/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease',
'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}, {'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/
images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a',
'volumeID': u'43dbb053-c5fe-45bf-9464-acf77546b96a', 'leasePath':
u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-83
27-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease', 'imageID':
u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} from=::ffff:10.4.8.242,45784,
flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
task_id=b825b7c7-ddae-441f-b9c8-2bd50ec144b5 (api:52)
2019-07-18 09:29:09,790+0200 INFO  (jsonrpc/2) [vds] prepared volume
path:
gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98
(clientIF:497)
2019-07-18 09:29:09,791+0200 INFO  (jsonrpc/2) [vdsm.api] START
teardownImage(sdUUID=u'4dabb6d6-4be5-458c-811d-6d5e87699640',
spUUID=u'00000001-0001-0001-0001-000000000311',
imgUUID=u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', volUUID=None)
from=::ffff:10.4.8.242,45784, flow
_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:46)

2019-07-18 09:29:09,792+0200 INFO  (jsonrpc/2) [storage.StorageDomain]
Removing image rundir link
u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50'
(fileSD:600)

2019-07-18 09:29:09,792+0200 INFO  (jsonrpc/2) [vdsm.api] FINISH
teardownImage return=None from=::ffff:10.4.8.242,45784,
flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:52)

2019-07-18 09:29:09,792+0200 ERROR (jsonrpc/2) [virt.vm]
(vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c') Unable to start
replication for sda to {u'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'volumeInfo': {'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/imag
es/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'type': 'network', 'hosts': [{'port': '0', 'transport': 'tcp', 'name':
'192.168.11.20'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv01'},
{'port': '0', 'transport': 'tcp', 'name': 'glusr
v02'}, {'port': '0', 'transport': 'tcp', 'name': 'glusrv03'}],
'protocol': 'gluster'}, 'format': 'cow', u'poolID':
u'00000001-0001-0001-0001-000000000311', u'device': 'disk', 'protocol':
'gluster', 'propagateErrors': 'off', u'diskType': u'network', 'cache':
'none', u'vol
umeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', u'imageID':
u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', 'hosts': [{'port': '0',
'transport': 'tcp', 'name': '192.168.11.20'}], 'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50
/62656632-8984-4b7e-8be1-fd2547ca0f98', 'volumeChain': [{'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'volu
meID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath':
u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease',
'imageID': u'd2964ff9-10f7-
4b92-8327-d68f3cfd5b50'}, {'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0, 'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a',
'volumeID': u'43dbb053-c5fe-45bf-94
64-acf77546b96a', 'leasePath':
u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease',
'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} (vm
:4710)
Traceback (most recent call last):
   File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4704,
in diskReplicateStart
     self._startDriveReplication(drive)
   File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4843,
in _startDriveReplication
     self._dom.blockCopy(drive.name, destxml, flags=flags)
   File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line
98, in f
     ret = attr(*args, **kwargs)
   File
"/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py",
line 130, in wrapper
     ret = f(*args, **kwargs)
   File "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line
92, in wrapper
     return func(inst, *args, **kwargs)
   File "/usr/lib64/python2.7/site-packages/libvirt.py", line 729, in
blockCopy
     if ret == -1: raise libvirtError ('virDomainBlockCopy() failed',
dom=self)
libvirtError: argument unsupported: non-file destination not supported yet

2019-07-18 09:29:09,796+0200 INFO  (jsonrpc/2) [api.virt] FINISH
diskReplicateStart return={'status': {'message': 'Drive replication
error', 'code': 55}} from=::ffff:10.4.8.242,45784,
flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
vmId=3b79d0c0-47e9-47c3-8511-980a8cfe147c
  (api:52)


On 18.07.19 10:42, Benny Zlotnik wrote:
> It should work, what is the engine and vdsm versions?
> Can you add vdsm logs as well?
>
> On Thu, Jul 18, 2019 at 11:16 AM Christoph Köhler
> <koehler@luis.uni-hannover.de <mailto:koehler@luis.uni-hannover.de>> wrote:
>
>     Hello,
>
>     I try to migrate a disk of a running vm from gluster 3.12.15 to gluster
>     3.12.15 but it fails. libGfApi set to true by engine-config.
>
>     ° taking a snapshot first, is working. Then at engine-log:
>
>     2019-07-18 09:29:13,932+02 ERROR
>     [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed in
>     'VmReplicateDiskStartVDS' method
>
>     2019-07-18 09:29:13,936+02 ERROR
>     [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] EVENT_ID:
>     VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ovvirt07 command
>     VmReplicateDiskStartVDS failed: Drive replication error
>
>     2019-07-18 09:29:13,936+02 INFO
>     [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command
>     'org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand'
>
>     return value 'StatusOnlyReturn [status=Status [code=55, message=Drive
>     replication error]]'
>
>     2019-07-18 09:29:13,936+02 INFO
>     [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] HostName = ovvirt07
>
>     2019-07-18 09:29:13,937+02 ERROR
>     [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command
>     'VmReplicateDiskStartVDSCommand(HostName = ovvirt07,
>     VmReplicateDiskParameters:{hostId='3a7bf85c-e92d-4559-908e-5eed2f5608d4',
>
>     vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c',
>     storagePoolId='00000001-0001-0001-0001-000000000311',
>     srcStorageDomainId='e54d835a-d8a5-44ae-8e17-fcba1c54e46f',
>     targetStorageDomainId='4dabb6d6-4be5-458c-811d-6d5e87699640',
>     imageGroupId='d2964ff9-10f7-4b92-8327-d68f3cfd5b50',
>     imageId='62656632-8984-4b7e-8be1-fd2547ca0f98'})' execution failed:
>     VDSGenericException: VDSErrorException: Failed to
>     VmReplicateDiskStartVDS, error = Drive replication error, code = 55
>
>     2019-07-18 09:29:13,937+02 INFO
>     [org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] FINISH,
>     VmReplicateDiskStartVDSCommand, return: , log id: 5b2afb0b
>
>     2019-07-18 09:29:13,937+02 ERROR
>     [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed VmReplicateDiskStart
>     (Disk
>     'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' , VM
>     '3b79d0c0-47e9-47c3-8511-980a8cfe147c')
>
>     2019-07-18 09:29:13,938+02 ERROR
>     [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id:
>     '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' with children
>     [03672b60-443b-47ba-834c-ac306d7129d0,
>     562522fc-6691-47fe-93bf-ef2c45e85676] failed when attempting to perform
>     the next operation, marking as 'ACTIVE'
>
>     2019-07-18 09:29:13,938+02 ERROR
>     [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] EngineException: Drive
>     replication error (Failed with error replicaErr and code 55):
>     org.ovirt.engine.core.common.errors.EngineException: EngineException:
>     Drive replication error (Failed with error replicaErr and code 55)
>               at
>     org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.replicateDiskStart(LiveMigrateDiskCommand.java:526)
>
>     [bll.jar:]
>               at
>     org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.performNextOperation(LiveMigrateDiskCommand.java:233)
>
>     [bll.jar:]
>               at
>     org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback.childCommandsExecutionEnded(SerialChildCommandsExecutionCallback.java:32)
>
>     [bll.jar:]
>               at
>     org.ovirt.engine.core.bll.ChildCommandsCallbackBase.doPolling(ChildCommandsCallbackBase.java:77)
>
>     [bll.jar:]
>               at
>     org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethodsImpl(CommandCallbacksPoller.java:175)
>
>     [bll.jar:]
>               at
>     org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethods(CommandCallbacksPoller.java:109)
>
>     [bll.jar:]
>               at
>     java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>     [rt.jar:1.8.0_212]
>               at
>     java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
>     [rt.jar:1.8.0_212]
>               at
>     org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.access$201(ManagedScheduledThreadPoolExecutor.java:383)
>
>     [javax.enterprise.concurrent-1.0.jar:]
>               at
>     org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.run(ManagedScheduledThreadPoolExecutor.java:534)
>
>     [javax.enterprise.concurrent-1.0.jar:]
>               at
>     java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>
>     [rt.jar:1.8.0_212]
>               at
>     java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>
>     [rt.jar:1.8.0_212]
>               at java.lang.Thread.run(Thread.java:748) [rt.jar:1.8.0_212]
>               at
>     org.glassfish.enterprise.concurrent.ManagedThreadFactoryImpl$ManagedThread.run(ManagedThreadFactoryImpl.java:250)
>
>     [javax.enterprise.concurrent-1.0.jar:]
>               at
>     org.jboss.as.ee.concurrent.service.ElytronManagedThreadFactory$ElytronManagedThread.run(ElytronManagedThreadFactory.java:78)
>
>     2019-07-18 09:29:13,938+02 INFO
>     [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
>     (EE-ManagedThreadFactory-engineScheduled-Thread-84)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id:
>     '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' child commands
>     '[03672b60-443b-47ba-834c-ac306d7129d0,
>     562522fc-6691-47fe-93bf-ef2c45e85676]' executions were completed,
>     status
>     'FAILED'
>
>     2019-07-18 09:29:15,019+02 ERROR
>     [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
>     (EE-ManagedThreadFactory-engineScheduled-Thread-1)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Ending command
>     'org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand' with
>     failure.
>
>     2019-07-18 09:29:15,019+02 ERROR
>     [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
>     (EE-ManagedThreadFactory-engineScheduled-Thread-1)
>     [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed during live storage
>     migration of disk 'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' of vm
>     '3b79d0c0-47e9-47c3-8511-980a8cfe147c', attempting to end replication
>     before deleting the target disk
>
>     //
>
>     LiveStorageMigration on gluster - should that work at all? Has someone
>     tried it?
>
>     Greetings!
>     Christoph Köhler
>     _______________________________________________
>     Users mailing list -- users@ovirt.org <mailto:users@ovirt.org>
>     To unsubscribe send an email to users-leave@ovirt.org
>     <mailto:users-leave@ovirt.org>
>     Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>     oVirt Code of Conduct:
>     https://www.ovirt.org/community/about/community-guidelines/
>     List Archives:
>     https://lists.ovirt.org/archives/list/users@ovirt.org/message/WSF45K47AJNSAJWYSAMSRKW34T5MWRWZ/
>