On Thu, Jul 18, 2019 at 12:05 PM Christoph Köhler <
koehler(a)luis.uni-hannover.de> wrote:
The engine runs on version 4.3.4
The hypervisors run on 4.2.8, (version vdsm: 4.20.46-1.el7)
with libvirt 4.5.0-10.el7_6.7
By vdsm.log :
2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.StorageDomain]
Creating domain run directory
u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640' (fileSD:577)
2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.fileUtils]
Creating directory:
/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640 mode: None
(fileUtils:197)
2019-07-18 09:29:09,744+0200 INFO (jsonrpc/2) [storage.StorageDomain]
Creating symlink from
/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50
to /var/run/vdsm/storage/4dabb6d6-4be5-458
c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50 (fileSD:580)
2019-07-18 09:29:09,789+0200 INFO (jsonrpc/2) [vdsm.api] FINISH
prepareImage return={'info': {'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'type': 'network', 'hosts': [{'port':
'0', 'transport': 'tcp', 'name':
'192.168.11.20'}, {'port': '0',
'transport': 'tcp', 'name': 'glusrv01'}, {'port':
'0', 'transport':
'tcp', 'name': 'glusrv02'}, {'port': '0',
'transport': 'tcp', 'name':
'glusrv03'}], 'protocol': 'gluster'}, 'path':
u'/rhev/data-center/mn
t/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'imgVolumesInfo': [{'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640',
'leaseOffset': 0, 'path': u'gluvol4/4dabb6d
6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'volumeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98',
'leasePath':
u'/rhev/data-center/mnt/glusterSD/192.168.11.20:
_gluvol4/4dabb6d6-4be5-458c-811d-6d5e8769964
0/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease',
'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'},
{'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0,
'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/
images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a',
'volumeID': u'43dbb053-c5fe-45bf-9464-acf77546b96a',
'leasePath':
u'/rhev/data-center/mnt/glusterSD/192.168.11.20:
_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-83
27-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease', 'imageID':
u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} from=::ffff:10.4.8.242,45784,
flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
task_id=b825b7c7-ddae-441f-b9c8-2bd50ec144b5 (api:52)
2019-07-18 09:29:09,790+0200 INFO (jsonrpc/2) [vds] prepared volume
path:
gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98
(clientIF:497)
2019-07-18 09:29:09,791+0200 INFO (jsonrpc/2) [vdsm.api] START
teardownImage(sdUUID=u'4dabb6d6-4be5-458c-811d-6d5e87699640',
spUUID=u'00000001-0001-0001-0001-000000000311',
imgUUID=u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', volUUID=None)
from=::ffff:10.4.8.242,45784, flow
_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:46)
2019-07-18 09:29:09,792+0200 INFO (jsonrpc/2) [storage.StorageDomain]
Removing image rundir link
u'/var/run/vdsm/storage/4dabb6d6-4be5-458c-811d-6d5e87699640/d2964ff9-10f7-4b92-8327-d68f3cfd5b50'
(fileSD:600)
2019-07-18 09:29:09,792+0200 INFO (jsonrpc/2) [vdsm.api] FINISH
teardownImage return=None from=::ffff:10.4.8.242,45784,
flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
task_id=c13f7ee9-8d01-4b73-8200-3052b4953385 (api:52)
2019-07-18 09:29:09,792+0200 ERROR (jsonrpc/2) [virt.vm]
(vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c') Unable to start
replication for sda to {u'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'volumeInfo': {'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/imag
es/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'type': 'network', 'hosts': [{'port': '0',
'transport': 'tcp', 'name':
'192.168.11.20'}, {'port': '0', 'transport':
'tcp', 'name': 'glusrv01'},
{'port': '0', 'transport': 'tcp', 'name':
'glusr
v02'}, {'port': '0', 'transport': 'tcp',
'name': 'glusrv03'}],
'protocol': 'gluster'}, 'format': 'cow',
u'poolID':
u'00000001-0001-0001-0001-000000000311', u'device': 'disk',
'protocol':
'gluster', 'propagateErrors': 'off', u'diskType':
u'network', 'cache':
'none', u'vol
umeID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', u'imageID':
u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50', 'hosts': [{'port':
'0',
'transport': 'tcp', 'name': '192.168.11.20'}],
'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50
/62656632-8984-4b7e-8be1-fd2547ca0f98', 'volumeChain': [{'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0,
'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98',
'volu
meID': u'62656632-8984-4b7e-8be1-fd2547ca0f98', 'leasePath':
u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/62656632-8984-4b7e-8be1-fd2547ca0f98.lease',
'imageID': u'd2964ff9-10f7-
4b92-8327-d68f3cfd5b50'}, {'domainID':
u'4dabb6d6-4be5-458c-811d-6d5e87699640', 'leaseOffset': 0,
'path':
u'gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a',
'volumeID': u'43dbb053-c5fe-45bf-94
64-acf77546b96a', 'leasePath':
u'/rhev/data-center/mnt/glusterSD/192.168.11.20:_gluvol4/4dabb6d6-4be5-458c-811d-6d5e87699640/images/d2964ff9-10f7-4b92-8327-d68f3cfd5b50/43dbb053-c5fe-45bf-9464-acf77546b96a.lease',
'imageID': u'd2964ff9-10f7-4b92-8327-d68f3cfd5b50'}]} (vm
:4710)
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4704,
in diskReplicateStart
self._startDriveReplication(drive)
File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 4843,
in _startDriveReplication
self._dom.blockCopy(drive.name, destxml, flags=flags)
File "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line
98, in f
ret = attr(*args, **kwargs)
File
"/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py",
line 130, in wrapper
ret = f(*args, **kwargs)
File "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line
92, in wrapper
return func(inst, *args, **kwargs)
File "/usr/lib64/python2.7/site-packages/libvirt.py", line 729, in
blockCopy
if ret == -1: raise libvirtError ('virDomainBlockCopy() failed',
dom=self)
libvirtError: argument unsupported: non-file destination not supported yet
2019-07-18 09:29:09,796+0200 INFO (jsonrpc/2) [api.virt] FINISH
diskReplicateStart return={'status': {'message': 'Drive replication
error', 'code': 55}} from=::ffff:10.4.8.242,45784,
flow_id=c957e011-37e0-43aa-abe7-9bb633c38c5f,
vmId=3b79d0c0-47e9-47c3-8511-980a8cfe147c
(api:52)
On 18.07.19 10:42, Benny Zlotnik wrote:
> It should work, what is the engine and vdsm versions?
> Can you add vdsm logs as well?
>
> On Thu, Jul 18, 2019 at 11:16 AM Christoph Köhler
> <koehler(a)luis.uni-hannover.de <mailto:koehler@luis.uni-hannover.de>>
wrote:
>
> Hello,
>
> I try to migrate a disk of a running vm from gluster 3.12.15 to
gluster
> 3.12.15 but it fails. libGfApi set to true by engine-config.
>
> ° taking a snapshot first, is working. Then at engine-log:
>
> 2019-07-18 09:29:13,932+02 ERROR
>
[org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed in
> 'VmReplicateDiskStartVDS' method
>
> 2019-07-18 09:29:13,936+02 ERROR
>
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] EVENT_ID:
> VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ovvirt07 command
> VmReplicateDiskStartVDS failed: Drive replication error
>
> 2019-07-18 09:29:13,936+02 INFO
>
[org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command
>
'org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand'
>
> return value 'StatusOnlyReturn [status=Status [code=55, message=Drive
> replication error]]'
>
> 2019-07-18 09:29:13,936+02 INFO
>
[org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] HostName = ovvirt07
>
> 2019-07-18 09:29:13,937+02 ERROR
>
[org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command
> 'VmReplicateDiskStartVDSCommand(HostName = ovvirt07,
>
VmReplicateDiskParameters:{hostId='3a7bf85c-e92d-4559-908e-5eed2f5608d4',
>
> vmId='3b79d0c0-47e9-47c3-8511-980a8cfe147c',
> storagePoolId='00000001-0001-0001-0001-000000000311',
> srcStorageDomainId='e54d835a-d8a5-44ae-8e17-fcba1c54e46f',
> targetStorageDomainId='4dabb6d6-4be5-458c-811d-6d5e87699640',
> imageGroupId='d2964ff9-10f7-4b92-8327-d68f3cfd5b50',
> imageId='62656632-8984-4b7e-8be1-fd2547ca0f98'})' execution failed:
> VDSGenericException: VDSErrorException: Failed to
> VmReplicateDiskStartVDS, error = Drive replication error, code = 55
>
> 2019-07-18 09:29:13,937+02 INFO
>
[org.ovirt.engine.core.vdsbroker.vdsbroker.VmReplicateDiskStartVDSCommand]
>
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] FINISH,
> VmReplicateDiskStartVDSCommand, return: , log id: 5b2afb0b
>
> 2019-07-18 09:29:13,937+02 ERROR
> [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed VmReplicateDiskStart
> (Disk
> 'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' , VM
> '3b79d0c0-47e9-47c3-8511-980a8cfe147c')
>
> 2019-07-18 09:29:13,938+02 ERROR
> [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id:
> '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' with children
> [03672b60-443b-47ba-834c-ac306d7129d0,
> 562522fc-6691-47fe-93bf-ef2c45e85676] failed when attempting to
perform
> the next operation, marking as 'ACTIVE'
>
> 2019-07-18 09:29:13,938+02 ERROR
> [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] EngineException: Drive
> replication error (Failed with error replicaErr and code 55):
> org.ovirt.engine.core.common.errors.EngineException: EngineException:
> Drive replication error (Failed with error replicaErr and code 55)
> at
>
org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.replicateDiskStart(LiveMigrateDiskCommand.java:526)
>
> [bll.jar:]
> at
>
org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand.performNextOperation(LiveMigrateDiskCommand.java:233)
>
> [bll.jar:]
> at
>
org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback.childCommandsExecutionEnded(SerialChildCommandsExecutionCallback.java:32)
>
> [bll.jar:]
> at
>
org.ovirt.engine.core.bll.ChildCommandsCallbackBase.doPolling(ChildCommandsCallbackBase.java:77)
>
> [bll.jar:]
> at
>
org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethodsImpl(CommandCallbacksPoller.java:175)
>
> [bll.jar:]
> at
>
org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethods(CommandCallbacksPoller.java:109)
>
> [bll.jar:]
> at
>
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> [rt.jar:1.8.0_212]
> at
> java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
> [rt.jar:1.8.0_212]
> at
>
org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.access$201(ManagedScheduledThreadPoolExecutor.java:383)
>
> [javax.enterprise.concurrent-1.0.jar:]
> at
>
org.glassfish.enterprise.concurrent.internal.ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.run(ManagedScheduledThreadPoolExecutor.java:534)
>
> [javax.enterprise.concurrent-1.0.jar:]
> at
>
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>
> [rt.jar:1.8.0_212]
> at
>
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>
> [rt.jar:1.8.0_212]
> at java.lang.Thread.run(Thread.java:748) [rt.jar:1.8.0_212]
> at
>
org.glassfish.enterprise.concurrent.ManagedThreadFactoryImpl$ManagedThread.run(ManagedThreadFactoryImpl.java:250)
>
> [javax.enterprise.concurrent-1.0.jar:]
> at
>
org.jboss.as.ee.concurrent.service.ElytronManagedThreadFactory$ElytronManagedThread.run(ElytronManagedThreadFactory.java:78)
>
> 2019-07-18 09:29:13,938+02 INFO
> [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
> (EE-ManagedThreadFactory-engineScheduled-Thread-84)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] Command 'LiveMigrateDisk' id:
> '8174c74c-8ab0-49fa-abfc-44d8b7c691e0' child commands
> '[03672b60-443b-47ba-834c-ac306d7129d0,
> 562522fc-6691-47fe-93bf-ef2c45e85676]' executions were completed,
> status
> 'FAILED'
>
> 2019-07-18 09:29:15,019+02 ERROR
> [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
> (EE-ManagedThreadFactory-engineScheduled-Thread-1)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] Ending command
> 'org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand' with
> failure.
>
> 2019-07-18 09:29:15,019+02 ERROR
> [org.ovirt.engine.core.bll.storage.lsm.LiveMigrateDiskCommand]
> (EE-ManagedThreadFactory-engineScheduled-Thread-1)
> [c957e011-37e0-43aa-abe7-9bb633c38c5f] Failed during live storage
> migration of disk 'd2964ff9-10f7-4b92-8327-d68f3cfd5b50' of vm
> '3b79d0c0-47e9-47c3-8511-980a8cfe147c', attempting to end replication
> before deleting the target disk
>
> //
>
> LiveStorageMigration on gluster - should that work at all? Has
someone
> tried it?
>
> Greetings!
> Christoph Köhler
> _______________________________________________
> Users mailing list -- users(a)ovirt.org <mailto:users@ovirt.org>
> To unsubscribe send an email to users-leave(a)ovirt.org
> <mailto:users-leave@ovirt.org>
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WSF45K47AJN...
>