VM live migration is supported and should work
Can you add engine and cinderlib logs?

On Tue, Jul 9, 2019 at 11:01 AM Dan Poltawski <dan.poltawski@tnp.net.uk> wrote:
On Tue, 2019-07-09 at 08:00 +0100, Dan Poltawski wrote:
> I've now managed to succesfully create/mount/delete volumes!

However, I'm seeing live migrations stay stuck. Is this supported?

(gdb) py-list
 345                        client.conf_set('rados_osd_op_timeout',
timeout)
 346                        client.conf_set('rados_mon_op_timeout',
timeout)
 347                        client.conf_set('client_mount_timeout',
timeout)
 348
 349                    client.connect()
>350                    ioctx = client.open_ioctx(pool)
 351                    return client, ioctx
 352                except self.rados.Error:
 353                    msg = _("Error connecting to ceph cluster.")
 354                    LOG.exception(msg)
 355                    client.shutdown()


(gdb) py-bt
#15 Frame 0x3ea0e50, for file /usr/lib/python2.7/site-
packages/cinder/volume/drivers/rbd.py, line 350, in _do_conn
(pool='storage-ssd', remote=None, timeout=-1, name='ceph',
conf='/etc/ceph/ceph.conf', user='ovirt', client=<rados.Rados at remote
0x7fb1f4f83a60>)
    ioctx = client.open_ioctx(pool)
#20 Frame 0x3ea4620, for file /usr/lib/python2.7/site-
packages/retrying.py, line 217, in call
(self=<Retrying(_retry_on_exception=<function at remote
0x7fb1f4f23488>, _wait_exponential_max=1073741823,
_wait_incrementing_start=0, stop=<function at remote 0x7fb1f4f23578>,
_stop_max_attempt_number=5, _wait_incrementing_increment=100,
_wait_random_max=1000, _retry_on_result=<instancemethod at remote
0x7fb1f51da550>, _stop_max_delay=100, _wait_fixed=1000,
_wrap_exception=False, _wait_random_min=0,
_wait_exponential_multiplier=1, wait=<function at remote
0x7fb1f4f23500>) at remote 0x7fb1f4f1ae90>, fn=<function at remote
0x7fb1f4f23668>, args=(None, None, None), kwargs={},
start_time=1562658179214, attempt_number=1)
    attempt = Attempt(fn(*args, **kwargs), attempt_number, False)
#25 Frame 0x3e49d50, for file /usr/lib/python2.7/site-
packages/cinder/utils.py, line 818, in _wrapper (args=(None, None,
None), kwargs={}, r=<Retrying(_retry_on_exception=<function at remote
0x7fb1f4f23488>, _wait_exponential_max=1073741823,
_wait_incrementing_start=0, stop=<function at remote 0x7fb1f4f23578>,
_stop_max_attempt_number=5, _wait_incrementing_increment=100,
_wait_random_max=1000, _retry_on_result=<instancemethod at remote
0x7fb1f51da550>, _stop_max_delay=100, _wait_fixed=1000,
_wrap_exception=False, _wait_random_min=0,
_wait_exponential_multiplier=1, wait=<function at remote
0x7fb1f4f23500>) at remote 0x7fb1f4f1ae90>)
    return r.call(f, *args, **kwargs)
#29 Frame 0x7fb1f4f9a810, for file /usr/lib/python2.7/site-
packages/cinder/volume/drivers/rbd.py, line 358, in _connect_to_rados
(self=<RBDDriver(_target_names=[], rbd=<module at remote
0x7fb20583e830>, _is_replication_enabled=False, _execute=<function at
remote 0x7fb2041242a8>, _active_config={'name': 'ceph', 'conf':
'/etc/ceph/ceph.conf', 'user': 'ovirt'}, _active_backend_id=None,
_initialized=False, db=<DBAPI(_backend=<module at remote
0x7fb203f8d520>, qos_specs_get=<instancemethod at remote
0x7fb1f677d460>, _lock=<Semaphore(counter=1,
_waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote
0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20,
'inc_retry_interval': True, 'retry_interval': 1, 'max_retry_interval':
10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'},
_backend_name='sqlalchemy', use_db_reconnect=False,
get_by_id=<instancemethod at remote 0x7fb1f61d8050>,
volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote
0x7fb2003aab10>, target_mapping={'tgtadm': 'cinder.vol...(truncated)
    return _do_conn(pool, remote, timeout)
#33 Frame 0x7fb1f4f5b220, for file /usr/lib/python2.7/site-
packages/cinder/volume/drivers/rbd.py, line 177, in __init__
(self=<RADOSClient(driver=<RBDDriver(_target_names=[], rbd=<module at
remote 0x7fb20583e830>, _is_replication_enabled=False,
_execute=<function at remote 0x7fb2041242a8>, _active_config={'name':
'ceph', 'conf': '/etc/ceph/ceph.conf', 'user': 'ovirt'},
_active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module
at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote
0x7fb1f677d460>, _lock=<Semaphore(counter=1,
_waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote
0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20,
'inc_retry_interval': True, 'retry_interval': 1, 'max_retry_interval':
10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'},
_backend_name='sqlalchemy', use_db_reconnect=False,
get_by_id=<instancemethod at remote 0x7fb1f61d8050>,
volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote
0x7fb2003aab10>, target_mapping={'tgtadm': ...(truncated)
    self.cluster, self.ioctx = driver._connect_to_rados(pool)
#44 Frame 0x7fb1f4f9a620, for file /usr/lib/python2.7/site-
packages/cinder/volume/drivers/rbd.py, line 298, in
check_for_setup_error (self=<RBDDriver(_target_names=[], rbd=<module at
remote 0x7fb20583e830>, _is_replication_enabled=False,
_execute=<function at remote 0x7fb2041242a8>, _active_config={'name':
'ceph', 'conf': '/etc/ceph/ceph.conf', 'user': 'ovirt'},
_active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module
at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote
0x7fb1f677d460>, _lock=<Semaphore(counter=1,
_waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote
0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20,
'inc_retry_interval': True, 'retry_interval': 1, 'max_retry_interval':
10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'},
_backend_name='sqlalchemy', use_db_reconnect=False,
get_by_id=<instancemethod at remote 0x7fb1f61d8050>,
volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote
0x7fb2003aab10>, target_mapping={'tgtadm': 'cinder...(truncated)
    with RADOSClient(self):
#48 Frame 0x3e5bef0, for file /usr/lib/python2.7/site-
packages/cinderlib/cinderlib.py, line 88, in __init__
(self=<Backend(driver=<RBDDriver(_target_names=[], rbd=<module at
remote 0x7fb20583e830>, _is_replication_enabled=False,
_execute=<function at remote 0x7fb2041242a8>, _active_config={'name':
'ceph', 'conf': '/etc/ceph/ceph.conf', 'user': 'ovirt'},
_active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module
at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote
0x7fb1f677d460>, _lock=<Semaphore(counter=1,
_waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote
0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20,
'inc_retry_interval': True, 'retry_interval': 1, 'max_retry_interval':
10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'},
_backend_name='sqlalchemy', use_db_reconnect=False,
get_by_id=<instancemethod at remote 0x7fb1f61d8050>,
volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote
0x7fb2003aab10>, target_mapping={'tgtadm':
'cinder.volume.t...(truncated)
    self.driver.check_for_setup_error()
#58 Frame 0x3d15cf0, for file ./cinderlib-client.py, line 210, in
load_backend (args=<Namespace(driver='{"volume_backend_name":"ceph-
storage-ssd-ovirt","volume---Type <return> to continue, or q <return>
to quit---
_driver":"cinder.volume.drivers.rbd.RBDDriver","rbd_ceph_conf":"/etc/ce
ph/ceph.conf","rbd_user":"ovirt","rbd_keyring_conf":"/etc/ceph/ceph.cli
ent.ovirt.keyring","rbd_pool":"storage-
ssd","use_multipath_for_image_xfer":"true"}',
db_url='postgresql+psycopg2://ovirt_cinderlib:ViRsNB3Dnwy5wmL0lLuDEq@lo
calhost:5432/ovirt_cinderlib', correlation_id='7733e7cc',
command=<function at remote 0x7fb1f52725f0>, volume_id='9d858d39-bbd4-
4cbe-9f2c-5bef25ed0525',
connector_info='{"ip":null,"host":"compute00.hq.tnp.infra","os_type":"l
inux2","platform":"x86_64","initiator":"iqn.1994-
05.com.redhat:c6ab662d439c","multipath":true,"do_local_attach":false}')
at remote 0x7fb1f5215910>, persistence_config={'connection':
'postgresql+psycopg2://ovirt_cinderlib:ViRsNB3Dnwy5wmL0lLuDEq@localhost
:5432/ovirt_cinderlib', 'storage': 'db'})
Python Exception <type 'exceptions.IOError'> [Errno 2] No such file or
directory: './cinderlib-client.py':
Error occurred in Python command: [Errno 2] No such file or directory:
'./cinderlib-client.py'




________________________________

The Networking People (TNP) Limited. Registered office: Network House, Caton Rd, Lancaster, LA1 3PE. Registered in England & Wales with company number: 07667393

This email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you have received this email in error please notify the system manager. This message contains confidential information and is intended only for the individual named. If you are not the named addressee you should not disseminate, distribute or copy this e-mail. Please notify the sender immediately by e-mail if you have received this e-mail by mistake and delete this e-mail from your system. If you are not the intended recipient you are notified that disclosing, copying, distributing or taking any action in reliance on the contents of this information is strictly prohibited.