On Tue, 2019-07-09 at 08:00 +0100, Dan Poltawski wrote:
I've now managed to succesfully create/mount/delete volumes!
However, I'm seeing live migrations stay stuck. Is this supported?
(gdb) py-list
345 client.conf_set('rados_osd_op_timeout',
timeout)
346 client.conf_set('rados_mon_op_timeout',
timeout)
347 client.conf_set('client_mount_timeout',
timeout)
348
349 client.connect()
350 ioctx = client.open_ioctx(pool)
351 return client, ioctx
352 except self.rados.Error:
353 msg = _("Error connecting to ceph cluster.")
354 LOG.exception(msg)
355 client.shutdown()
(gdb) py-bt
#15 Frame 0x3ea0e50, for file /usr/lib/python2.7/site-
packages/cinder/volume/drivers/rbd.py, line 350, in _do_conn
(pool='storage-ssd', remote=None, timeout=-1, name='ceph',
conf='/etc/ceph/ceph.conf', user='ovirt', client=<rados.Rados at
remote
0x7fb1f4f83a60>)
ioctx = client.open_ioctx(pool)
#20 Frame 0x3ea4620, for file /usr/lib/python2.7/site-
packages/retrying.py, line 217, in call
(self=<Retrying(_retry_on_exception=<function at remote
0x7fb1f4f23488>, _wait_exponential_max=1073741823,
_wait_incrementing_start=0, stop=<function at remote 0x7fb1f4f23578>,
_stop_max_attempt_number=5, _wait_incrementing_increment=100,
_wait_random_max=1000, _retry_on_result=<instancemethod at remote
0x7fb1f51da550>, _stop_max_delay=100, _wait_fixed=1000,
_wrap_exception=False, _wait_random_min=0,
_wait_exponential_multiplier=1, wait=<function at remote
0x7fb1f4f23500>) at remote 0x7fb1f4f1ae90>, fn=<function at remote
0x7fb1f4f23668>, args=(None, None, None), kwargs={},
start_time=1562658179214, attempt_number=1)
attempt = Attempt(fn(*args, **kwargs), attempt_number, False)
#25 Frame 0x3e49d50, for file /usr/lib/python2.7/site-
packages/cinder/utils.py, line 818, in _wrapper (args=(None, None,
None), kwargs={}, r=<Retrying(_retry_on_exception=<function at remote
0x7fb1f4f23488>, _wait_exponential_max=1073741823,
_wait_incrementing_start=0, stop=<function at remote 0x7fb1f4f23578>,
_stop_max_attempt_number=5, _wait_incrementing_increment=100,
_wait_random_max=1000, _retry_on_result=<instancemethod at remote
0x7fb1f51da550>, _stop_max_delay=100, _wait_fixed=1000,
_wrap_exception=False, _wait_random_min=0,
_wait_exponential_multiplier=1, wait=<function at remote
0x7fb1f4f23500>) at remote 0x7fb1f4f1ae90>)
return r.call(f, *args, **kwargs)
#29 Frame 0x7fb1f4f9a810, for file /usr/lib/python2.7/site-
packages/cinder/volume/drivers/rbd.py, line 358, in _connect_to_rados
(self=<RBDDriver(_target_names=[], rbd=<module at remote
0x7fb20583e830>, _is_replication_enabled=False, _execute=<function at
remote 0x7fb2041242a8>, _active_config={'name': 'ceph',
'conf':
'/etc/ceph/ceph.conf', 'user': 'ovirt'},
_active_backend_id=None,
_initialized=False, db=<DBAPI(_backend=<module at remote
0x7fb203f8d520>, qos_specs_get=<instancemethod at remote
0x7fb1f677d460>, _lock=<Semaphore(counter=1,
_waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote
0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20,
'inc_retry_interval': True, 'retry_interval': 1,
'max_retry_interval':
10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'},
_backend_name='sqlalchemy', use_db_reconnect=False,
get_by_id=<instancemethod at remote 0x7fb1f61d8050>,
volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote
0x7fb2003aab10>, target_mapping={'tgtadm': 'cinder.vol...(truncated)
return _do_conn(pool, remote, timeout)
#33 Frame 0x7fb1f4f5b220, for file /usr/lib/python2.7/site-
packages/cinder/volume/drivers/rbd.py, line 177, in __init__
(self=<RADOSClient(driver=<RBDDriver(_target_names=[], rbd=<module at
remote 0x7fb20583e830>, _is_replication_enabled=False,
_execute=<function at remote 0x7fb2041242a8>, _active_config={'name':
'ceph', 'conf': '/etc/ceph/ceph.conf', 'user':
'ovirt'},
_active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module
at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote
0x7fb1f677d460>, _lock=<Semaphore(counter=1,
_waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote
0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20,
'inc_retry_interval': True, 'retry_interval': 1,
'max_retry_interval':
10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'},
_backend_name='sqlalchemy', use_db_reconnect=False,
get_by_id=<instancemethod at remote 0x7fb1f61d8050>,
volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote
0x7fb2003aab10>, target_mapping={'tgtadm': ...(truncated)
self.cluster, self.ioctx = driver._connect_to_rados(pool)
#44 Frame 0x7fb1f4f9a620, for file /usr/lib/python2.7/site-
packages/cinder/volume/drivers/rbd.py, line 298, in
check_for_setup_error (self=<RBDDriver(_target_names=[], rbd=<module at
remote 0x7fb20583e830>, _is_replication_enabled=False,
_execute=<function at remote 0x7fb2041242a8>, _active_config={'name':
'ceph', 'conf': '/etc/ceph/ceph.conf', 'user':
'ovirt'},
_active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module
at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote
0x7fb1f677d460>, _lock=<Semaphore(counter=1,
_waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote
0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20,
'inc_retry_interval': True, 'retry_interval': 1,
'max_retry_interval':
10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'},
_backend_name='sqlalchemy', use_db_reconnect=False,
get_by_id=<instancemethod at remote 0x7fb1f61d8050>,
volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote
0x7fb2003aab10>, target_mapping={'tgtadm': 'cinder...(truncated)
with RADOSClient(self):
#48 Frame 0x3e5bef0, for file /usr/lib/python2.7/site-
packages/cinderlib/cinderlib.py, line 88, in __init__
(self=<Backend(driver=<RBDDriver(_target_names=[], rbd=<module at
remote 0x7fb20583e830>, _is_replication_enabled=False,
_execute=<function at remote 0x7fb2041242a8>, _active_config={'name':
'ceph', 'conf': '/etc/ceph/ceph.conf', 'user':
'ovirt'},
_active_backend_id=None, _initialized=False, db=<DBAPI(_backend=<module
at remote 0x7fb203f8d520>, qos_specs_get=<instancemethod at remote
0x7fb1f677d460>, _lock=<Semaphore(counter=1,
_waiters=<collections.deque at remote 0x7fb1f5246d70>) at remote
0x7fb1f5205bd0>, _wrap_db_kwargs={'max_retries': 20,
'inc_retry_interval': True, 'retry_interval': 1,
'max_retry_interval':
10}, _backend_mapping={'sqlalchemy': 'cinder.db.sqlalchemy.api'},
_backend_name='sqlalchemy', use_db_reconnect=False,
get_by_id=<instancemethod at remote 0x7fb1f61d8050>,
volume_type_get=<instancemethod at remote 0x7fb1f61c0f50>) at remote
0x7fb2003aab10>, target_mapping={'tgtadm':
'cinder.volume.t...(truncated)
self.driver.check_for_setup_error()
#58 Frame 0x3d15cf0, for file ./cinderlib-client.py, line 210, in
load_backend
(args=<Namespace(driver='{"volume_backend_name":"ceph-
storage-ssd-ovirt","volume---Type <return> to continue, or q
<return>
to quit---
_driver":"cinder.volume.drivers.rbd.RBDDriver","rbd_ceph_conf":"/etc/ce
ph/ceph.conf","rbd_user":"ovirt","rbd_keyring_conf":"/etc/ceph/ceph.cli
ent.ovirt.keyring","rbd_pool":"storage-
ssd","use_multipath_for_image_xfer":"true"}',
db_url='postgresql+psycopg2://ovirt_cinderlib:ViRsNB3Dnwy5wmL0lLuDEq@lo
calhost:5432/ovirt_cinderlib', correlation_id='7733e7cc',
command=<function at remote 0x7fb1f52725f0>, volume_id='9d858d39-bbd4-
4cbe-9f2c-5bef25ed0525',
connector_info='{"ip":null,"host":"compute00.hq.tnp.infra","os_type":"l
inux2","platform":"x86_64","initiator":"iqn.1994-
05.com.redhat:c6ab662d439c","multipath":true,"do_local_attach":false}')
at remote 0x7fb1f5215910>, persistence_config={'connection':
'postgresql+psycopg2://ovirt_cinderlib:ViRsNB3Dnwy5wmL0lLuDEq@localhost
:5432/ovirt_cinderlib', 'storage': 'db'})
Python Exception <type 'exceptions.IOError'> [Errno 2] No such file or
directory: './cinderlib-client.py':
Error occurred in Python command: [Errno 2] No such file or directory:
'./cinderlib-client.py'
________________________________
The Networking People (TNP) Limited. Registered office: Network House, Caton Rd,
Lancaster, LA1 3PE. Registered in England & Wales with company number: 07667393
This email and any files transmitted with it are confidential and intended solely for the
use of the individual or entity to whom they are addressed. If you have received this
email in error please notify the system manager. This message contains confidential
information and is intended only for the individual named. If you are not the named
addressee you should not disseminate, distribute or copy this e-mail. Please notify the
sender immediately by e-mail if you have received this e-mail by mistake and delete this
e-mail from your system. If you are not the intended recipient you are notified that
disclosing, copying, distributing or taking any action in reliance on the contents of this
information is strictly prohibited.