
Can you try to create mutliple ceph volumes manually via rbd from the engine machine, so we can simulate what cinderlib does without using it, this can be done $ rbd -c ceph.conf create <pool_name>/vol1 --size 100M $ rbd -c ceph.conf create <pool_name>/vol2 --size 100M On Mon, Jul 8, 2019 at 4:58 PM Dan Poltawski <dan.poltawski@tnp.net.uk> wrote:
On Mon, 2019-07-08 at 16:49 +0300, Benny Zlotnik wrote:
Not too useful unfortunately :\ Can you try py-list instead of py-bt? Perhaps it will provide better results
(gdb) py-list 57 if get_errno(ex) != errno.EEXIST: 58 raise 59 return listener 60 61 def do_poll(self, seconds):
62 return self.poll.poll(seconds)
Thanks for you help,
Dan
On Mon, Jul 8, 2019 at 4:41 PM Dan Poltawski < dan.poltawski@tnp.net.uk> wrote:
On Mon, 2019-07-08 at 16:25 +0300, Benny Zlotnik wrote:
Hi,
You have a typo, it's py-bt and I just tried it myself, I only had to install: $ yum install -y python-devel (in addition to the packages specified in the link)
Thanks - this is what I get:
#3 Frame 0x7f2046b59ad0, for file /usr/lib/python2.7/site- packages/eventlet/hubs/epolls.py, line 62, in do_poll (self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>, debug_exceptions=True, debug_blocking_resolution=1, modify=<built- in method modify of select.epoll object at remote 0x7f2048455168>, running=True, debug_blocking=False, listeners={'read': {20: <FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet object at remote 0x7f2046878410>, spent=False, greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read', mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in method throw of greenlet.greenlet object at remote 0x7f2046878410>) at remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0, greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[], stopping=False, timers=[(<float at remote 0x2830e20>, <Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}), called=F...(truncated) return self.poll.poll(seconds) #6 Frame 0x32fbf30, for file /usr/lib/python2.7/site- packages/eventlet/hubs/poll.py, line 85, in wait (self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>, debug_exceptions=True, debug_blocking_resolution=1, modify=<built- in method modify of select.epoll object at remote 0x7f2048455168>, running=True, debug_blocking=False, listeners={'read': {20: <FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet object at remote 0x7f2046878410>, spent=False, greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read', mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in method throw of greenlet.greenlet object at remote 0x7f2046878410>) at remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0, greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[], stopping=False, timers=[(<float at remote 0x2830e20>, <Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}), called=False) at r...(truncated) presult = self.do_poll(seconds) #10 Frame 0x7f2046afca00, for file /usr/lib/python2.7/site- packages/eventlet/hubs/hub.py, line 346, in run (self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>, debug_exceptions=True, debug_blocking_resolution=1, modify=<built- in method modify of select.epoll object at remote 0x7f2048455168>, running=True, debug_blocking=False, listeners={'read': {20: <FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet object at remote 0x7f2046878410>, spent=False, greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read', mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in method throw of greenlet.greenlet object at remote 0x7f2046878410>) at remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0, greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[], stopping=False, timers=[(<float at remote 0x2830e20>, <Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}), called=False) ...(truncated) self.wait(sleep_time)
On Mon, Jul 8, 2019 at 2:40 PM Dan Poltawski < dan.poltawski@tnp.net.uk> wrote:
Hi,
On Sun, 2019-07-07 at 09:31 +0300, Benny Zlotnik wrote:
> Any chance you can setup gdb[1] so we can find out where
it's
stuck > exactly?
Yes, abolutely - but I will need some assistance in getting GDB configured in the engine as I am not very familar with it - or how to enable the correct repos to get the debug info.
$ gdb python 54654
[...]
Reading symbols from /lib64/libfreeblpriv3.so...Reading symbols from /lib64/libfreeblpriv3.so...(no debugging symbols found)...done. (no debugging symbols found)...done. Loaded symbols for /lib64/libfreeblpriv3.so 0x00007fcf82256483 in epoll_wait () from /lib64/libc.so.6 Missing separate debuginfos, use: debuginfo-install python- 2.7.5- 80.el7_6.x86_64 (gdb) pt-bt Undefined command: "pt-bt". Try "help".
> Also, which version of ovirt are you using?
Using 4.3.4
> Can you also check the ceph logs for anything suspicious?
I haven't seen anything so far, but is an entirely resonable possibility this is ceph misoconfiguraiton as we are learning about both tools.
thanks,
Dan
> > > [1] - https://wiki.python.org/moin/DebuggingWithGdb > $ gdb python <cinderlib pid> > then `py-bt` > > On Thu, Jul 4, 2019 at 7:00 PM <dan.poltawski@tnp.net.uk> wrote: > > > > Can you provide logs? mainly engine.log and cinderlib.log > > > > (/var/log/ovirt-engine/cinderlib/cinderlib.log > > > > > > > > > If I create two volumes, the first one succeeds > successfully, the > > > second one hangs. If I look in the processlist after > creating the > > > second volume which doesn't succceed, I see the python > ./cinderlib- > > > client.py create_volume [...] command still running. > > > > > > On the ceph side, I can see only the one rbd volume. > > > > > > Logs below: > > > > > > > > > > > > --- cinderlib.log -- > > > > > > 2019-07-04 16:46:30,863 - cinderlib-client - INFO - Fetch > backend > > > stats [b07698bb-1688-472f-841b-70a9d52a250d] > > > 2019-07-04 16:46:56,308 - cinderlib-client - INFO - > Creating volume > > > '236285cc-ac01-4239-821c-4beadd66923f', with size '2' GB > [0b0f0d6f- > > > cb20-440a-bacb-7f5ead2b4b4d] > > > 2019-07-04 16:47:21,671 - cinderlib-client - INFO - > Creating volume > > > '84886485-554a-44ca-964c-9758b4a16aae', with size '2' GB > [a793bfc9- > > > fc37-4711-a144-d74c100cc75b] > > > > > > --- engine.log --- > > > > > > 2019-07-04 16:46:54,062+01 INFO > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand] > (default > > > task-22) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running > command: > > > AddDiskCommand internal: false. Entities affected : ID: > 31536d80- > > > ff45-496b-9820-15441d505924 Type: StorageAction group > CREATE_DISK > > > with role type USER > > > 2019-07-04 16:46:54,150+01 INFO > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged > Bloc > > > kStorageDiskCommand] (EE-ManagedThreadFactory- > commandCoordinator- > > > Thread-1) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running > command: > > > AddManagedBlockStorageDiskCommand internal: true. > > > 2019-07-04 16:46:56,863+01 INFO > > > > [org.ovirt.engine.core.common.utils.cinderlib.CinderlibExecutor > ] > > > (EE-ManagedThreadFactory-commandCoordinator-Thread-1) > [0b0f0d6f- > > > cb20-440a-bacb-7f5ead2b4b4d] cinderlib output: > > > 2019-07-04 16:46:56,912+01 INFO > > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi > rect > > > or] (default task-22) [] EVENT_ID: > > > USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk 'test0' was > > > successfully added. > > > 2019-07-04 16:47:00,126+01 INFO > > > > [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCall > back > > > ] (EE-ManagedThreadFactory-engineScheduled-Thread-95) > [0b0f0d6f- > > > cb20-440a-bacb-7f5ead2b4b4d] Command 'AddDisk' id: > '15fe157d-7adb- > > > 4031-9e81-f51aa0b6528f' child commands '[d056397a-7ed9- > 4c01-b880- > > > dd518421a2c6]' executions were completed, status > 'SUCCEEDED' > > > 2019-07-04 16:47:01,136+01 INFO > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand] > (EE- > > > ManagedThreadFactory-engineScheduled-Thread-99) [0b0f0d6f- > cb20- > > > 440a-bacb-7f5ead2b4b4d] Ending command > > > 'org.ovirt.engine.core.bll.storage.disk.AddDiskCommand' > > > successfully. > > > 2019-07-04 16:47:01,141+01 INFO > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged > Bloc > > > kStorageDiskCommand] (EE-ManagedThreadFactory- > engineScheduled- > > > Thread-99) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Ending > command > > > > 'org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged > Bloc > > > kStorageDiskCommand' successfully. > > > 2019-07-04 16:47:01,145+01 WARN > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand] > (EE- > > > ManagedThreadFactory-engineScheduled-Thread-99) [] VM is > null - no > > > unlocking > > > 2019-07-04 16:47:01,186+01 INFO > > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi > rect > > > or] (EE-ManagedThreadFactory-engineScheduled-Thread-99) [] > > > EVENT_ID: USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk > 'test0' > > > was successfully added. > > > 2019-07-04 16:47:19,446+01 INFO > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand] > (default > > > task-22) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running > command: > > > AddDiskCommand internal: false. Entities affected : ID: > 31536d80- > > > ff45-496b-9820-15441d505924 Type: StorageAction group > CREATE_DISK > > > with role type USER > > > 2019-07-04 16:47:19,464+01 INFO > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged > Bloc > > > kStorageDiskCommand] (EE-ManagedThreadFactory- > commandCoordinator- > > > Thread-2) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running > command: > > > AddManagedBlockStorageDiskCommand internal: true. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'commandCoordinator' is using 1 threads out of 10, 1 > threads > > > waiting for tasks. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'default' is using 0 threads out of 1, 5 threads > waiting for > > > tasks. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'engine' is using 0 threads out of 500, 9 threads > waiting for > > > tasks and 0 tasks in queue. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'engineScheduled' is using 0 threads out of 100, 100 > threads > > > waiting for tasks. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'engineThreadMonitoring' is using 1 threads out of 1, > 0 > > > threads waiting for tasks. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'hostUpdatesChecker' is using 0 threads out of 5, 1 > threads > > > waiting for tasks. > > > > > > > > > > > > 2019-07-04 16:46:54,062+01 INFO > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand] > (default > > > task-22) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running > command: > > > AddDiskCommand internal: false. Entities affected : ID: > 31536d80- > > > ff45-496b-9820-15441d505924 Type: StorageAction group > CREATE_DISK > > > with role type USER > > > 2019-07-04 16:46:54,150+01 INFO > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged > Bloc > > > kStorageDiskCommand] (EE-ManagedThreadFactory- > commandCoordinator- > > > Thread-1) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running > command: > > > AddManagedBlockStorageDiskCommand internal: true. > > > 2019-07-04 16:46:56,863+01 INFO > > > > [org.ovirt.engine.core.common.utils.cinderlib.CinderlibExecutor > ] > > > (EE-ManagedThreadFactory-commandCoordinator-Thread-1) > [0b0f0d6f- > > > cb20-440a-bacb-7f5ead2b4b4d] cinderlib output: > > > 2019-07-04 16:46:56,912+01 INFO > > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi > rect > > > or] (default task-22) [] EVENT_ID: > > > USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk 'test0' was > > > successfully added. > > > 2019-07-04 16:47:00,126+01 INFO > > > > [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCall > back > > > ] (EE-ManagedThreadFactory-engineScheduled-Thread-95) > [0b0f0d6f- > > > cb20-440a-bacb-7f5ead2b4b4d] Command 'AddDisk' id: > '15fe157d-7adb- > > > 4031-9e81-f51aa0b6528f' child commands '[d056397a-7ed9- > 4c01-b880- > > > dd518421a2c6]' executions were completed, status > 'SUCCEEDED' > > > 2019-07-04 16:47:01,136+01 INFO > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand] > (EE- > > > ManagedThreadFactory-engineScheduled-Thread-99) [0b0f0d6f- > cb20- > > > 440a-bacb-7f5ead2b4b4d] Ending command > > > 'org.ovirt.engine.core.bll.storage.disk.AddDiskCommand' > > > successfully. > > > 2019-07-04 16:47:01,141+01 INFO > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged > Bloc > > > kStorageDiskCommand] (EE-ManagedThreadFactory- > engineScheduled- > > > Thread-99) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Ending > command > > > > 'org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged > Bloc > > > kStorageDiskCommand' successfully. > > > 2019-07-04 16:47:01,145+01 WARN > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand] > (EE- > > > ManagedThreadFactory-engineScheduled-Thread-99) [] VM is > null - no > > > unlocking > > > 2019-07-04 16:47:01,186+01 INFO > > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi > rect > > > or] (EE-ManagedThreadFactory-engineScheduled-Thread-99) [] > > > EVENT_ID: USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk > 'test0' > > > was successfully added. > > > 2019-07-04 16:47:19,446+01 INFO > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand] > (default > > > task-22) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running > command: > > > AddDiskCommand internal: false. Entities affected : ID: > 31536d80- > > > ff45-496b-9820-15441d505924 Type: StorageAction group > CREATE_DISK > > > with role type USER > > > 2019-07-04 16:47:19,464+01 INFO > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged > Bloc > > > kStorageDiskCommand] (EE-ManagedThreadFactory- > commandCoordinator- > > > Thread-2) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running > command: > > > AddManagedBlockStorageDiskCommand internal: true. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'commandCoordinator' is using 1 threads out of 10, 1 > threads > > > waiting for tasks. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'default' is using 0 threads out of 1, 5 threads > waiting for > > > tasks. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'engine' is using 0 threads out of 500, 9 threads > waiting for > > > tasks and 0 tasks in queue. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'engineScheduled' is using 0 threads out of 100, 100 > threads > > > waiting for tasks. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'engineThreadMonitoring' is using 1 threads out of 1, > 0 > > > threads waiting for tasks. > > > 2019-07-04 16:48:19,501+01 INFO > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService] > (EE- > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] > Thread > > > pool 'hostUpdatesChecker' is using 0 threads out of 5, 1 > threads > > > waiting for tasks. > > > _______________________________________________ > > > Users mailing list -- users@ovirt.org > > > To unsubscribe send an email to users-leave@ovirt.org > > > Privacy Statement: > https://www.ovirt.org/site/privacy-policy/ > > > oVirt Code of Conduct: > > > https://www.ovirt.org/community/about/community-guidelines/ > > > List Archives: > > > >
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VW5AZHXGU2YTSB...
The Networking People (TNP) Limited. Registered office: Network House, Caton Rd, Lancaster, LA1 3PE. Registered in England &
with company number: 07667393 This email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you have received this email in error
Wales please
notify the system manager. This message contains confidential information and is intended only for the individual named. If you are not the named addressee you should not disseminate, distribute or copy this e-mail. Please notify the sender immediately by e- mail if you have received this e-mail by mistake and delete this e- mail from your system. If you are not the intended recipient you are notified that disclosing, copying, distributing or taking any action in reliance on the contents of this information is strictly prohibited.
________________________________
The Networking People (TNP) Limited. Registered office: Network House, Caton Rd, Lancaster, LA1 3PE. Registered in England & Wales with company number: 07667393
This email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you have received this email in error please notify the system manager. This message contains confidential information and is intended only for the individual named. If you are not the named addressee you should not disseminate, distribute or copy this e-mail. Please notify the sender immediately by e-mail if you have received this e-mail by mistake and delete this e-mail from your system. If you are not the intended recipient you are notified that disclosing, copying, distributing or taking any action in reliance on the contents of this information is strictly prohibited.
________________________________
The Networking People (TNP) Limited. Registered office: Network House, Caton Rd, Lancaster, LA1 3PE. Registered in England & Wales with company number: 07667393
This email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you have received this email in error please notify the system manager. This message contains confidential information and is intended only for the individual named. If you are not the named addressee you should not disseminate, distribute or copy this e-mail. Please notify the sender immediately by e-mail if you have received this e-mail by mistake and delete this e-mail from your system. If you are not the intended recipient you are notified that disclosing, copying, distributing or taking any action in reliance on the contents of this information is strictly prohibited.