On Mon, 2019-07-08 at 16:25 +0300, Benny Zlotnik wrote:
> Hi,
>
> You have a typo, it's py-bt and I just tried it myself, I only had to
> install:
> $ yum install -y python-devel
> (in addition to the packages specified in the link)
Thanks - this is what I get:
#3 Frame 0x7f2046b59ad0, for file /usr/lib/python2.7/site-
packages/eventlet/hubs/epolls.py, line 62, in do_poll
(self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>,
debug_exceptions=True, debug_blocking_resolution=1, modify=<built-in
method modify of select.epoll object at remote 0x7f2048455168>,
running=True, debug_blocking=False, listeners={'read': {20:
<FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet
object at remote 0x7f2046878410>, spent=False,
greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read',
mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in
method throw of greenlet.greenlet object at remote 0x7f2046878410>) at
remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0,
greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[],
stopping=False, timers=[(<float at remote 0x2830e20>,
<Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch
of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}),
called=F...(truncated)
return self.poll.poll(seconds)
#6 Frame 0x32fbf30, for file /usr/lib/python2.7/site-
packages/eventlet/hubs/poll.py, line 85, in wait
(self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>,
debug_exceptions=True, debug_blocking_resolution=1, modify=<built-in
method modify of select.epoll object at remote 0x7f2048455168>,
running=True, debug_blocking=False, listeners={'read': {20:
<FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet
object at remote 0x7f2046878410>, spent=False,
greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read',
mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in
method throw of greenlet.greenlet object at remote 0x7f2046878410>) at
remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0,
greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[],
stopping=False, timers=[(<float at remote 0x2830e20>,
<Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch
of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}),
called=False) at r...(truncated)
presult = self.do_poll(seconds)
#10 Frame 0x7f2046afca00, for file /usr/lib/python2.7/site-
packages/eventlet/hubs/hub.py, line 346, in run
(self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>,
debug_exceptions=True, debug_blocking_resolution=1, modify=<built-in
method modify of select.epoll object at remote 0x7f2048455168>,
running=True, debug_blocking=False, listeners={'read': {20:
<FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet
object at remote 0x7f2046878410>, spent=False,
greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read',
mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in
method throw of greenlet.greenlet object at remote 0x7f2046878410>) at
remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0,
greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[],
stopping=False, timers=[(<float at remote 0x2830e20>,
<Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch
of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}),
called=False) ...(truncated)
self.wait(sleep_time)
>
> On Mon, Jul 8, 2019 at 2:40 PM Dan Poltawski <
> dan.poltawski@tnp.net.uk> wrote:
> > Hi,
> >
> > On Sun, 2019-07-07 at 09:31 +0300, Benny Zlotnik wrote:
> > > > Any chance you can setup gdb[1] so we can find out where it's
> > > stuck
> > > > exactly?
> >
> > Yes, abolutely - but I will need some assistance in getting GDB
> > configured in the engine as I am not very familar with it - or how
> > to enable the correct repos to get the debug info.
> >
> > $ gdb python 54654
> >
> > [...]
> >
> > Reading symbols from /lib64/libfreeblpriv3.so...Reading symbols
> > from /lib64/libfreeblpriv3.so...(no debugging symbols
> > found)...done.
> > (no debugging symbols found)...done.
> > Loaded symbols for /lib64/libfreeblpriv3.so
> > 0x00007fcf82256483 in epoll_wait () from /lib64/libc.so.6
> > Missing separate debuginfos, use: debuginfo-install python-2.7.5-
> > 80.el7_6.x86_64
> > (gdb) pt-bt
> > Undefined command: "pt-bt". Try "help".
> >
> >
> > > > Also, which version of ovirt are you using?
> >
> > Using 4.3.4
> >
> > > > Can you also check the ceph logs for anything suspicious?
> >
> > I haven't seen anything so far, but is an entirely resonable
> > possibility this is ceph misoconfiguraiton as we are learning about
> > both tools.
> >
> >
> > thanks,
> >
> > Dan
> >
> > > >
> > > >
> > > > [1] - https://wiki.python.org/moin/DebuggingWithGdb
> > > > $ gdb python <cinderlib pid>
> > > > then `py-bt`
> > > >
> > > > On Thu, Jul 4, 2019 at 7:00 PM <dan.poltawski@tnp.net.uk>
> > > wrote:
> > > > > > > Can you provide logs? mainly engine.log and cinderlib.log
> > > > > > > (/var/log/ovirt-engine/cinderlib/cinderlib.log
> > > > > >
> > > > > >
> > > > > > If I create two volumes, the first one succeeds
> > > > successfully, the
> > > > > > second one hangs. If I look in the processlist after
> > > > creating the
> > > > > > second volume which doesn't succceed, I see the python
> > > > ./cinderlib-
> > > > > > client.py create_volume [...] command still running.
> > > > > >
> > > > > > On the ceph side, I can see only the one rbd volume.
> > > > > >
> > > > > > Logs below:
> > > > > >
> > > > > >
> > > > > >
> > > > > > --- cinderlib.log --
> > > > > >
> > > > > > 2019-07-04 16:46:30,863 - cinderlib-client - INFO - Fetch
> > > > backend
> > > > > > stats [b07698bb-1688-472f-841b-70a9d52a250d]
> > > > > > 2019-07-04 16:46:56,308 - cinderlib-client - INFO -
> > > > Creating volume
> > > > > > '236285cc-ac01-4239-821c-4beadd66923f', with size '2' GB
> > > > [0b0f0d6f-
> > > > > > cb20-440a-bacb-7f5ead2b4b4d]
> > > > > > 2019-07-04 16:47:21,671 - cinderlib-client - INFO -
> > > > Creating volume
> > > > > > '84886485-554a-44ca-964c-9758b4a16aae', with size '2' GB
> > > > [a793bfc9-
> > > > > > fc37-4711-a144-d74c100cc75b]
> > > > > >
> > > > > > --- engine.log ---
> > > > > >
> > > > > > 2019-07-04 16:46:54,062+01 INFO
> > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > (default
> > > > > > task-22) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running
> > > > command:
> > > > > > AddDiskCommand internal: false. Entities affected : ID:
> > > > 31536d80-
> > > > > > ff45-496b-9820-15441d505924 Type: StorageAction group
> > > > CREATE_DISK
> > > > > > with role type USER
> > > > > > 2019-07-04 16:46:54,150+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > Bloc
> > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > commandCoordinator-
> > > > > > Thread-1) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running
> > > > command:
> > > > > > AddManagedBlockStorageDiskCommand internal: true.
> > > > > > 2019-07-04 16:46:56,863+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.common.utils.cinderlib.CinderlibExecutor
> > > > ]
> > > > > > (EE-ManagedThreadFactory-commandCoordinator-Thread-1)
> > > > [0b0f0d6f-
> > > > > > cb20-440a-bacb-7f5ead2b4b4d] cinderlib output:
> > > > > > 2019-07-04 16:46:56,912+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi
> > > > rect
> > > > > > or] (default task-22) [] EVENT_ID:
> > > > > > USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk 'test0' was
> > > > > > successfully added.
> > > > > > 2019-07-04 16:47:00,126+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCall
> > > > back
> > > > > > ] (EE-ManagedThreadFactory-engineScheduled-Thread-95)
> > > > [0b0f0d6f-
> > > > > > cb20-440a-bacb-7f5ead2b4b4d] Command 'AddDisk' id:
> > > > '15fe157d-7adb-
> > > > > > 4031-9e81-f51aa0b6528f' child commands '[d056397a-7ed9-
> > > > 4c01-b880-
> > > > > > dd518421a2c6]' executions were completed, status
> > > > 'SUCCEEDED'
> > > > > > 2019-07-04 16:47:01,136+01 INFO
> > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineScheduled-Thread-99) [0b0f0d6f-
> > > > cb20-
> > > > > > 440a-bacb-7f5ead2b4b4d] Ending command
> > > > > > 'org.ovirt.engine.core.bll.storage.disk.AddDiskCommand'
> > > > > > successfully.
> > > > > > 2019-07-04 16:47:01,141+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > Bloc
> > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > engineScheduled-
> > > > > > Thread-99) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Ending
> > > > command
> > > > > >
> > > > 'org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > Bloc
> > > > > > kStorageDiskCommand' successfully.
> > > > > > 2019-07-04 16:47:01,145+01 WARN
> > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineScheduled-Thread-99) [] VM is
> > > > null - no
> > > > > > unlocking
> > > > > > 2019-07-04 16:47:01,186+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi
> > > > rect
> > > > > > or] (EE-ManagedThreadFactory-engineScheduled-Thread-99) []
> > > > > > EVENT_ID: USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk
> > > > 'test0'
> > > > > > was successfully added.
> > > > > > 2019-07-04 16:47:19,446+01 INFO
> > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > (default
> > > > > > task-22) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running
> > > > command:
> > > > > > AddDiskCommand internal: false. Entities affected : ID:
> > > > 31536d80-
> > > > > > ff45-496b-9820-15441d505924 Type: StorageAction group
> > > > CREATE_DISK
> > > > > > with role type USER
> > > > > > 2019-07-04 16:47:19,464+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > Bloc
> > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > commandCoordinator-
> > > > > > Thread-2) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running
> > > > command:
> > > > > > AddManagedBlockStorageDiskCommand internal: true.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'commandCoordinator' is using 1 threads out of 10, 1
> > > > threads
> > > > > > waiting for tasks.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'default' is using 0 threads out of 1, 5 threads
> > > > waiting for
> > > > > > tasks.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'engine' is using 0 threads out of 500, 9 threads
> > > > waiting for
> > > > > > tasks and 0 tasks in queue.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'engineScheduled' is using 0 threads out of 100, 100
> > > > threads
> > > > > > waiting for tasks.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'engineThreadMonitoring' is using 1 threads out of 1,
> > > > 0
> > > > > > threads waiting for tasks.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'hostUpdatesChecker' is using 0 threads out of 5, 1
> > > > threads
> > > > > > waiting for tasks.
> > > > > >
> > > > > >
> > > > > >
> > > > > > 2019-07-04 16:46:54,062+01 INFO
> > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > (default
> > > > > > task-22) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running
> > > > command:
> > > > > > AddDiskCommand internal: false. Entities affected : ID:
> > > > 31536d80-
> > > > > > ff45-496b-9820-15441d505924 Type: StorageAction group
> > > > CREATE_DISK
> > > > > > with role type USER
> > > > > > 2019-07-04 16:46:54,150+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > Bloc
> > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > commandCoordinator-
> > > > > > Thread-1) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running
> > > > command:
> > > > > > AddManagedBlockStorageDiskCommand internal: true.
> > > > > > 2019-07-04 16:46:56,863+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.common.utils.cinderlib.CinderlibExecutor
> > > > ]
> > > > > > (EE-ManagedThreadFactory-commandCoordinator-Thread-1)
> > > > [0b0f0d6f-
> > > > > > cb20-440a-bacb-7f5ead2b4b4d] cinderlib output:
> > > > > > 2019-07-04 16:46:56,912+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi
> > > > rect
> > > > > > or] (default task-22) [] EVENT_ID:
> > > > > > USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk 'test0' was
> > > > > > successfully added.
> > > > > > 2019-07-04 16:47:00,126+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCall
> > > > back
> > > > > > ] (EE-ManagedThreadFactory-engineScheduled-Thread-95)
> > > > [0b0f0d6f-
> > > > > > cb20-440a-bacb-7f5ead2b4b4d] Command 'AddDisk' id:
> > > > '15fe157d-7adb-
> > > > > > 4031-9e81-f51aa0b6528f' child commands '[d056397a-7ed9-
> > > > 4c01-b880-
> > > > > > dd518421a2c6]' executions were completed, status
> > > > 'SUCCEEDED'
> > > > > > 2019-07-04 16:47:01,136+01 INFO
> > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineScheduled-Thread-99) [0b0f0d6f-
> > > > cb20-
> > > > > > 440a-bacb-7f5ead2b4b4d] Ending command
> > > > > > 'org.ovirt.engine.core.bll.storage.disk.AddDiskCommand'
> > > > > > successfully.
> > > > > > 2019-07-04 16:47:01,141+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > Bloc
> > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > engineScheduled-
> > > > > > Thread-99) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Ending
> > > > command
> > > > > >
> > > > 'org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > Bloc
> > > > > > kStorageDiskCommand' successfully.
> > > > > > 2019-07-04 16:47:01,145+01 WARN
> > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineScheduled-Thread-99) [] VM is
> > > > null - no
> > > > > > unlocking
> > > > > > 2019-07-04 16:47:01,186+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi
> > > > rect
> > > > > > or] (EE-ManagedThreadFactory-engineScheduled-Thread-99) []
> > > > > > EVENT_ID: USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk
> > > > 'test0'
> > > > > > was successfully added.
> > > > > > 2019-07-04 16:47:19,446+01 INFO
> > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > (default
> > > > > > task-22) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running
> > > > command:
> > > > > > AddDiskCommand internal: false. Entities affected : ID:
> > > > 31536d80-
> > > > > > ff45-496b-9820-15441d505924 Type: StorageAction group
> > > > CREATE_DISK
> > > > > > with role type USER
> > > > > > 2019-07-04 16:47:19,464+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > Bloc
> > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > commandCoordinator-
> > > > > > Thread-2) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running
> > > > command:
> > > > > > AddManagedBlockStorageDiskCommand internal: true.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'commandCoordinator' is using 1 threads out of 10, 1
> > > > threads
> > > > > > waiting for tasks.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'default' is using 0 threads out of 1, 5 threads
> > > > waiting for
> > > > > > tasks.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'engine' is using 0 threads out of 500, 9 threads
> > > > waiting for
> > > > > > tasks and 0 tasks in queue.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'engineScheduled' is using 0 threads out of 100, 100
> > > > threads
> > > > > > waiting for tasks.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'engineThreadMonitoring' is using 1 threads out of 1,
> > > > 0
> > > > > > threads waiting for tasks.
> > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > >
> > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > (EE-
> > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > Thread
> > > > > > pool 'hostUpdatesChecker' is using 0 threads out of 5, 1
> > > > threads
> > > > > > waiting for tasks.
> > > > > > _______________________________________________
> > > > > > Users mailing list -- users@ovirt.org
> > > > > > To unsubscribe send an email to users-leave@ovirt.org
> > > > > > Privacy Statement:
> > > > https://www.ovirt.org/site/privacy-policy/
> > > > > > oVirt Code of Conduct:
> > > > > > https://www.ovirt.org/community/about/community-guidelines/
> > > > > > List Archives:
> > > > > >
> > > > https://lists.ovirt.org/archives/list/users@ovirt.org/message/VW5AZHXGU2YTSBBYJI4YEUNBSI6CDXUS/
> >
> > The Networking People (TNP) Limited. Registered office: Network
> > House, Caton Rd, Lancaster, LA1 3PE. Registered in England & Wales
> > with company number: 07667393
> > This email and any files transmitted with it are confidential and
> > intended solely for the use of the individual or entity to whom
> > they are addressed. If you have received this email in error please
> > notify the system manager. This message contains confidential
> > information and is intended only for the individual named. If you
> > are not the named addressee you should not disseminate, distribute
> > or copy this e-mail. Please notify the sender immediately by e-mail
> > if you have received this e-mail by mistake and delete this e-mail
> > from your system. If you are not the intended recipient you are
> > notified that disclosing, copying, distributing or taking any
> > action in reliance on the contents of this information is strictly
> > prohibited.
________________________________
The Networking People (TNP) Limited. Registered office: Network House, Caton Rd, Lancaster, LA1 3PE. Registered in England & Wales with company number: 07667393
This email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you have received this email in error please notify the system manager. This message contains confidential information and is intended only for the individual named. If you are not the named addressee you should not disseminate, distribute or copy this e-mail. Please notify the sender immediately by e-mail if you have received this e-mail by mistake and delete this e-mail from your system. If you are not the intended recipient you are notified that disclosing, copying, distributing or taking any action in reliance on the contents of this information is strictly prohibited.