Not too useful unfortunately :\
Can you try py-list instead of py-bt? Perhaps it will provide better results

On Mon, Jul 8, 2019 at 4:41 PM Dan Poltawski <dan.poltaw...@tnp.net.uk>
wrote:

> On Mon, 2019-07-08 at 16:25 +0300, Benny Zlotnik wrote:
> > Hi,
> >
> > You have a typo, it's py-bt and I just tried it myself, I only had to
> > install:
> > $ yum install -y python-devel
> > (in addition to the packages specified in the link)
>
> Thanks - this is what I get:
>
> #3 Frame 0x7f2046b59ad0, for file /usr/lib/python2.7/site-
> packages/eventlet/hubs/epolls.py, line 62, in do_poll
> (self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>,
> debug_exceptions=True, debug_blocking_resolution=1, modify=<built-in
> method modify of select.epoll object at remote 0x7f2048455168>,
> running=True, debug_blocking=False, listeners={'read': {20:
> <FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet
> object at remote 0x7f2046878410>, spent=False,
> greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read',
> mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in
> method throw of greenlet.greenlet object at remote 0x7f2046878410>) at
> remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0,
> greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[],
> stopping=False, timers=[(<float at remote 0x2830e20>,
> <Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch
> of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}),
> called=F...(truncated)
>     return self.poll.poll(seconds)
> #6 Frame 0x32fbf30, for file /usr/lib/python2.7/site-
> packages/eventlet/hubs/poll.py, line 85, in wait
> (self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>,
> debug_exceptions=True, debug_blocking_resolution=1, modify=<built-in
> method modify of select.epoll object at remote 0x7f2048455168>,
> running=True, debug_blocking=False, listeners={'read': {20:
> <FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet
> object at remote 0x7f2046878410>, spent=False,
> greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read',
> mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in
> method throw of greenlet.greenlet object at remote 0x7f2046878410>) at
> remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0,
> greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[],
> stopping=False, timers=[(<float at remote 0x2830e20>,
> <Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch
> of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}),
> called=False) at r...(truncated)
>     presult = self.do_poll(seconds)
> #10 Frame 0x7f2046afca00, for file /usr/lib/python2.7/site-
> packages/eventlet/hubs/hub.py, line 346, in run
> (self=<Hub(next_timers=[], clock=<function at remote 0x7f20661059b0>,
> debug_exceptions=True, debug_blocking_resolution=1, modify=<built-in
> method modify of select.epoll object at remote 0x7f2048455168>,
> running=True, debug_blocking=False, listeners={'read': {20:
> <FdListener(fileno=20, cb=<built-in method switch of greenlet.greenlet
> object at remote 0x7f2046878410>, spent=False,
> greenlet=<greenlet.greenlet at remote 0x7f2046878410>, evtype='read',
> mark_as_closed=<instancemethod at remote 0x7f2046b09500>, tb=<built-in
> method throw of greenlet.greenlet object at remote 0x7f2046878410>) at
> remote 0x7f20468cda10>}, 'write': {}}, timers_canceled=0,
> greenlet=<greenlet.greenlet at remote 0x7f2046af70f0>, closed=[],
> stopping=False, timers=[(<float at remote 0x2830e20>,
> <Timer(seconds=<float at remote 0xced610>, tpl=(<built-in method switch
> of greenlet.greenlet object at remote 0x7f2046934c30>, (), {}),
> called=False) ...(truncated)
>     self.wait(sleep_time)
>
>
>
> >
> > On Mon, Jul 8, 2019 at 2:40 PM Dan Poltawski <
> > dan.poltaw...@tnp.net.uk> wrote:
> > > Hi,
> > >
> > > On Sun, 2019-07-07 at 09:31 +0300, Benny Zlotnik wrote:
> > > > > Any chance you can setup gdb[1] so we can find out where it's
> > > > stuck
> > > > > exactly?
> > >
> > > Yes, abolutely - but I will need some assistance in getting GDB
> > > configured in the engine as I am not very familar with it - or how
> > > to enable the correct repos to get the debug info.
> > >
> > > $ gdb python 54654
> > >
> > > [...]
> > >
> > > Reading symbols from /lib64/libfreeblpriv3.so...Reading symbols
> > > from /lib64/libfreeblpriv3.so...(no debugging symbols
> > > found)...done.
> > > (no debugging symbols found)...done.
> > > Loaded symbols for /lib64/libfreeblpriv3.so
> > > 0x00007fcf82256483 in epoll_wait () from /lib64/libc.so.6
> > > Missing separate debuginfos, use: debuginfo-install python-2.7.5-
> > > 80.el7_6.x86_64
> > > (gdb) pt-bt
> > > Undefined command: "pt-bt".  Try "help".
> > >
> > >
> > > > > Also, which version of ovirt are you using?
> > >
> > > Using 4.3.4
> > >
> > > > > Can you also check the ceph logs for anything suspicious?
> > >
> > > I haven't seen anything so far, but is an entirely resonable
> > > possibility this is ceph misoconfiguraiton as we are learning about
> > > both tools.
> > >
> > >
> > > thanks,
> > >
> > > Dan
> > >
> > > > >
> > > > >
> > > > > [1] - https://wiki.python.org/moin/DebuggingWithGdb
> > > > > $ gdb python <cinderlib pid>
> > > > > then `py-bt`
> > > > >
> > > > > On Thu, Jul 4, 2019 at 7:00 PM <dan.poltaw...@tnp.net.uk>
> > > > wrote:
> > > > > > > > Can you provide logs? mainly engine.log and cinderlib.log
> > > > > > > > (/var/log/ovirt-engine/cinderlib/cinderlib.log
> > > > > > >
> > > > > > >
> > > > > > > If I create two volumes, the first one succeeds
> > > > > successfully, the
> > > > > > > second one hangs. If I look in the processlist after
> > > > > creating the
> > > > > > > second volume which doesn't succceed, I see the python
> > > > > ./cinderlib-
> > > > > > > client.py create_volume [...] command still running.
> > > > > > >
> > > > > > > On the ceph side, I can see only the one rbd volume.
> > > > > > >
> > > > > > > Logs below:
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > --- cinderlib.log --
> > > > > > >
> > > > > > > 2019-07-04 16:46:30,863 - cinderlib-client - INFO - Fetch
> > > > > backend
> > > > > > > stats [b07698bb-1688-472f-841b-70a9d52a250d]
> > > > > > > 2019-07-04 16:46:56,308 - cinderlib-client - INFO -
> > > > > Creating volume
> > > > > > > '236285cc-ac01-4239-821c-4beadd66923f', with size '2' GB
> > > > > [0b0f0d6f-
> > > > > > > cb20-440a-bacb-7f5ead2b4b4d]
> > > > > > > 2019-07-04 16:47:21,671 - cinderlib-client - INFO -
> > > > > Creating volume
> > > > > > > '84886485-554a-44ca-964c-9758b4a16aae', with size '2' GB
> > > > > [a793bfc9-
> > > > > > > fc37-4711-a144-d74c100cc75b]
> > > > > > >
> > > > > > > --- engine.log ---
> > > > > > >
> > > > > > > 2019-07-04 16:46:54,062+01 INFO
> > > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > > (default
> > > > > > > task-22) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running
> > > > > command:
> > > > > > > AddDiskCommand internal: false. Entities affected :  ID:
> > > > > 31536d80-
> > > > > > > ff45-496b-9820-15441d505924 Type: StorageAction group
> > > > > CREATE_DISK
> > > > > > > with role type USER
> > > > > > > 2019-07-04 16:46:54,150+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > > Bloc
> > > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > > commandCoordinator-
> > > > > > > Thread-1) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running
> > > > > command:
> > > > > > > AddManagedBlockStorageDiskCommand internal: true.
> > > > > > > 2019-07-04 16:46:56,863+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.common.utils.cinderlib.CinderlibExecutor
> > > > > ]
> > > > > > > (EE-ManagedThreadFactory-commandCoordinator-Thread-1)
> > > > > [0b0f0d6f-
> > > > > > > cb20-440a-bacb-7f5ead2b4b4d] cinderlib output:
> > > > > > > 2019-07-04 16:46:56,912+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi
> > > > > rect
> > > > > > > or] (default task-22) [] EVENT_ID:
> > > > > > > USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk 'test0' was
> > > > > > > successfully added.
> > > > > > > 2019-07-04 16:47:00,126+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCall
> > > > > back
> > > > > > > ] (EE-ManagedThreadFactory-engineScheduled-Thread-95)
> > > > > [0b0f0d6f-
> > > > > > > cb20-440a-bacb-7f5ead2b4b4d] Command 'AddDisk' id:
> > > > > '15fe157d-7adb-
> > > > > > > 4031-9e81-f51aa0b6528f' child commands '[d056397a-7ed9-
> > > > > 4c01-b880-
> > > > > > > dd518421a2c6]' executions were completed, status
> > > > > 'SUCCEEDED'
> > > > > > > 2019-07-04 16:47:01,136+01 INFO
> > > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineScheduled-Thread-99) [0b0f0d6f-
> > > > > cb20-
> > > > > > > 440a-bacb-7f5ead2b4b4d] Ending command
> > > > > > > 'org.ovirt.engine.core.bll.storage.disk.AddDiskCommand'
> > > > > > > successfully.
> > > > > > > 2019-07-04 16:47:01,141+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > > Bloc
> > > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > > engineScheduled-
> > > > > > > Thread-99) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Ending
> > > > > command
> > > > > > >
> > > > > 'org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > > Bloc
> > > > > > > kStorageDiskCommand' successfully.
> > > > > > > 2019-07-04 16:47:01,145+01 WARN
> > > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineScheduled-Thread-99) [] VM is
> > > > > null - no
> > > > > > > unlocking
> > > > > > > 2019-07-04 16:47:01,186+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi
> > > > > rect
> > > > > > > or] (EE-ManagedThreadFactory-engineScheduled-Thread-99) []
> > > > > > > EVENT_ID: USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk
> > > > > 'test0'
> > > > > > > was successfully added.
> > > > > > > 2019-07-04 16:47:19,446+01 INFO
> > > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > > (default
> > > > > > > task-22) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running
> > > > > command:
> > > > > > > AddDiskCommand internal: false. Entities affected :  ID:
> > > > > 31536d80-
> > > > > > > ff45-496b-9820-15441d505924 Type: StorageAction group
> > > > > CREATE_DISK
> > > > > > > with role type USER
> > > > > > > 2019-07-04 16:47:19,464+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > > Bloc
> > > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > > commandCoordinator-
> > > > > > > Thread-2) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running
> > > > > command:
> > > > > > > AddManagedBlockStorageDiskCommand internal: true.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'commandCoordinator' is using 1 threads out of 10, 1
> > > > > threads
> > > > > > > waiting for tasks.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'default' is using 0 threads out of 1, 5 threads
> > > > > waiting for
> > > > > > > tasks.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'engine' is using 0 threads out of 500, 9 threads
> > > > > waiting for
> > > > > > > tasks and 0 tasks in queue.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'engineScheduled' is using 0 threads out of 100, 100
> > > > > threads
> > > > > > > waiting for tasks.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'engineThreadMonitoring' is using 1 threads out of 1,
> > > > > 0
> > > > > > > threads waiting for tasks.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'hostUpdatesChecker' is using 0 threads out of 5, 1
> > > > > threads
> > > > > > > waiting for tasks.
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > 2019-07-04 16:46:54,062+01 INFO
> > > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > > (default
> > > > > > > task-22) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running
> > > > > command:
> > > > > > > AddDiskCommand internal: false. Entities affected :  ID:
> > > > > 31536d80-
> > > > > > > ff45-496b-9820-15441d505924 Type: StorageAction group
> > > > > CREATE_DISK
> > > > > > > with role type USER
> > > > > > > 2019-07-04 16:46:54,150+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > > Bloc
> > > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > > commandCoordinator-
> > > > > > > Thread-1) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Running
> > > > > command:
> > > > > > > AddManagedBlockStorageDiskCommand internal: true.
> > > > > > > 2019-07-04 16:46:56,863+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.common.utils.cinderlib.CinderlibExecutor
> > > > > ]
> > > > > > > (EE-ManagedThreadFactory-commandCoordinator-Thread-1)
> > > > > [0b0f0d6f-
> > > > > > > cb20-440a-bacb-7f5ead2b4b4d] cinderlib output:
> > > > > > > 2019-07-04 16:46:56,912+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi
> > > > > rect
> > > > > > > or] (default task-22) [] EVENT_ID:
> > > > > > > USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk 'test0' was
> > > > > > > successfully added.
> > > > > > > 2019-07-04 16:47:00,126+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCall
> > > > > back
> > > > > > > ] (EE-ManagedThreadFactory-engineScheduled-Thread-95)
> > > > > [0b0f0d6f-
> > > > > > > cb20-440a-bacb-7f5ead2b4b4d] Command 'AddDisk' id:
> > > > > '15fe157d-7adb-
> > > > > > > 4031-9e81-f51aa0b6528f' child commands '[d056397a-7ed9-
> > > > > 4c01-b880-
> > > > > > > dd518421a2c6]' executions were completed, status
> > > > > 'SUCCEEDED'
> > > > > > > 2019-07-04 16:47:01,136+01 INFO
> > > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineScheduled-Thread-99) [0b0f0d6f-
> > > > > cb20-
> > > > > > > 440a-bacb-7f5ead2b4b4d] Ending command
> > > > > > > 'org.ovirt.engine.core.bll.storage.disk.AddDiskCommand'
> > > > > > > successfully.
> > > > > > > 2019-07-04 16:47:01,141+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > > Bloc
> > > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > > engineScheduled-
> > > > > > > Thread-99) [0b0f0d6f-cb20-440a-bacb-7f5ead2b4b4d] Ending
> > > > > command
> > > > > > >
> > > > > 'org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > > Bloc
> > > > > > > kStorageDiskCommand' successfully.
> > > > > > > 2019-07-04 16:47:01,145+01 WARN
> > > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineScheduled-Thread-99) [] VM is
> > > > > null - no
> > > > > > > unlocking
> > > > > > > 2019-07-04 16:47:01,186+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDi
> > > > > rect
> > > > > > > or] (EE-ManagedThreadFactory-engineScheduled-Thread-99) []
> > > > > > > EVENT_ID: USER_ADD_DISK_FINISHED_SUCCESS(2,021), The disk
> > > > > 'test0'
> > > > > > > was successfully added.
> > > > > > > 2019-07-04 16:47:19,446+01 INFO
> > > > > > > [org.ovirt.engine.core.bll.storage.disk.AddDiskCommand]
> > > > > (default
> > > > > > > task-22) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running
> > > > > command:
> > > > > > > AddDiskCommand internal: false. Entities affected :  ID:
> > > > > 31536d80-
> > > > > > > ff45-496b-9820-15441d505924 Type: StorageAction group
> > > > > CREATE_DISK
> > > > > > > with role type USER
> > > > > > > 2019-07-04 16:47:19,464+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.storage.disk.managedblock.AddManaged
> > > > > Bloc
> > > > > > > kStorageDiskCommand] (EE-ManagedThreadFactory-
> > > > > commandCoordinator-
> > > > > > > Thread-2) [a793bfc9-fc37-4711-a144-d74c100cc75b] Running
> > > > > command:
> > > > > > > AddManagedBlockStorageDiskCommand internal: true.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'commandCoordinator' is using 1 threads out of 10, 1
> > > > > threads
> > > > > > > waiting for tasks.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'default' is using 0 threads out of 1, 5 threads
> > > > > waiting for
> > > > > > > tasks.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'engine' is using 0 threads out of 500, 9 threads
> > > > > waiting for
> > > > > > > tasks and 0 tasks in queue.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'engineScheduled' is using 0 threads out of 100, 100
> > > > > threads
> > > > > > > waiting for tasks.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'engineThreadMonitoring' is using 1 threads out of 1,
> > > > > 0
> > > > > > > threads waiting for tasks.
> > > > > > > 2019-07-04 16:48:19,501+01 INFO
> > > > > > >
> > > > > [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
> > > > > (EE-
> > > > > > > ManagedThreadFactory-engineThreadMonitoring-Thread-1) []
> > > > > Thread
> > > > > > > pool 'hostUpdatesChecker' is using 0 threads out of 5, 1
> > > > > threads
> > > > > > > waiting for tasks.
> > > > > > > _______________________________________________
> > > > > > > Users mailing list -- users@ovirt.org
> > > > > > > To unsubscribe send an email to users-le...@ovirt.org
> > > > > > > Privacy Statement:
> > > > > https://www.ovirt.org/site/privacy-policy/
> > > > > > > oVirt Code of Conduct:
> > > > > > > https://www.ovirt.org/community/about/community-guidelines/
> > > > > > > List Archives:
> > > > > > >
> > > > >
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VW5AZHXGU2YTSBBYJI4YEUNBSI6CDXUS/
> > >
> > > The Networking People (TNP) Limited. Registered office: Network
> > > House, Caton Rd, Lancaster, LA1 3PE. Registered in England & Wales
> > > with company number: 07667393
> > > This email and any files transmitted with it are confidential and
> > > intended solely for the use of the individual or entity to whom
> > > they are addressed. If you have received this email in error please
> > > notify the system manager. This message contains confidential
> > > information and is intended only for the individual named. If you
> > > are not the named addressee you should not disseminate, distribute
> > > or copy this e-mail. Please notify the sender immediately by e-mail
> > > if you have received this e-mail by mistake and delete this e-mail
> > > from your system. If you are not the intended recipient you are
> > > notified that disclosing, copying, distributing or taking any
> > > action in reliance on the contents of this information is strictly
> > > prohibited.
>
> ________________________________
>
> The Networking People (TNP) Limited. Registered office: Network House,
> Caton Rd, Lancaster, LA1 3PE. Registered in England & Wales with company
> number: 07667393
>
> This email and any files transmitted with it are confidential and intended
> solely for the use of the individual or entity to whom they are addressed.
> If you have received this email in error please notify the system manager.
> This message contains confidential information and is intended only for the
> individual named. If you are not the named addressee you should not
> disseminate, distribute or copy this e-mail. Please notify the sender
> immediately by e-mail if you have received this e-mail by mistake and
> delete this e-mail from your system. If you are not the intended recipient
> you are notified that disclosing, copying, distributing or taking any
> action in reliance on the contents of this information is strictly
> prohibited.
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MSWD6OQBCXFYM7CJR5EDXRQKIKHRLETH/

Reply via email to