Please strace both virsh and libvirtd (you can attach to it by pid),
and make sure that the strace command uses the "-f" switch (i.e.
traces all threads).

On Wed, Jul 8, 2020 at 6:20 PM Andrei Mikhailovsky <and...@arhont.com> wrote:
>
> Jason,
>
> After adding the 1:storage to the log line of the config and restarting the 
> service I do not see anything in the logs. I've started the "virsh pool-list" 
> command several times and there is absolutely nothing in the logs. The 
> command keeps hanging
>
>
> running the strace virsh pool-list shows (the last 50-100 lines or so):
>
>
>
> ioctl(0, TCGETS, {B38400 opost isig icanon echo ...}) = 0
> ioctl(0, TCGETS, {B38400 opost isig icanon echo ...}) = 0
> getuid()                                = 0
> geteuid()                               = 0
> openat(AT_FDCWD, "/usr/lib/x86_64-linux-gnu/gconv/gconv-modules.cache", 
> O_RDONLY) = 3
> fstat(3, {st_mode=S_IFREG|0644, st_size=26376, ...}) = 0
> mmap(NULL, 26376, PROT_READ, MAP_SHARED, 3, 0) = 0x7fe979933000
> close(3)                                = 0
> futex(0x7fe978505a08, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> uname({sysname="Linux", nodename="ais-cloudhost1", ...}) = 0
> futex(0x7fe9790bfce0, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> socket(AF_INET6, SOCK_DGRAM, IPPROTO_IP) = 3
> close(3)                                = 0
> futex(0x7fe9790c0700, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> pipe2([3, 4], O_NONBLOCK|O_CLOEXEC)     = 0
> mmap(NULL, 8392704, PROT_NONE, MAP_PRIVATE|MAP_ANONYMOUS|MAP_STACK, -1, 0) = 
> 0x7fe96ca98000
> mprotect(0x7fe96ca99000, 8388608, PROT_READ|PROT_WRITE) = 0
> clone(child_stack=0x7fe96d297db0, 
> flags=CLONE_VM|CLONE_FS|CLONE_FILES|CLONE_SIGHAND|CLONE_THREAD|CLONE_SYSVSEM|CLONE_SE
> TTLS|CLONE_PARENT_SETTID|CLONE_CHILD_CLEARTID, parent_tidptr=0x7fe96d2989d0, 
> tls=0x7fe96d298700, child_tidptr=0x7fe96d2
> 989d0) = 54218
> futex(0x7fe9790bffb8, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> futex(0x7fe9790c06f8, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> geteuid()                               = 0
> access("/etc/libvirt/libvirt.conf", F_OK) = 0
> openat(AT_FDCWD, "/etc/libvirt/libvirt.conf", O_RDONLY) = 5
> read(5, "#\n# This can be used to setup UR"..., 8192) = 547
> read(5, "", 7645)                       = 0
> close(5)                                = 0
> getuid()                                = 0
> geteuid()                               = 0
> access("/proc/vz", F_OK)                = -1 ENOENT (No such file or 
> directory)
> geteuid()                               = 0
> getuid()                                = 0
> geteuid()                               = 0
> socket(AF_UNIX, SOCK_STREAM, 0)         = 5
> connect(5, {sa_family=AF_UNIX, sun_path="/var/run/libvirt/libvirt-sock"}, 
> 110) = 0
> getsockname(5, {sa_family=AF_UNIX}, [128->2]) = 0
> futex(0x7fe9790c0a08, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> fcntl(5, F_GETFD)                       = 0
> fcntl(5, F_SETFD, FD_CLOEXEC)           = 0
> fcntl(5, F_GETFL)                       = 0x2 (flags O_RDWR)
> fcntl(5, F_SETFL, O_RDWR|O_NONBLOCK)    = 0
> futex(0x7fe9790c0908, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> pipe2([6, 7], O_CLOEXEC)                = 0
> write(4, "\0", 1)                       = 1
> futex(0x7fe9790bfb60, FUTEX_WAKE_PRIVATE, 1) = 1
> futex(0x7fe9790c09d0, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> futex(0x7fe9790c0920, FUTEX_WAKE_PRIVATE, 2147483647) = 0
> brk(0x5598ffebb000)                     = 0x5598ffebb000
> write(4, "\0", 1)                       = 1
> futex(0x7fe9790bfb60, FUTEX_WAKE_PRIVATE, 1) = 1
> rt_sigprocmask(SIG_BLOCK, [PIPE CHLD WINCH], [], 8) = 0
> poll([{fd=5, events=POLLOUT}, {fd=6, events=POLLIN}], 2, -1) = 1 ([{fd=5, 
> revents=POLLOUT}])
> rt_sigprocmask(SIG_SETMASK, [], NULL, 8) = 0
> write(5, "\0\0\0\34 \0\200\206\0\0\0\1\0\0\0B\0\0\0\0\0\0\0\0\0\0\0\0", 28) = 
> 28
> rt_sigprocmask(SIG_BLOCK, [PIPE CHLD WINCH], [], 8) = 0
> poll([{fd=5, events=POLLIN}, {fd=6, events=POLLIN}], 2, -1
>
>
>
>
>
>
> It get's stuck at the last line and there is nothing happening.
>
> Andrei
>
> ----- Original Message -----
> > From: "Jason Dillaman" <jdill...@redhat.com>
> > To: "Andrei Mikhailovsky" <and...@arhont.com>
> > Cc: "ceph-users" <ceph-users@ceph.io>
> > Sent: Tuesday, 7 July, 2020 23:33:03
> > Subject: Re: [ceph-users] Re: Octopus upgrade breaks Ubuntu 18.04 libvirt
>
> > On Tue, Jul 7, 2020 at 5:14 PM Andrei Mikhailovsky <and...@arhont.com> 
> > wrote:
> >>
> >> Hi Jason,
> >> The extract from the debug log file is given below in the first message. 
> >> It just
> >> repeats those lines every so often.
> >>
> >> I can't find anything else.
> >
> > I would expect lots of debug logs from the storage backend. Do you
> > have a "1:storage" entry in your libvirtd.conf?
> >
> >>  Cheers
> >> ----- Original Message -----
> >> > From: "Jason Dillaman" <jdill...@redhat.com>
> >> > To: "Andrei Mikhailovsky" <and...@arhont.com>
> >> > Cc: "ceph-users" <ceph-users@ceph.io>
> >> > Sent: Tuesday, 7 July, 2020 16:33:25
> >> > Subject: Re: [ceph-users] Re: Octopus upgrade breaks Ubuntu 18.04 libvirt
> >>
> >> > On Tue, Jul 7, 2020 at 11:07 AM Andrei Mikhailovsky <and...@arhont.com> 
> >> > wrote:
> >> >>
> >> >> I've left the virsh pool-list command 'hang' for a while and it did 
> >> >> eventually
> >> >> get the results back. In about 4 hours!
> >> >
> >> > Perhaps enable the debug logging of libvirt [1] to determine what it's
> >> > spending its time on?
> >> >
> >> >> root@ais-cloudhost1:/home/andrei# time virsh pool-list
> >> >>  Name                 State      Autostart
> >> >> -------------------------------------------
> >> >>  12ca033f-e673-4060-8db9-909d79650f39 active     no
> >> >>  bcc753c6-e47a-3b7c-904a-fcc1d0a594c5 active     no
> >> >>  cf771bc7-8998-354d-8e10-5564585a3c20 active     no
> >> >>  d8d5ec36-3cb0-39af-8fc6-084a4abd5d28 active     no
> >> >>
> >> >>
> >> >> real    234m23.877s
> >> >> user    0m0.351s
> >> >> sys     0m0.506s
> >> >>
> >> >>
> >> >>
> >> >> The second attempt was a mere 2 hours with a bit.
> >> >>
> >> >>
> >> >> root@ais-cloudhost1:/home/andrei# time virsh pool-list
> >> >>  Name                 State      Autostart
> >> >> -------------------------------------------
> >> >>  12ca033f-e673-4060-8db9-909d79650f39 active     no
> >> >>  bcc753c6-e47a-3b7c-904a-fcc1d0a594c5 active     no
> >> >>  cf771bc7-8998-354d-8e10-5564585a3c20 active     no
> >> >>  d8d5ec36-3cb0-39af-8fc6-084a4abd5d28 active     no
> >> >>
> >> >>
> >> >> real    148m54.763s
> >> >> user    0m0.241s
> >> >> sys     0m0.304s
> >> >>
> >> >>
> >> >>
> >> >> Am I the only person having these issues with libvirt and Octopus 
> >> >> release?
> >> >>
> >> >> Cheers
> >> >>
> >> >> ----- Original Message -----
> >> >> > From: "Andrei Mikhailovsky" <and...@arhont.com>
> >> >> > To: "ceph-users" <ceph-users@ceph.io>
> >> >> > Sent: Monday, 6 July, 2020 19:27:25
> >> >> > Subject: [ceph-users] Re: Octopus upgrade breaks Ubuntu 18.04 libvirt
> >> >>
> >> >> > A quick update.
> >> >> >
> >> >> > I have done a fresh install of the CloudStack host server running 
> >> >> > Ubuntu 18.04
> >> >> > with the latest updates. I've installed ceph 12.x and connected it to
> >> >> > Cloudstack which uses kvm/libvirt/ceph/rbd. The rest of the ceph 
> >> >> > services
> >> >> > (mon,mgr,osd,etc) are all running 15.2.3. Works like a charm.
> >> >> >
> >> >> > As soon as I've updated the host server to version 15.2.3, Libvirt 
> >> >> > stopped
> >> >> > working. It just hangs without doing much it seems. Common commands 
> >> >> > like 'virsh
> >> >> > pool-list' or 'virsh list' are just hanging. I've strace the process 
> >> >> > and it
> >> >> > just doesn't show any activity.
> >> >> >
> >> >> >
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: info : 
> >> >> > virEventPollUpdateTimeout:265 :
> >> >> > EVENT_POLL_UPDATE_TIMEOUT: timer=993 frequen
> >> >> > cy=5000
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: debug : 
> >> >> > virEventPollUpdateTimeout:282 : Set
> >> >> > timer freq=5000 expires=1594059521930
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: debug : 
> >> >> > virEventPollInterruptLocked:722 :
> >> >> > Skip interrupt, 1 140123172218240
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: info : 
> >> >> > virEventPollUpdateHandle:152 :
> >> >> > EVENT_POLL_UPDATE_HANDLE: watch=1004 events=1
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: debug : 
> >> >> > virEventPollInterruptLocked:722 :
> >> >> > Skip interrupt, 1 140123172218240
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: debug : 
> >> >> > virEventPollCleanupTimeouts:525 :
> >> >> > Cleanup 8
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: debug : 
> >> >> > virEventPollCleanupHandles:574 :
> >> >> > Cleanup 22
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: debug : 
> >> >> > virEventRunDefaultImpl:324 : running
> >> >> > default event implementation
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: debug : 
> >> >> > virEventPollCleanupTimeouts:525 :
> >> >> > Cleanup 8
> >> >> > 2020-07-06 18:18:36.930+0000: 3273: debug : 
> >> >> > virEventPollCleanupHandles:574 :
> >> >> > Cleanup 22
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=0 w=1, f=5 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=1 w=2, f=7 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=2 w=3, f=10 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=3 w=4, f=11 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=4 w=5, f=12 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=5 w=6, f=13 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=6 w=7, f=14 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=7 w=8, f=15 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=8 w=9, f=16 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=9 w=10, f=17 e=1 d=0
> >> >> > 2020-07-06 18:18:36.931+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=10 w=11, f=18 e=1 d=0
> >> >> >
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=11 w=12, f=19 e=0 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=12 w=13, f=19 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=13 w=14, f=24 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=14 w=15, f=25 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=15 w=19, f=26 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=16 w=82, f=79 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=17 w=944, f=22 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=18 w=993, f=82 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=19 w=1001, f=30 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=20 w=1002, f=33 e=1 d=0
> >> >> > 2020-07-06 18:18:36.932+0000: 3273: debug : 
> >> >> > virEventPollMakePollFDs:401 :
> >> >> > Prepare n=21 w=1004, f=83 e=1 d=0
> >> >> > 2020-07-06 18:18:36.933+0000: 3273: debug : 
> >> >> > virEventPollCalculateTimeout:338 :
> >> >> > Calculate expiry of 8 timers
> >> >> > 2020-07-06 18:18:36.933+0000: 3273: debug : 
> >> >> > virEventPollCalculateTimeout:346 :
> >> >> > Got a timeout scheduled for 1594059521930
> >> >> > 2020-07-06 18:18:36.933+0000: 3273: debug : 
> >> >> > virEventPollCalculateTimeout:359 :
> >> >> > Schedule timeout then=1594059521930 now=1594059516933
> >> >> > 2020-07-06 18:18:36.933+0000: 3273: debug : 
> >> >> > virEventPollCalculateTimeout:369 :
> >> >> > Timeout at 1594059521930 due in 4997 ms
> >> >> > 2020-07-06 18:18:36.933+0000: 3273: info : virEventPollRunOnce:640 :
> >> >> > EVENT_POLL_RUN: nhandles=21 timeout=4997
> >> >> >
> >> >> >
> >> >> > The ceph itself seems to work, i.e. I can execute ceph -s / rbd -p 
> >> >> > <pool> ls -l,
> >> >> > etc. That produces the output. It's just the libvirt seems to be no 
> >> >> > joy.
> >> >> >
> >> >> > The version of libvirt installed is:
> >> >> >
> >> >> > libvirt-bin             4.0.0-1ubuntu8.1
> >> >> >
> >> >> >
> >> >> >
> >> >> > Any idea how I can make ceph Octopus to play nicely with libvirt?
> >> >> >
> >> >> > Cheers
> >> >> >
> >> >> > Andrei
> >> >> >
> >> >> > ----- Original Message -----
> >> >> >> From: "Andrei Mikhailovsky" <and...@arhont.com>
> >> >> >> To: "ceph-users" <ceph-users@ceph.io>
> >> >> >> Sent: Monday, 29 June, 2020 20:40:01
> >> >> >> Subject: [ceph-users] Octopus upgrade breaks Ubuntu 18.04 libvirt
> >> >> >
> >> >> >> Hello,
> >> >> >>
> >> >> >> I've upgraded ceph to Octopus (15.2.3 from repo) on one of the 
> >> >> >> Ubuntu 18.04 host
> >> >> >> servers. The update caused problem with libvirtd which hangs when it 
> >> >> >> tries to
> >> >> >> access the storage pools. The problem doesn't exist on Nautilus. The 
> >> >> >> libvirtd
> >> >> >> process simply hangs. Nothing seem to happen. The log file for the 
> >> >> >> libvirtd
> >> >> >> shows:
> >> >> >>
> >> >> >> 2020-06-29 19:30:51.556+0000: 12040: debug : 
> >> >> >> virNetlinkEventCallback:707 :
> >> >> >> dispatching to max 0 clients, called from event watch 11
> >> >> >> 2020-06-29 19:30:51.556+0000: 12040: debug : 
> >> >> >> virNetlinkEventCallback:720 : event
> >> >> >> not handled.
> >> >> >> 2020-06-29 19:30:51.556+0000: 12040: debug : 
> >> >> >> virNetlinkEventCallback:707 :
> >> >> >> dispatching to max 0 clients, called from event watch 11
> >> >> >> 2020-06-29 19:30:51.556+0000: 12040: debug : 
> >> >> >> virNetlinkEventCallback:720 : event
> >> >> >> not handled.
> >> >> >> 2020-06-29 19:30:51.557+0000: 12040: debug : 
> >> >> >> virNetlinkEventCallback:707 :
> >> >> >> dispatching to max 0 clients, called from event watch 11
> >> >> >> 2020-06-29 19:30:51.557+0000: 12040: debug : 
> >> >> >> virNetlinkEventCallback:720 : event
> >> >> >> not handled.
> >> >> >> 2020-06-29 19:30:51.591+0000: 12040: debug : 
> >> >> >> virNetlinkEventCallback:707 :
> >> >> >> dispatching to max 0 clients, called from event watch 11
> >> >> >> 2020-06-29 19:30:51.591+0000: 12040: debug : 
> >> >> >> virNetlinkEventCallback:720 : event
> >> >> >> not handled.
> >> >> >>
> >> >> >> Running strace on the libvirtd process shows:
> >> >> >>
> >> >> >> root@ais-cloudhost1:/home/andrei# strace -p 12040
> >> >> >> strace: Process 12040 attached
> >> >> >> restart_syscall(<... resuming interrupted poll ...>
> >> >> >>
> >> >> >>
> >> >> >> Nothing happens after that point.
> >> >> >>
> >> >> >> The same host server can get access to the ceph cluster and the 
> >> >> >> pools by running
> >> >> >> ceph -s or rbd -p <pool> ls -l commands for example.
> >> >> >>
> >> >> >> Need some help to get the host servers working again with Octopus.
> >> >> >>
> >> >> >> Cheers
> >> >> >> _______________________________________________
> >> >> >> ceph-users mailing list -- ceph-users@ceph.io
> >> >> >> To unsubscribe send an email to ceph-users-le...@ceph.io
> >> >> > _______________________________________________
> >> >> > ceph-users mailing list -- ceph-users@ceph.io
> >> >> > To unsubscribe send an email to ceph-users-le...@ceph.io
> >> >> _______________________________________________
> >> >> ceph-users mailing list -- ceph-users@ceph.io
> >> >> To unsubscribe send an email to ceph-users-le...@ceph.io
> >> >>
> >> >
> >> > [1] https://wiki.libvirt.org/page/DebugLogs
> >> >
> >> > --
> >> > Jason
> >>
> >
> >
> > --
> > Jason
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io



-- 
Alexander E. Patrakov
CV: http://pc.cd/PLz7
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to