On Wed, Jan 31, 2024 at 07:49:51PM -0300, Fabiano Rosas wrote: > pet...@redhat.com writes: > > > From: Peter Xu <pet...@redhat.com> > > > > This patchset contains quite a few refactorings to current multifd: > > > > - It picked up some patches from an old series of mine [0] (the last > > patches were dropped, though; I did the cleanup slightly differently): > > > > I still managed to include one patch to split pending_job, but I > > rewrote the patch here. > > > > - It tries to cleanup multiple multifd paths here and there, the ultimate > > goal is to redefine send_prepare() to be something like: > > > > p->pages -----------> send_prepare() -------------> IOVs > > > > So that there's no obvious change yet on multifd_ops besides redefined > > interface for send_prepare(). We may want a separate OPs for file > > later. > > > > For 2), one benefit is already presented by Fabiano in his other series [1] > > on cleaning up zero copy, but this patchset addressed it quite differently, > > and hopefully also more gradually. The other benefit is for sure if we > > have a more concrete API for send_prepare() and if we can reach an initial > > consensus, then we can have the recent compression accelerators rebased on > > top of this one. > > > > This also prepares for the case where the input can be extended to even not > > any p->pages, but arbitrary data (like VFIO's potential use case in the > > future?). But that will also for later even if reasonable. > > > > Please have a look. Thanks, > > > > [0] https://lore.kernel.org/r/20231022201211.452861-1-pet...@redhat.com > > [1] > > https://lore.kernel.org/qemu-devel/20240126221943.26628-1-faro...@suse.de > > > > Peter Xu (14): > > migration/multifd: Drop stale comment for multifd zero copy > > migration/multifd: multifd_send_kick_main() > > migration/multifd: Drop MultiFDSendParams.quit, cleanup error paths > > migration/multifd: Postpone reset of MultiFDPages_t > > migration/multifd: Drop MultiFDSendParams.normal[] array > > migration/multifd: Separate SYNC request with normal jobs > > migration/multifd: Simplify locking in sender thread > > migration/multifd: Drop pages->num check in sender thread > > migration/multifd: Rename p->num_packets and clean it up > > migration/multifd: Move total_normal_pages accounting > > migration/multifd: Move trace_multifd_send|recv() > > migration/multifd: multifd_send_prepare_header() > > migration/multifd: Move header prepare/fill into send_prepare() > > migration/multifd: Forbid spurious wakeups > > > > migration/multifd.h | 34 +++-- > > migration/multifd-zlib.c | 11 +- > > migration/multifd-zstd.c | 11 +- > > migration/multifd.c | 291 +++++++++++++++++++-------------------- > > 4 files changed, 182 insertions(+), 165 deletions(-) > > This series didn't survive my 9999 iterations test on the opensuse > machine. > > # Running /x86_64/migration/multifd/tcp/tls/x509/reject-anon-client > ... > kill_qemu() detected QEMU death from signal 11 (Segmentation fault) (core > dumped) > > > #0 0x00005575dda06399 in qemu_mutex_lock_impl (mutex=0x18, > file=0x5575ddce9cc3 "../util/qemu-thread-posix.c", line=275) at > ../util/qemu-thread-posix.c:92 > #1 0x00005575dda06a94 in qemu_sem_post (sem=0x18) at > ../util/qemu-thread-posix.c:275 > #2 0x00005575dd56a512 in multifd_send_thread (opaque=0x5575df054ef8) at > ../migration/multifd.c:720 > #3 0x00005575dda0709b in qemu_thread_start (args=0x7fd404001d50) at > ../util/qemu-thread-posix.c:541 > #4 0x00007fd45e8a26ea in start_thread (arg=0x7fd3faffd700) at > pthread_create.c:477 > #5 0x00007fd45cd2150f in clone () at > ../sysdeps/unix/sysv/linux/x86_64/clone.S:95 > > The multifd thread is posting channels_ready with an already freed > multifd_send_state. > > This is the bug Avihai has hit. We're going into multifd_save_cleanup() > so early that multifd_new_send_channel_async() hasn't even had the > chance to set p->running. So it misses the join and frees everything up > while a second multifd thread is just starting.
Thanks for doing that. Would this series makes that bug easier to happen? I didn't do a lot of test on it, it only survived the smoke test and the kicked CI job. I think we can still decide to fix that issues separately; but if this series makes that easier to happen then that's definitely bad.. -- Peter Xu