On Mon, Aug 04, 2025 at 04:27:45PM -0300, Fabiano Rosas wrote: > Juraj Marcin <jmar...@redhat.com> writes: > > > Hi Daniel, > > > > On 2025-08-01 18:02, Daniel P. Berrangé wrote: > >> This is a followup to previously merged patches that claimed to > >> workaround the gnutls bug impacting migration, but in fact were > >> essentially non-functional. Juraj Marcin pointed this out, and > >> this new patch tweaks the workaround to make it actually do > >> something useful. > >> > >> Daniel P. Berrangé (2): > >> migration: simplify error reporting after channel read > >> migration: fix workaround for gnutls thread safety > >> > >> crypto/tlssession.c | 16 ---------------- > >> migration/qemu-file.c | 22 +++++++++++++++++----- > >> 2 files changed, 17 insertions(+), 21 deletions(-) > >> > > > > thanks for finding a fix for the workaround. I have tested it and it > > resolves the issue. > > > > However, it significantly slows down migration, even with the workaround > > disabled (and thus no locking). When benchmarking, I used the fixed > > version of GNUTLS, VM with 20GB of RAM which were fully written to > > before starting a normal migration with no workload during the > > migration. > > > > Test cases: > > [1]: before this patchset > > [2]: with this patchset applied and GNUTLS workaround enabled > > [2]: with this patchset applied and GNUTLS workaround disabled > > > > | Total time | Throughput | Transfered bytes | > > --+------------+------------+------------------+ > > 1 | 31 192 ms | 5450 mpbs | 21 230 973 763 | > > 2 | 74 147 ms | 2291 mbps | 21 232 849 066 | > > 3 | 72 426 ms | 2343 mbps | 21 215 009 392 | > > Thanks testing this. I had just managed to convince myself that there > wouldn't be any performance issues. > > The yield at every buffer fill on the incoming side is probably way more > impactful than the poll on the RP.
Yeah, that's an unacceptable penalty on the incoming side for sure. How about we simply change the outbound migration channel to be in non-blocking mode ? I originally put it in blocking mode way back in 9e4d2b98ee98f4cee50d671e500eceeefa751ee0, but if I look at the QEMUFile impl of qemu_fill_buffer and qemu_fflush, but should be coping with a non-blocking socket. qemu_fill_buffer has explicit code to wait, and qemu_fflush uses the _all() variant whcih has built-in support for waiting. So I'm not seeing an obvious need to run the channel in blocking mode. Using non-blocking will prevent the return path thuread setting in a read() call, so we won't have mutual exclusion between read and write which this patch was trying to avoid ie, this delta on top of this patch: diff --git a/migration/migration.c b/migration/migration.c index 10c216d25d..1eaabc1f19 100644 --- a/migration/migration.c +++ b/migration/migration.c @@ -4002,7 +4002,7 @@ void migration_connect(MigrationState *s, Error *error_in) } migration_rate_set(rate_limit); - qemu_file_set_blocking(s->to_dst_file, true); + qemu_file_set_blocking(s->to_dst_file, false); /* * Open the return path. For postcopy, it is used exclusively. For diff --git a/migration/qemu-file.c b/migration/qemu-file.c index cf6115e699..8ee44c5ac9 100644 --- a/migration/qemu-file.c +++ b/migration/qemu-file.c @@ -338,22 +338,6 @@ static ssize_t coroutine_mixed_fn qemu_fill_buffer(QEMUFile *f) return 0; } - /* - * This feature triggers acquisition of mutexes around every - * read and write. Thus we must not sit in a blocking read - * if this is set, but must instead poll proactively. This - * does not work with some channel types, however, so must - * only pre-poll when the featre is set. - */ - if (qio_channel_has_feature(f->ioc, - QIO_CHANNEL_FEATURE_CONCURRENT_IO)) { - if (qemu_in_coroutine()) { - qio_channel_yield(f->ioc, G_IO_IN); - } else { - qio_channel_wait(f->ioc, G_IO_IN); - } - } - do { struct iovec iov = { f->buf + pending, IO_BUF_SIZE - pending }; len = qio_channel_readv_full(f->ioc, &iov, 1, pfds, pnfd, 0, With regards, Daniel -- |: https://berrange.com -o- https://www.flickr.com/photos/dberrange :| |: https://libvirt.org -o- https://fstop138.berrange.com :| |: https://entangle-photo.org -o- https://www.instagram.com/dberrange :|