On Fri, Jun 06, 2025 at 03:37:56PM -0400, Steven Sistare wrote: > On 6/6/2025 2:06 PM, JAEHOON KIM wrote: > > On 6/6/2025 12:04 PM, Steven Sistare wrote: > > > On 6/6/2025 12:06 PM, Daniel P. Berrangé wrote: > > > > On Fri, Jun 06, 2025 at 11:50:10AM -0400, Steven Sistare wrote: > > > > > On 6/6/2025 11:43 AM, Daniel P. Berrangé wrote: > > > > > > On Fri, Jun 06, 2025 at 10:37:28AM -0500, JAEHOON KIM wrote: > > > > > > > On 6/6/2025 10:12 AM, Steven Sistare wrote: > > > > > > > > On 6/6/2025 11:06 AM, JAEHOON KIM wrote: > > > > > > > > > On 6/6/2025 9:14 AM, Steven Sistare wrote: > > > > > > > > > > On 6/6/2025 9:53 AM, Daniel P. Berrangé wrote: > > > > > > > > > > > On Thu, Jun 05, 2025 at 06:08:08PM -0500, Jaehoon Kim > > > > > > > > > > > wrote: > > > > > > > > > > > > When the source VM attempts to connect to the > > > > > > > > > > > > destination VM's Unix > > > > > > > > > > > > domain socket(cpr.sock) during CPR transfer, the socket > > > > > > > > > > > > file might not > > > > > > > > > > > > yet be exist if the destination side hasn't completed > > > > > > > > > > > > the bind > > > > > > > > > > > > operation. This can lead to connection failures when > > > > > > > > > > > > running tests with > > > > > > > > > > > > the qtest framework. > > > > > > > > > > > > > > > > > > > > > > This sounds like a flawed test impl to me - whatever is > > > > > > > > > > > initiating > > > > > > > > > > > the cpr operation on the source has done so prematurely - > > > > > > > > > > > it should > > > > > > > > > > > ensure the dest is ready before starting the operation. > > > > > > > > > > > > > > > > > > > > > > > To address this, add cpr_validate_socket_path(), which > > > > > > > > > > > > wait for the > > > > > > > > > > > > socket file to appear. This avoids intermittent qtest > > > > > > > > > > > > failures caused by > > > > > > > > > > > > early connection attempts. > > > > > > > > > > > > > > > > > > > > > > IMHO it is dubious to special case cpr in this way. > > > > > > > > > > > > > > > > > > > > I agree with Daniel, and unfortunately it is not just a > > > > > > > > > > test issue; > > > > > > > > > > every management framework that supports cpr-transfer must > > > > > > > > > > add logic to > > > > > > > > > > wait for the cpr socket to appear in the target before > > > > > > > > > > proceeding. > > > > > > > > > > > > > > > > > > > > This is analogous to waiting for the monitor socket to > > > > > > > > > > appear before > > > > > > > > > > connecting to it. > > > > > > > > > > > > > > > > > > > > - Steve > > > > > > > > > > > > > > > > > > Thank you very much for your valuable review and feedback. > > > > > > > > > > > > > > > > > > Just to clarify, the added cpr_validate_socket_path() > > > > > > > > > function is > > > > > > > > > not limited to the test framework. > > > > > > > > > It is part of the actual CPR implementation and is intended to > > > > > > > > > ensure correct behavior in all cases, including outside of > > > > > > > > > tests. > > > > > > > > > > > > > > > > > > I mentioned the qtest failure simply as an example where this > > > > > > > > > issue > > > > > > > > > became apparent. > > > > > > > > > > > > > > > > Yes, I understand that you understand :) > > > > > > > > Are you willing to move your fix to the qtest? > > > > > > > > > > > > > > > > - Steve > > > > > > > > > > > > > > Thank you for your question and feedback. > > > > > > > > > > > > > > I agree that the issue could be addressed within the qtest > > > > > > > framework to > > > > > > > improve stability. > > > > > > > > > > > > > > However, this socket readiness check is a fundamental part of CPR > > > > > > > transfer > > > > > > > process, > > > > > > > and I believe that incorporating cpr_validate_socket_path() > > > > > > > directly into > > > > > > > the CPR implementation helps ensure more reliable behavior > > > > > > > across all environments - not only during testing. > > > > > > > > > > > > > > Just from my perspective, adding this logic to the CPR code does > > > > > > > not > > > > > > > introduce significant overhead or side effects. > > > > > > > I would appreciate if you could share more details about your > > > > > > > concerns, so I > > > > > > > can better address them. > > > > > > > > > > > > Requiring a busy wait like this is a sign of a design problem. > > > > > > > > > > > > There needs to be a way to setup the incoming socket listener > > > > > > without resorting to busy waiting - that's showing a lack of > > > > > > synchronization. > > > > > > > > > > How is this a design problem? If I start a program that creates a > > > > > listening unix > > > > > domain socket, I cannot attempt to connect to it until the socket is > > > > > created and > > > > > listening. Clients face the same issue when starting qemu and > > > > > connecting to the > > > > > monitor socket. > > > > > > > > Yes, the monitor has the same conceptual problem, and this caused > > > > problems > > > > for libvirt starting QEMU for many years. > > > > > > > > With the busy wait you risk looping forever if the child (target) QEMU > > > > already exited for some reason without ever creating the socket. You > > > > can mitigate this by using 'kill($PID, 0)' in the loop and looking > > > > for -ERSCH, but this only works if you know the pid involved. > > > > > > > > One option is to use 'daemonize' such that when the parent sees the > > > > initial > > > > QEMU process leader exit, the parent has a guarantee that the daemonized > > > > QEMU already has the UNIX listener open, and any failure indicates QEMU > > > > already quit. > > > > > > > > The other option is to use FD passing such that QEMU is not responsible > > > > for opening the listener socket - it gets passed a pre-opened listener > > > > FD, so the parent has a guarantee it can successfull connect immediately > > > > and any failure indicates QEMU already quit. > > > > > > > > For the tests, passing a pre-opened UNIX socket FD could work, but I'm > > > > still curious why this is only a problem for the CPR tests, and not > > > > the other migration tests which don't use 'defer'. What has made CPR > > > > special to expose a race ? > > > > > > For normal migration, target qemu listens on the migration socket, then > > > listens > > > on the monitor. After the client connects to the monitor (waiting for it > > > to appear > > > as needed), them the migration socket already exists. > > > > > > For cpr, target qemu creates the cpr socket and listens before the > > > monitor is > > > started, which is necessary because cpr state is needed before backends or > > > devices are created. > > > > > > A few months back I sent a series to start the monitor first (I think I > > > called > > > it the precreate phase), but it was derailed over discussions about > > > allowing > > > qemu to start with no arguments and be configured exclusively via the > > > monitor. > > > > > > - Steve > > > > Thank you for sharing your thoughts. > > > > I agree that busy waiting is not ideal. > > However, considering the timing of when target QEMU creates and begins > > listening on the socket, > > I think there is currently no reliable way for the host to check the > > socket's listening state. > > This also implies that FD passing is not a viable option in this case. > > > > As for the 'defer' option in qtest, > > it doesn't cause race-condition issues because the target enters the > > listening state during the option processing. > > > > Of course, to address this issue, > > I could create a wait_for_socket() function similar to wait_for_serial() in > > qtest framework. > > Since the socket might already be created, I cannot simply wait for the > > file to appear using file system notification APIs like inotify, > > so busy-waiting would still be necessary. > > > > I would appreciate hearing any further thoughts you might have on this. > > The easiest solution, with no interface changes, is adding wait_for_socket() > in qtest, > with this addition from Daniel: > > "With the busy wait you risk looping forever if the child (target) QEMU > already exited for some reason without ever creating the socket. You > can mitigate this by using 'kill($PID, 0)' in the loop and looking > for -ERSCH, but this only works if you know the pid involved." > > Daniel also suggested: > "For the tests, passing a pre-opened UNIX socket FD could work" > > Note we can not use any of the standard chardev options to specify such a > socket, > because the cpr socket is created before chardevs are created. > > Perhaps we could specify the fd in an extension of the MigrationChannel > MigrationAddress. > { 'union': 'MigrationAddress', > 'base': { 'transport' : 'MigrationAddressType'}, > 'discriminator': 'transport', > 'data': { > 'socket': 'SocketAddress', > 'exec': 'MigrationExecCommand', > 'rdma': 'InetSocketAddress', > 'file': 'FileMigrationArgs', > 'fd': 'FdMigrationArgs' } } <-- add this > > That would be useful for all clients, but this is asking a lot from you, > when you are just trying to fix the tests.
Note, 'SocketAddress' already has an option for declaring a FD that represents a socket. With regards, Daniel -- |: https://berrange.com -o- https://www.flickr.com/photos/dberrange :| |: https://libvirt.org -o- https://fstop138.berrange.com :| |: https://entangle-photo.org -o- https://www.instagram.com/dberrange :|