On Fri, Jun 06, 2025 at 03:37:56PM -0400, Steven Sistare wrote:
> On 6/6/2025 2:06 PM, JAEHOON KIM wrote:
> > On 6/6/2025 12:04 PM, Steven Sistare wrote:
> > > On 6/6/2025 12:06 PM, Daniel P. Berrangé wrote:
> > > > On Fri, Jun 06, 2025 at 11:50:10AM -0400, Steven Sistare wrote:
> > > > > On 6/6/2025 11:43 AM, Daniel P. Berrangé wrote:
> > > > > > On Fri, Jun 06, 2025 at 10:37:28AM -0500, JAEHOON KIM wrote:
> > > > > > > On 6/6/2025 10:12 AM, Steven Sistare wrote:
> > > > > > > > On 6/6/2025 11:06 AM, JAEHOON KIM wrote:
> > > > > > > > > On 6/6/2025 9:14 AM, Steven Sistare wrote:
> > > > > > > > > > On 6/6/2025 9:53 AM, Daniel P. Berrangé wrote:
> > > > > > > > > > > On Thu, Jun 05, 2025 at 06:08:08PM -0500, Jaehoon Kim 
> > > > > > > > > > > wrote:
> > > > > > > > > > > > When the source VM attempts to connect to the 
> > > > > > > > > > > > destination VM's Unix
> > > > > > > > > > > > domain socket(cpr.sock) during CPR transfer, the socket
> > > > > > > > > > > > file might not
> > > > > > > > > > > > yet be exist if the destination side hasn't completed 
> > > > > > > > > > > > the bind
> > > > > > > > > > > > operation. This can lead to connection failures when
> > > > > > > > > > > > running tests with
> > > > > > > > > > > > the qtest framework.
> > > > > > > > > > > 
> > > > > > > > > > > This sounds like a flawed test impl to me - whatever is 
> > > > > > > > > > > initiating
> > > > > > > > > > > the cpr operation on the source has done so prematurely - 
> > > > > > > > > > > it should
> > > > > > > > > > > ensure the dest is ready before starting the operation.
> > > > > > > > > > > 
> > > > > > > > > > > > To address this, add cpr_validate_socket_path(), which 
> > > > > > > > > > > > wait for the
> > > > > > > > > > > > socket file to appear. This avoids intermittent qtest
> > > > > > > > > > > > failures caused by
> > > > > > > > > > > > early connection attempts.
> > > > > > > > > > > 
> > > > > > > > > > > IMHO it is dubious to special case cpr in this way.
> > > > > > > > > > 
> > > > > > > > > > I agree with Daniel, and unfortunately it is not just a 
> > > > > > > > > > test issue;
> > > > > > > > > > every management framework that supports cpr-transfer must 
> > > > > > > > > > add logic to
> > > > > > > > > > wait for the cpr socket to appear in the target before 
> > > > > > > > > > proceeding.
> > > > > > > > > > 
> > > > > > > > > > This is analogous to waiting for the monitor socket to 
> > > > > > > > > > appear before
> > > > > > > > > > connecting to it.
> > > > > > > > > > 
> > > > > > > > > > - Steve
> > > > > > > > > 
> > > > > > > > > Thank you very much for your valuable review and feedback.
> > > > > > > > > 
> > > > > > > > > Just to clarify, the added cpr_validate_socket_path() 
> > > > > > > > > function is
> > > > > > > > > not limited to the test framework.
> > > > > > > > > It is part of the actual CPR implementation and is intended to
> > > > > > > > > ensure correct behavior in all cases, including outside of 
> > > > > > > > > tests.
> > > > > > > > > 
> > > > > > > > > I mentioned the qtest failure simply as an example where this 
> > > > > > > > > issue
> > > > > > > > > became apparent.
> > > > > > > > 
> > > > > > > > Yes, I understand that you understand :)
> > > > > > > > Are you willing to move your fix to the qtest?
> > > > > > > > 
> > > > > > > > - Steve
> > > > > > > 
> > > > > > > Thank you for your question and feedback.
> > > > > > > 
> > > > > > > I agree that the issue could be addressed within the qtest 
> > > > > > > framework to
> > > > > > > improve stability.
> > > > > > > 
> > > > > > > However, this socket readiness check is a fundamental part of CPR 
> > > > > > > transfer
> > > > > > > process,
> > > > > > > and I believe that incorporating cpr_validate_socket_path() 
> > > > > > > directly into
> > > > > > > the CPR implementation helps ensure more reliable behavior
> > > > > > > across all environments - not only during testing.
> > > > > > > 
> > > > > > > Just from my perspective, adding this logic to the CPR code does 
> > > > > > > not
> > > > > > > introduce significant overhead or side effects.
> > > > > > > I would appreciate if you could share more details about your 
> > > > > > > concerns, so I
> > > > > > > can better address them.
> > > > > > 
> > > > > > Requiring a busy wait like this is a sign of a design problem.
> > > > > > 
> > > > > > There needs to be a way to setup the incoming socket listener
> > > > > > without resorting to busy waiting - that's showing a lack of
> > > > > > synchronization.
> > > > > 
> > > > > How is this a design problem?  If I start a program that creates a 
> > > > > listening unix
> > > > > domain socket, I cannot attempt to connect to it until the socket is 
> > > > > created and
> > > > > listening. Clients face the same issue when starting qemu and 
> > > > > connecting to the
> > > > > monitor socket.
> > > > 
> > > > Yes, the monitor has the same conceptual problem, and this caused 
> > > > problems
> > > > for libvirt starting QEMU for many years.
> > > > 
> > > > With the busy wait you risk looping forever if the child (target) QEMU
> > > > already exited for some reason without ever creating the socket. You
> > > > can mitigate this by using 'kill($PID, 0)' in the loop and looking
> > > > for -ERSCH, but this only works if you know the pid involved.
> > > > 
> > > > One option is to use 'daemonize' such that when the parent sees the 
> > > > initial
> > > > QEMU process leader exit, the parent has a guarantee that the daemonized
> > > > QEMU already has the UNIX listener open, and any failure indicates QEMU
> > > > already quit.
> > > > 
> > > > The other option is to use FD passing such that QEMU is not responsible
> > > > for opening the listener socket - it gets passed a pre-opened listener
> > > > FD, so the parent has a guarantee it can successfull connect immediately
> > > > and any failure indicates QEMU already quit.
> > > > 
> > > > For the tests, passing a pre-opened UNIX socket FD could work, but I'm
> > > > still curious why this is only a problem for the CPR tests, and not
> > > > the other migration tests which don't use 'defer'. What has made CPR
> > > > special to expose a race ?
> > > 
> > > For normal migration, target qemu listens on the migration socket, then 
> > > listens
> > > on the monitor.  After the client connects to the monitor (waiting for it 
> > > to appear
> > > as needed), them the migration socket already exists.
> > > 
> > > For cpr, target qemu creates the cpr socket and listens before the 
> > > monitor is
> > > started, which is necessary because cpr state is needed before backends or
> > > devices are created.
> > > 
> > > A few months back I sent a series to start the monitor first (I think I 
> > > called
> > > it the precreate phase), but it was derailed over discussions about 
> > > allowing
> > > qemu to start with no arguments and be configured exclusively via the 
> > > monitor.
> > > 
> > > - Steve
> > 
> > Thank you for sharing your thoughts.
> > 
> > I agree that busy waiting is not ideal.
> > However, considering the timing of when target QEMU creates and begins 
> > listening on the socket,
> > I think there is currently no reliable way for the host to check the 
> > socket's listening state.
> > This also implies that FD passing is not a viable option in this case.
> > 
> > As for the 'defer' option in qtest,
> > it doesn't cause race-condition issues because the target enters the 
> > listening state during the option processing.
> > 
> > Of course, to address this issue,
> > I could create a wait_for_socket() function similar to wait_for_serial() in 
> > qtest framework.
> > Since the socket might already be created, I cannot simply wait for the 
> > file to appear using file system notification APIs like inotify,
> > so busy-waiting would still be necessary.
> > 
> > I would appreciate hearing any further thoughts you might have on this.
> 
> The easiest solution, with no interface changes, is adding wait_for_socket() 
> in qtest,
> with this addition from Daniel:
> 
>   "With the busy wait you risk looping forever if the child (target) QEMU
>   already exited for some reason without ever creating the socket. You
>   can mitigate this by using 'kill($PID, 0)' in the loop and looking
>   for -ERSCH, but this only works if you know the pid involved."
> 
> Daniel also suggested:
>   "For the tests, passing a pre-opened UNIX socket FD could work"
> 
> Note we can not use any of the standard chardev options to specify such a 
> socket,
> because the cpr socket is created before chardevs are created.
> 
> Perhaps we could specify the fd in an extension of the MigrationChannel 
> MigrationAddress.
> { 'union': 'MigrationAddress',
>   'base': { 'transport' : 'MigrationAddressType'},
>   'discriminator': 'transport',
>   'data': {
>     'socket': 'SocketAddress',
>     'exec': 'MigrationExecCommand',
>     'rdma': 'InetSocketAddress',
>     'file': 'FileMigrationArgs',
>     'fd':   'FdMigrationArgs' } }           <-- add this
> 
> That would be useful for all clients, but this is asking a lot from you,
> when you are just trying to fix the tests.

Note, 'SocketAddress' already has an option for declaring a FD that
represents a socket.

With regards,
Daniel
-- 
|: https://berrange.com      -o-    https://www.flickr.com/photos/dberrange :|
|: https://libvirt.org         -o-            https://fstop138.berrange.com :|
|: https://entangle-photo.org    -o-    https://www.instagram.com/dberrange :|


Reply via email to